File size: 2,931 Bytes
fbcfe0e | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 | ---
language: en
license: mit
tags:
- federated-learning
- finance
- sentiment-analysis
- bert
- finbert
- fedavg
library_name: transformers
pipeline_tag: text-classification
authors:
- Harsh Prasad
- Sai Dhole
---
## FinBERTโFedAvg: Federated Averaging for Financial Sentiment Analysis
---
### ๐ Model Summary
This model is a **federated version of FinBERT** fine-tuned for
**financial sentiment classification (Positive / Negative / Neutral)**.
Training is performed across **three clients**:
* Financial Twitter posts
* Financial news headlines
* Financial reports & statements
This model is trained using the **Federated Averaging (FedAvg)** algorithm,
where each client trains locally on its own data and only **model weights** are shared.
No raw data is exchanged, supporting privacy-preserving learning.
This model is part of a research project comparing:
* FedAvg
* FedProx
* Adaptive Aggregation
for federated financial NLP.
---
### ๐ง Intended Use
Designed for:
* Financial sentiment research
* Risk & market analytics
* Academic exploration of federated learning
Not intended for automated trading without expert oversight.
---
### ๐ Model Architecture
Base Model:
```
ProsusAI/finbert
```
Task:
```
Sequence classification โ 3 classes
```
Training Setup:
```
3 federation clients
10 global rounds
3 local epochs
FedAvg aggregation
````
---
### ๐ Client Data Sources
| Client | Data Type |
| -------- | ----------------- |
| Client-1 | Financial Twitter |
| Client-2 | Financial News |
| Client-3 | Financial Reports |
No raw data is shared between clients.
---
### ๐ Privacy Advantage
Only model updates are exchanged โ not text data.
This supports data governance and privacy-aware ML.
---
### ๐ Performance (Validation)
| Method | Final Avg F1-Score |
| ------ | ------------------ |
| FedAvg | **0.846** |
FedAvg provided **strong and stable global performance**
across heterogeneous financial text sources.
---
### ๐ Example Usage
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained(
"harshprasad03/FinBERT-FedAvg"
)
tokenizer = AutoTokenizer.from_pretrained(
"harshprasad03/FinBERT-FedAvg"
)
text = "Tech stocks fell after negative earnings guidance."
inputs = tokenizer(text, return_tensors="pt")
outputs = model(**inputs)
prob = torch.softmax(outputs.logits, dim=1)
print(prob)
````
---
### โ ๏ธ Limitations
* Trained only on finance-domain text
* Sentiment โ market prediction
* Model may inherit dataset biases
* Designed for research use
---
### ๐ Citation
```
Harsh Prasad, Sai Dhole (2025).
FedAvg-based Federated FinBERT for Financial Sentiment Analysis.
```
---
### ๐จโ๐ป Authors
**Harsh Prasad**
AI and ML Research
**Sai Dhole**
AI and ML Research
--- |