π Incremental label | Acc: 0.950, F1: 0.965
Browse files- README.md +15 -9
- {checkpoint-5 β checkpoint-32}/config.json +0 -0
- {checkpoint-5 β checkpoint-32}/model.safetensors +1 -1
- {checkpoint-5 β checkpoint-32}/special_tokens_map.json +0 -0
- {checkpoint-5 β checkpoint-32}/tokenizer.json +0 -0
- {checkpoint-5 β checkpoint-32}/tokenizer_config.json +0 -0
- {checkpoint-5 β checkpoint-32}/trainer_state.json +38 -17
- {checkpoint-5 β checkpoint-32}/vocab.txt +0 -0
- model.safetensors +1 -1
- training_args.bin +1 -1
README.md
CHANGED
|
@@ -6,6 +6,7 @@ tags:
|
|
| 6 |
- sentiment-analysis
|
| 7 |
- finbert
|
| 8 |
- trading
|
|
|
|
| 9 |
pipeline_tag: text-classification
|
| 10 |
---
|
| 11 |
|
|
@@ -20,17 +21,19 @@ Fine-tuned FinBERT model for financial sentiment analysis in TradePulse.
|
|
| 20 |
|
| 21 |
## Performance
|
| 22 |
|
| 23 |
-
*Last training: 2025-07-24
|
| 24 |
-
*Dataset: `news_20250724.csv` (
|
| 25 |
|
| 26 |
| Metric | Value |
|
| 27 |
|--------|-------|
|
| 28 |
-
| Loss | 0.
|
| 29 |
-
| Accuracy |
|
| 30 |
-
| F1 Score |
|
| 31 |
-
|
| 32 |
-
|
|
| 33 |
-
|
|
|
|
|
|
|
| 34 |
|
| 35 |
## Training Details
|
| 36 |
|
|
@@ -41,10 +44,12 @@ Fine-tuned FinBERT model for financial sentiment analysis in TradePulse.
|
|
| 41 |
- **Batch Size**: 4
|
| 42 |
- **Class Balancing**: None
|
| 43 |
|
|
|
|
| 44 |
## Usage
|
| 45 |
|
| 46 |
```python
|
| 47 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
|
|
|
| 48 |
|
| 49 |
tokenizer = AutoTokenizer.from_pretrained("Bencode92/tradepulse-finbert-sentiment")
|
| 50 |
model = AutoModelForSequenceClassification.from_pretrained("Bencode92/tradepulse-finbert-sentiment")
|
|
@@ -53,10 +58,11 @@ model = AutoModelForSequenceClassification.from_pretrained("Bencode92/tradepulse
|
|
| 53 |
text = "Apple reported strong quarterly earnings beating expectations"
|
| 54 |
inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True)
|
| 55 |
outputs = model(**inputs)
|
|
|
|
| 56 |
predictions = outputs.logits.softmax(dim=-1)
|
| 57 |
```
|
| 58 |
|
| 59 |
## Model Card Authors
|
| 60 |
|
| 61 |
- TradePulse ML Team
|
| 62 |
-
- Auto-generated on 2025-07-24
|
|
|
|
| 6 |
- sentiment-analysis
|
| 7 |
- finbert
|
| 8 |
- trading
|
| 9 |
+
|
| 10 |
pipeline_tag: text-classification
|
| 11 |
---
|
| 12 |
|
|
|
|
| 21 |
|
| 22 |
## Performance
|
| 23 |
|
| 24 |
+
*Last training: 2025-07-24 15:44*
|
| 25 |
+
*Dataset: `news_20250724.csv` (171 samples)*
|
| 26 |
|
| 27 |
| Metric | Value |
|
| 28 |
|--------|-------|
|
| 29 |
+
| Loss | 0.2699 |
|
| 30 |
+
| Accuracy | 0.9535 |
|
| 31 |
+
| F1 Score | 0.9535 |
|
| 32 |
+
|
| 33 |
+
| F1 Macro | 0.9535 |
|
| 34 |
+
|
| 35 |
+
| Precision | 0.9579 |
|
| 36 |
+
| Recall | 0.9535 |
|
| 37 |
|
| 38 |
## Training Details
|
| 39 |
|
|
|
|
| 44 |
- **Batch Size**: 4
|
| 45 |
- **Class Balancing**: None
|
| 46 |
|
| 47 |
+
|
| 48 |
## Usage
|
| 49 |
|
| 50 |
```python
|
| 51 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
| 52 |
+
import torch
|
| 53 |
|
| 54 |
tokenizer = AutoTokenizer.from_pretrained("Bencode92/tradepulse-finbert-sentiment")
|
| 55 |
model = AutoModelForSequenceClassification.from_pretrained("Bencode92/tradepulse-finbert-sentiment")
|
|
|
|
| 58 |
text = "Apple reported strong quarterly earnings beating expectations"
|
| 59 |
inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True)
|
| 60 |
outputs = model(**inputs)
|
| 61 |
+
|
| 62 |
predictions = outputs.logits.softmax(dim=-1)
|
| 63 |
```
|
| 64 |
|
| 65 |
## Model Card Authors
|
| 66 |
|
| 67 |
- TradePulse ML Team
|
| 68 |
+
- Auto-generated on 2025-07-24 15:44:30
|
{checkpoint-5 β checkpoint-32}/config.json
RENAMED
|
File without changes
|
{checkpoint-5 β checkpoint-32}/model.safetensors
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 439039996
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0c790fd6e8d95e12a74a0f232817351992936b6aa478caf158acf325cd08cf87
|
| 3 |
size 439039996
|
{checkpoint-5 β checkpoint-32}/special_tokens_map.json
RENAMED
|
File without changes
|
{checkpoint-5 β checkpoint-32}/tokenizer.json
RENAMED
|
File without changes
|
{checkpoint-5 β checkpoint-32}/tokenizer_config.json
RENAMED
|
File without changes
|
{checkpoint-5 β checkpoint-32}/trainer_state.json
RENAMED
|
@@ -1,31 +1,52 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric":
|
| 3 |
-
"best_model_checkpoint": "hf-sentiment-production/checkpoint-
|
| 4 |
"epoch": 1.0,
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
| 10 |
"log_history": [
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
{
|
| 12 |
"epoch": 1.0,
|
| 13 |
-
"eval_accuracy":
|
| 14 |
-
"eval_f1":
|
| 15 |
-
"eval_f1_macro":
|
| 16 |
-
"eval_loss": 0.
|
| 17 |
-
"eval_precision":
|
| 18 |
-
"eval_precision_macro":
|
| 19 |
-
"eval_recall":
|
| 20 |
-
"eval_recall_macro":
|
| 21 |
-
"eval_runtime":
|
| 22 |
-
"eval_samples_per_second": 1.
|
| 23 |
-
"eval_steps_per_second": 0.
|
| 24 |
-
"step":
|
| 25 |
}
|
| 26 |
],
|
| 27 |
"logging_steps": 10,
|
| 28 |
-
"max_steps":
|
| 29 |
"num_input_tokens_seen": 0,
|
| 30 |
"num_train_epochs": 2,
|
| 31 |
"save_steps": 500,
|
|
@@ -50,7 +71,7 @@
|
|
| 50 |
"attributes": {}
|
| 51 |
}
|
| 52 |
},
|
| 53 |
-
"total_flos":
|
| 54 |
"train_batch_size": 4,
|
| 55 |
"trial_name": null,
|
| 56 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.9534883720930233,
|
| 3 |
+
"best_model_checkpoint": "hf-sentiment-production/checkpoint-32",
|
| 4 |
"epoch": 1.0,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 32,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
| 10 |
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.3125,
|
| 13 |
+
"grad_norm": 20.869625091552734,
|
| 14 |
+
"learning_rate": 3.125e-06,
|
| 15 |
+
"loss": 0.2232,
|
| 16 |
+
"step": 10
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"epoch": 0.625,
|
| 20 |
+
"grad_norm": 24.07443618774414,
|
| 21 |
+
"learning_rate": 6.25e-06,
|
| 22 |
+
"loss": 0.142,
|
| 23 |
+
"step": 20
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"epoch": 0.9375,
|
| 27 |
+
"grad_norm": 5.75204610824585,
|
| 28 |
+
"learning_rate": 9.375000000000001e-06,
|
| 29 |
+
"loss": 0.203,
|
| 30 |
+
"step": 30
|
| 31 |
+
},
|
| 32 |
{
|
| 33 |
"epoch": 1.0,
|
| 34 |
+
"eval_accuracy": 0.9534883720930233,
|
| 35 |
+
"eval_f1": 0.9534883720930233,
|
| 36 |
+
"eval_f1_macro": 0.9666666666666667,
|
| 37 |
+
"eval_loss": 0.26990270614624023,
|
| 38 |
+
"eval_precision": 0.9579180509413068,
|
| 39 |
+
"eval_precision_macro": 0.9682539682539683,
|
| 40 |
+
"eval_recall": 0.9534883720930233,
|
| 41 |
+
"eval_recall_macro": 0.9682539682539683,
|
| 42 |
+
"eval_runtime": 23.8102,
|
| 43 |
+
"eval_samples_per_second": 1.806,
|
| 44 |
+
"eval_steps_per_second": 0.462,
|
| 45 |
+
"step": 32
|
| 46 |
}
|
| 47 |
],
|
| 48 |
"logging_steps": 10,
|
| 49 |
+
"max_steps": 64,
|
| 50 |
"num_input_tokens_seen": 0,
|
| 51 |
"num_train_epochs": 2,
|
| 52 |
"save_steps": 500,
|
|
|
|
| 71 |
"attributes": {}
|
| 72 |
}
|
| 73 |
},
|
| 74 |
+
"total_flos": 33678517469184.0,
|
| 75 |
"train_batch_size": 4,
|
| 76 |
"trial_name": null,
|
| 77 |
"trial_params": null
|
{checkpoint-5 β checkpoint-32}/vocab.txt
RENAMED
|
File without changes
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 439039996
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0c790fd6e8d95e12a74a0f232817351992936b6aa478caf158acf325cd08cf87
|
| 3 |
size 439039996
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5521
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0797cb80249f45aa53e3f7e5498774d0cda687d153caa26fca11f969aecb203d
|
| 3 |
size 5521
|