Bencode92 commited on
Commit
a3d3208
Β·
1 Parent(s): fa4b395

πŸ”„ Incremental importance | Acc: 1.000, F1: 1.000

Browse files
README.md CHANGED
@@ -6,6 +6,7 @@ tags:
6
  - sentiment-analysis
7
  - finbert
8
  - trading
 
9
  pipeline_tag: text-classification
10
  ---
11
 
@@ -20,17 +21,19 @@ Fine-tuned FinBERT model for financial importance analysis in TradePulse.
20
 
21
  ## Performance
22
 
23
- *Last training: 2025-07-24 14:48*
24
- *Dataset: `news_20250724.csv` (27 samples)*
25
 
26
  | Metric | Value |
27
  |--------|-------|
28
- | Loss | 0.3668 |
29
- | Accuracy | 0.8571 |
30
- | F1 Score | 0.8508 |
31
- | F1 Macro | 0.8508 |
32
- | Precision | 0.8857 |
33
- | Recall | 0.8571 |
 
 
34
 
35
  ## Training Details
36
 
@@ -41,10 +44,12 @@ Fine-tuned FinBERT model for financial importance analysis in TradePulse.
41
  - **Batch Size**: 4
42
  - **Class Balancing**: None
43
 
 
44
  ## Usage
45
 
46
  ```python
47
  from transformers import AutoTokenizer, AutoModelForSequenceClassification
 
48
 
49
  tokenizer = AutoTokenizer.from_pretrained("Bencode92/tradepulse-finbert-importance")
50
  model = AutoModelForSequenceClassification.from_pretrained("Bencode92/tradepulse-finbert-importance")
@@ -53,10 +58,11 @@ model = AutoModelForSequenceClassification.from_pretrained("Bencode92/tradepulse
53
  text = "Apple reported strong quarterly earnings beating expectations"
54
  inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True)
55
  outputs = model(**inputs)
 
56
  predictions = outputs.logits.softmax(dim=-1)
57
  ```
58
 
59
  ## Model Card Authors
60
 
61
  - TradePulse ML Team
62
- - Auto-generated on 2025-07-24 14:48:46
 
6
  - sentiment-analysis
7
  - finbert
8
  - trading
9
+
10
  pipeline_tag: text-classification
11
  ---
12
 
 
21
 
22
  ## Performance
23
 
24
+ *Last training: 2025-07-24 16:00*
25
+ *Dataset: `news_20250724.csv` (190 samples)*
26
 
27
  | Metric | Value |
28
  |--------|-------|
29
+ | Loss | 0.1901 |
30
+ | Accuracy | 1.0000 |
31
+ | F1 Score | 1.0000 |
32
+
33
+ | F1 Macro | 1.0000 |
34
+
35
+ | Precision | 1.0000 |
36
+ | Recall | 1.0000 |
37
 
38
  ## Training Details
39
 
 
44
  - **Batch Size**: 4
45
  - **Class Balancing**: None
46
 
47
+
48
  ## Usage
49
 
50
  ```python
51
  from transformers import AutoTokenizer, AutoModelForSequenceClassification
52
+ import torch
53
 
54
  tokenizer = AutoTokenizer.from_pretrained("Bencode92/tradepulse-finbert-importance")
55
  model = AutoModelForSequenceClassification.from_pretrained("Bencode92/tradepulse-finbert-importance")
 
58
  text = "Apple reported strong quarterly earnings beating expectations"
59
  inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True)
60
  outputs = model(**inputs)
61
+
62
  predictions = outputs.logits.softmax(dim=-1)
63
  ```
64
 
65
  ## Model Card Authors
66
 
67
  - TradePulse ML Team
68
+ - Auto-generated on 2025-07-24 16:00:13
{checkpoint-5 β†’ checkpoint-48}/config.json RENAMED
File without changes
{checkpoint-5 β†’ checkpoint-48}/model.safetensors RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b414d5bce5d60abc2176d405792d4cba1579262f941401da44fcf89168e31c4a
3
  size 439039996
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a21169b75795694e2eade74c7631920af5a7199ac0a38451ea6675feafcbd2c
3
  size 439039996
{checkpoint-5 β†’ checkpoint-48}/special_tokens_map.json RENAMED
File without changes
{checkpoint-5 β†’ checkpoint-48}/tokenizer.json RENAMED
File without changes
{checkpoint-5 β†’ checkpoint-48}/tokenizer_config.json RENAMED
File without changes
{checkpoint-5 β†’ checkpoint-48}/trainer_state.json RENAMED
@@ -1,31 +1,59 @@
1
  {
2
- "best_metric": 0.8507936507936508,
3
- "best_model_checkpoint": "hf-importance-production/checkpoint-5",
4
  "epoch": 1.0,
5
  "eval_steps": 500,
6
- "global_step": 5,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  {
12
  "epoch": 1.0,
13
- "eval_accuracy": 0.8571428571428571,
14
- "eval_f1": 0.8507936507936508,
15
- "eval_f1_macro": 0.8444444444444444,
16
- "eval_loss": 0.3668130338191986,
17
- "eval_precision": 0.8857142857142858,
18
- "eval_precision_macro": 0.9,
19
- "eval_recall": 0.8571428571428571,
20
- "eval_recall_macro": 0.8333333333333333,
21
- "eval_runtime": 3.9487,
22
- "eval_samples_per_second": 1.773,
23
- "eval_steps_per_second": 0.507,
24
- "step": 5
25
  }
26
  ],
27
  "logging_steps": 10,
28
- "max_steps": 10,
29
  "num_input_tokens_seen": 0,
30
  "num_train_epochs": 2,
31
  "save_steps": 500,
@@ -50,7 +78,7 @@
50
  "attributes": {}
51
  }
52
  },
53
- "total_flos": 5262268354560.0,
54
  "train_batch_size": 4,
55
  "trial_name": null,
56
  "trial_params": null
 
1
  {
2
+ "best_metric": 1.0,
3
+ "best_model_checkpoint": "hf-importance-production/checkpoint-48",
4
  "epoch": 1.0,
5
  "eval_steps": 500,
6
+ "global_step": 48,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
+ {
12
+ "epoch": 0.20833333333333334,
13
+ "grad_norm": 29.484235763549805,
14
+ "learning_rate": 2.1276595744680853e-06,
15
+ "loss": 0.139,
16
+ "step": 10
17
+ },
18
+ {
19
+ "epoch": 0.4166666666666667,
20
+ "grad_norm": 0.19761385023593903,
21
+ "learning_rate": 4.255319148936171e-06,
22
+ "loss": 0.0851,
23
+ "step": 20
24
+ },
25
+ {
26
+ "epoch": 0.625,
27
+ "grad_norm": 75.20068359375,
28
+ "learning_rate": 6.382978723404256e-06,
29
+ "loss": 0.2067,
30
+ "step": 30
31
+ },
32
+ {
33
+ "epoch": 0.8333333333333334,
34
+ "grad_norm": 0.0050959475338459015,
35
+ "learning_rate": 8.510638297872341e-06,
36
+ "loss": 0.2117,
37
+ "step": 40
38
+ },
39
  {
40
  "epoch": 1.0,
41
+ "eval_accuracy": 1.0,
42
+ "eval_f1": 1.0,
43
+ "eval_f1_macro": 1.0,
44
+ "eval_loss": 0.19008873403072357,
45
+ "eval_precision": 1.0,
46
+ "eval_precision_macro": 1.0,
47
+ "eval_recall": 1.0,
48
+ "eval_recall_macro": 1.0,
49
+ "eval_runtime": 0.5835,
50
+ "eval_samples_per_second": 1.714,
51
+ "eval_steps_per_second": 1.714,
52
+ "step": 48
53
  }
54
  ],
55
  "logging_steps": 10,
56
+ "max_steps": 96,
57
  "num_input_tokens_seen": 0,
58
  "num_train_epochs": 2,
59
  "save_steps": 500,
 
78
  "attributes": {}
79
  }
80
  },
81
+ "total_flos": 49728435950592.0,
82
  "train_batch_size": 4,
83
  "trial_name": null,
84
  "trial_params": null
{checkpoint-5 β†’ checkpoint-48}/vocab.txt RENAMED
File without changes
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b414d5bce5d60abc2176d405792d4cba1579262f941401da44fcf89168e31c4a
3
  size 439039996
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a21169b75795694e2eade74c7631920af5a7199ac0a38451ea6675feafcbd2c
3
  size 439039996
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe388da48683bddd7c46856d680946afceeef503e46feec65f61458ffe1ef4f0
3
  size 5521
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d80d1e81c1a0fc229e86f25920b3a900302cd322f503a1e512596b0bb8d9a54
3
  size 5521