Bencode92 commited on
Commit
c88a3a8
Β·
1 Parent(s): eb869f4

πŸ”„ Incremental correlations | F1Β΅: 0.000, Hamming: 0.968

Browse files
README.md CHANGED
@@ -21,12 +21,12 @@ Fine-tuned FinBERT model for financial correlations analysis in TradePulse.
21
 
22
  ## Performance
23
 
24
- *Last training: 2025-07-25 16:03*
25
- *Dataset: `news_20250725.csv` (224 samples)*
26
 
27
  | Metric | Value |
28
  |--------|-------|
29
- | Loss | 0.3343 |
30
  | Subset Accuracy | 1.0000 |
31
  | F1 Score | 0.0000 |
32
  | F1 Micro | 0.0000 |
@@ -65,4 +65,4 @@ predictions = torch.sigmoid(outputs.logits).squeeze() > 0.5
65
  ## Model Card Authors
66
 
67
  - TradePulse ML Team
68
- - Auto-generated on 2025-07-25 16:03:24
 
21
 
22
  ## Performance
23
 
24
+ *Last training: 2025-07-26 18:16*
25
+ *Dataset: `news_20250726.csv` (206 samples)*
26
 
27
  | Metric | Value |
28
  |--------|-------|
29
+ | Loss | 0.2760 |
30
  | Subset Accuracy | 1.0000 |
31
  | F1 Score | 0.0000 |
32
  | F1 Micro | 0.0000 |
 
65
  ## Model Card Authors
66
 
67
  - TradePulse ML Team
68
+ - Auto-generated on 2025-07-26 18:16:11
{checkpoint-56 β†’ checkpoint-52}/config.json RENAMED
File without changes
{checkpoint-56 β†’ checkpoint-52}/model.safetensors RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:613f427274186184eabd58541b382754192e1137a7b717dedca28555cdf0b949
3
  size 439415268
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdbbb53017891b99599c43bd7a67e43ff4688fc51dba6967be2894c90a11ca96
3
  size 439415268
{checkpoint-56 β†’ checkpoint-52}/special_tokens_map.json RENAMED
File without changes
{checkpoint-56 β†’ checkpoint-52}/tokenizer.json RENAMED
File without changes
{checkpoint-56 β†’ checkpoint-52}/tokenizer_config.json RENAMED
File without changes
{checkpoint-56 β†’ checkpoint-52}/trainer_state.json RENAMED
@@ -1,46 +1,46 @@
1
  {
2
  "best_metric": 0.0,
3
- "best_model_checkpoint": "hf-correlations-production/checkpoint-56",
4
  "epoch": 1.0,
5
  "eval_steps": 500,
6
- "global_step": 56,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.17857142857142858,
13
- "grad_norm": 0.6497295498847961,
14
- "learning_rate": 1.8181818181818183e-06,
15
- "loss": 0.4226,
16
  "step": 10
17
  },
18
  {
19
- "epoch": 0.35714285714285715,
20
- "grad_norm": 0.634713351726532,
21
- "learning_rate": 3.6363636363636366e-06,
22
- "loss": 0.4162,
23
  "step": 20
24
  },
25
  {
26
- "epoch": 0.5357142857142857,
27
- "grad_norm": 0.6155627369880676,
28
- "learning_rate": 5.4545454545454545e-06,
29
- "loss": 0.4035,
30
  "step": 30
31
  },
32
  {
33
- "epoch": 0.7142857142857143,
34
- "grad_norm": 0.6066541075706482,
35
- "learning_rate": 7.272727272727273e-06,
36
- "loss": 0.3882,
37
  "step": 40
38
  },
39
  {
40
- "epoch": 0.8928571428571429,
41
- "grad_norm": 0.5941498875617981,
42
- "learning_rate": 9.090909090909091e-06,
43
- "loss": 0.3675,
44
  "step": 50
45
  },
46
  {
@@ -51,22 +51,22 @@
51
  "eval_f1_micro": 0.0,
52
  "eval_hamming_loss": 0.0,
53
  "eval_hamming_score": 1.0,
54
- "eval_loss": 0.3342878520488739,
55
  "eval_precision": 0.0,
56
  "eval_precision_macro": 0.0,
57
  "eval_precision_micro": 0.0,
58
  "eval_recall": 0.0,
59
  "eval_recall_macro": 0.0,
60
  "eval_recall_micro": 0.0,
61
- "eval_runtime": 0.5973,
62
- "eval_samples_per_second": 1.674,
63
- "eval_steps_per_second": 1.674,
64
  "eval_subset_accuracy": 1.0,
65
- "step": 56
66
  }
67
  ],
68
  "logging_steps": 10,
69
- "max_steps": 112,
70
  "num_input_tokens_seen": 0,
71
  "num_train_epochs": 2,
72
  "save_steps": 500,
@@ -91,7 +91,7 @@
91
  "attributes": {}
92
  }
93
  },
94
- "total_flos": 58738562737152.0,
95
  "train_batch_size": 4,
96
  "trial_name": null,
97
  "trial_params": null
 
1
  {
2
  "best_metric": 0.0,
3
+ "best_model_checkpoint": "hf-correlations-production/checkpoint-52",
4
  "epoch": 1.0,
5
  "eval_steps": 500,
6
+ "global_step": 52,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.19230769230769232,
13
+ "grad_norm": 0.5568640828132629,
14
+ "learning_rate": 1.96078431372549e-06,
15
+ "loss": 0.3432,
16
  "step": 10
17
  },
18
  {
19
+ "epoch": 0.38461538461538464,
20
+ "grad_norm": 0.5578851103782654,
21
+ "learning_rate": 3.92156862745098e-06,
22
+ "loss": 0.3362,
23
  "step": 20
24
  },
25
  {
26
+ "epoch": 0.5769230769230769,
27
+ "grad_norm": 0.5405699610710144,
28
+ "learning_rate": 5.882352941176471e-06,
29
+ "loss": 0.328,
30
  "step": 30
31
  },
32
  {
33
+ "epoch": 0.7692307692307693,
34
+ "grad_norm": 0.5361686944961548,
35
+ "learning_rate": 7.84313725490196e-06,
36
+ "loss": 0.3155,
37
  "step": 40
38
  },
39
  {
40
+ "epoch": 0.9615384615384616,
41
+ "grad_norm": 0.5128586292266846,
42
+ "learning_rate": 9.803921568627451e-06,
43
+ "loss": 0.2977,
44
  "step": 50
45
  },
46
  {
 
51
  "eval_f1_micro": 0.0,
52
  "eval_hamming_loss": 0.0,
53
  "eval_hamming_score": 1.0,
54
+ "eval_loss": 0.2760178744792938,
55
  "eval_precision": 0.0,
56
  "eval_precision_macro": 0.0,
57
  "eval_precision_micro": 0.0,
58
  "eval_recall": 0.0,
59
  "eval_recall_macro": 0.0,
60
  "eval_recall_micro": 0.0,
61
+ "eval_runtime": 0.5983,
62
+ "eval_samples_per_second": 1.671,
63
+ "eval_steps_per_second": 1.671,
64
  "eval_subset_accuracy": 1.0,
65
+ "step": 52
66
  }
67
  ],
68
  "logging_steps": 10,
69
+ "max_steps": 104,
70
  "num_input_tokens_seen": 0,
71
  "num_train_epochs": 2,
72
  "save_steps": 500,
 
91
  "attributes": {}
92
  }
93
  },
94
+ "total_flos": 53997333457920.0,
95
  "train_batch_size": 4,
96
  "trial_name": null,
97
  "trial_params": null
{checkpoint-56 β†’ checkpoint-52}/vocab.txt RENAMED
File without changes
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:613f427274186184eabd58541b382754192e1137a7b717dedca28555cdf0b949
3
  size 439415268
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdbbb53017891b99599c43bd7a67e43ff4688fc51dba6967be2894c90a11ca96
3
  size 439415268