Hartunka commited on
Commit
ff3026a
·
verified ·
1 Parent(s): 196bf54

End of training

Browse files
README.md CHANGED
@@ -1,13 +1,28 @@
1
  ---
2
  library_name: transformers
 
 
3
  base_model: Hartunka/bert_base_km_20_v1
4
  tags:
5
  - generated_from_trainer
 
 
6
  metrics:
7
  - spearmanr
8
  model-index:
9
  - name: bert_base_km_20_v1_stsb
10
- results: []
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -15,12 +30,12 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # bert_base_km_20_v1_stsb
17
 
18
- This model is a fine-tuned version of [Hartunka/bert_base_km_20_v1](https://huggingface.co/Hartunka/bert_base_km_20_v1) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 2.4996
21
- - Pearson: 0.3085
22
- - Spearmanr: 0.3151
23
- - Combined Score: 0.3118
24
 
25
  ## Model description
26
 
 
1
  ---
2
  library_name: transformers
3
+ language:
4
+ - en
5
  base_model: Hartunka/bert_base_km_20_v1
6
  tags:
7
  - generated_from_trainer
8
+ datasets:
9
+ - glue
10
  metrics:
11
  - spearmanr
12
  model-index:
13
  - name: bert_base_km_20_v1_stsb
14
+ results:
15
+ - task:
16
+ name: Text Classification
17
+ type: text-classification
18
+ dataset:
19
+ name: GLUE STSB
20
+ type: glue
21
+ args: stsb
22
+ metrics:
23
+ - name: Spearmanr
24
+ type: spearmanr
25
+ value: 0.2806848266082787
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
30
 
31
  # bert_base_km_20_v1_stsb
32
 
33
+ This model is a fine-tuned version of [Hartunka/bert_base_km_20_v1](https://huggingface.co/Hartunka/bert_base_km_20_v1) on the GLUE STSB dataset.
34
  It achieves the following results on the evaluation set:
35
+ - Loss: 2.1734
36
+ - Pearson: 0.2765
37
+ - Spearmanr: 0.2807
38
+ - Combined Score: 0.2786
39
 
40
  ## Model description
41
 
all_results.json CHANGED
@@ -3,15 +3,15 @@
3
  "eval_combined_score": 0.27859606936951653,
4
  "eval_loss": 2.1733975410461426,
5
  "eval_pearson": 0.27650731213075436,
6
- "eval_runtime": 0.9705,
7
  "eval_samples": 1500,
8
- "eval_samples_per_second": 1545.65,
9
  "eval_spearmanr": 0.2806848266082787,
10
- "eval_steps_per_second": 6.183,
11
  "total_flos": 6050447504044032.0,
12
  "train_loss": 1.3817211441371753,
13
- "train_runtime": 98.0352,
14
  "train_samples": 5749,
15
- "train_samples_per_second": 2932.112,
16
- "train_steps_per_second": 11.73
17
  }
 
3
  "eval_combined_score": 0.27859606936951653,
4
  "eval_loss": 2.1733975410461426,
5
  "eval_pearson": 0.27650731213075436,
6
+ "eval_runtime": 0.9765,
7
  "eval_samples": 1500,
8
+ "eval_samples_per_second": 1536.087,
9
  "eval_spearmanr": 0.2806848266082787,
10
+ "eval_steps_per_second": 6.144,
11
  "total_flos": 6050447504044032.0,
12
  "train_loss": 1.3817211441371753,
13
+ "train_runtime": 98.9888,
14
  "train_samples": 5749,
15
+ "train_samples_per_second": 2903.865,
16
+ "train_steps_per_second": 11.617
17
  }
eval_results.json CHANGED
@@ -3,9 +3,9 @@
3
  "eval_combined_score": 0.27859606936951653,
4
  "eval_loss": 2.1733975410461426,
5
  "eval_pearson": 0.27650731213075436,
6
- "eval_runtime": 0.9705,
7
  "eval_samples": 1500,
8
- "eval_samples_per_second": 1545.65,
9
  "eval_spearmanr": 0.2806848266082787,
10
- "eval_steps_per_second": 6.183
11
  }
 
3
  "eval_combined_score": 0.27859606936951653,
4
  "eval_loss": 2.1733975410461426,
5
  "eval_pearson": 0.27650731213075436,
6
+ "eval_runtime": 0.9765,
7
  "eval_samples": 1500,
8
+ "eval_samples_per_second": 1536.087,
9
  "eval_spearmanr": 0.2806848266082787,
10
+ "eval_steps_per_second": 6.144
11
  }
logs/events.out.tfevents.1745050757.s_005_m.2788007.13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba840a98d6ea8e5a3e41f017768645745eeb57affde3c13c7ae76c38fcee8310
3
+ size 473
train_results.json CHANGED
@@ -2,8 +2,8 @@
2
  "epoch": 8.0,
3
  "total_flos": 6050447504044032.0,
4
  "train_loss": 1.3817211441371753,
5
- "train_runtime": 98.0352,
6
  "train_samples": 5749,
7
- "train_samples_per_second": 2932.112,
8
- "train_steps_per_second": 11.73
9
  }
 
2
  "epoch": 8.0,
3
  "total_flos": 6050447504044032.0,
4
  "train_loss": 1.3817211441371753,
5
+ "train_runtime": 98.9888,
6
  "train_samples": 5749,
7
+ "train_samples_per_second": 2903.865,
8
+ "train_steps_per_second": 11.617
9
  }
trainer_state.json CHANGED
@@ -21,10 +21,10 @@
21
  "eval_combined_score": 0.21883840300992635,
22
  "eval_loss": 2.362926721572876,
23
  "eval_pearson": 0.22315559305765864,
24
- "eval_runtime": 0.9514,
25
- "eval_samples_per_second": 1576.685,
26
  "eval_spearmanr": 0.21452121296219404,
27
- "eval_steps_per_second": 6.307,
28
  "step": 23
29
  },
30
  {
@@ -39,10 +39,10 @@
39
  "eval_combined_score": 0.2250433116133167,
40
  "eval_loss": 2.223330020904541,
41
  "eval_pearson": 0.22643903431708493,
42
- "eval_runtime": 0.9766,
43
- "eval_samples_per_second": 1536.013,
44
  "eval_spearmanr": 0.2236475889095485,
45
- "eval_steps_per_second": 6.144,
46
  "step": 46
47
  },
48
  {
@@ -57,10 +57,10 @@
57
  "eval_combined_score": 0.27859606936951653,
58
  "eval_loss": 2.1733975410461426,
59
  "eval_pearson": 0.27650731213075436,
60
- "eval_runtime": 0.9499,
61
- "eval_samples_per_second": 1579.115,
62
  "eval_spearmanr": 0.2806848266082787,
63
- "eval_steps_per_second": 6.316,
64
  "step": 69
65
  },
66
  {
@@ -75,10 +75,10 @@
75
  "eval_combined_score": 0.2857232862338934,
76
  "eval_loss": 2.284653902053833,
77
  "eval_pearson": 0.28481215085198996,
78
- "eval_runtime": 0.9437,
79
- "eval_samples_per_second": 1589.456,
80
  "eval_spearmanr": 0.28663442161579694,
81
- "eval_steps_per_second": 6.358,
82
  "step": 92
83
  },
84
  {
@@ -93,10 +93,10 @@
93
  "eval_combined_score": 0.2801317348206342,
94
  "eval_loss": 2.755934238433838,
95
  "eval_pearson": 0.27054585694075417,
96
- "eval_runtime": 0.997,
97
- "eval_samples_per_second": 1504.452,
98
  "eval_spearmanr": 0.2897176127005142,
99
- "eval_steps_per_second": 6.018,
100
  "step": 115
101
  },
102
  {
@@ -111,10 +111,10 @@
111
  "eval_combined_score": 0.30899397408639784,
112
  "eval_loss": 2.474417209625244,
113
  "eval_pearson": 0.3009404135965045,
114
- "eval_runtime": 0.9611,
115
- "eval_samples_per_second": 1560.688,
116
  "eval_spearmanr": 0.31704753457629126,
117
- "eval_steps_per_second": 6.243,
118
  "step": 138
119
  },
120
  {
@@ -129,10 +129,10 @@
129
  "eval_combined_score": 0.2980624641456129,
130
  "eval_loss": 2.5624561309814453,
131
  "eval_pearson": 0.29500261904772385,
132
- "eval_runtime": 0.994,
133
- "eval_samples_per_second": 1508.988,
134
  "eval_spearmanr": 0.301122309243502,
135
- "eval_steps_per_second": 6.036,
136
  "step": 161
137
  },
138
  {
@@ -147,10 +147,10 @@
147
  "eval_combined_score": 0.31179634058913464,
148
  "eval_loss": 2.499624252319336,
149
  "eval_pearson": 0.3084881703414194,
150
- "eval_runtime": 0.9583,
151
- "eval_samples_per_second": 1565.335,
152
  "eval_spearmanr": 0.3151045108368498,
153
- "eval_steps_per_second": 6.261,
154
  "step": 184
155
  },
156
  {
@@ -158,9 +158,9 @@
158
  "step": 184,
159
  "total_flos": 6050447504044032.0,
160
  "train_loss": 1.3817211441371753,
161
- "train_runtime": 98.0352,
162
- "train_samples_per_second": 2932.112,
163
- "train_steps_per_second": 11.73
164
  }
165
  ],
166
  "logging_steps": 1,
 
21
  "eval_combined_score": 0.21883840300992635,
22
  "eval_loss": 2.362926721572876,
23
  "eval_pearson": 0.22315559305765864,
24
+ "eval_runtime": 0.9537,
25
+ "eval_samples_per_second": 1572.84,
26
  "eval_spearmanr": 0.21452121296219404,
27
+ "eval_steps_per_second": 6.291,
28
  "step": 23
29
  },
30
  {
 
39
  "eval_combined_score": 0.2250433116133167,
40
  "eval_loss": 2.223330020904541,
41
  "eval_pearson": 0.22643903431708493,
42
+ "eval_runtime": 0.9462,
43
+ "eval_samples_per_second": 1585.334,
44
  "eval_spearmanr": 0.2236475889095485,
45
+ "eval_steps_per_second": 6.341,
46
  "step": 46
47
  },
48
  {
 
57
  "eval_combined_score": 0.27859606936951653,
58
  "eval_loss": 2.1733975410461426,
59
  "eval_pearson": 0.27650731213075436,
60
+ "eval_runtime": 0.9646,
61
+ "eval_samples_per_second": 1555.115,
62
  "eval_spearmanr": 0.2806848266082787,
63
+ "eval_steps_per_second": 6.22,
64
  "step": 69
65
  },
66
  {
 
75
  "eval_combined_score": 0.2857232862338934,
76
  "eval_loss": 2.284653902053833,
77
  "eval_pearson": 0.28481215085198996,
78
+ "eval_runtime": 0.9772,
79
+ "eval_samples_per_second": 1534.935,
80
  "eval_spearmanr": 0.28663442161579694,
81
+ "eval_steps_per_second": 6.14,
82
  "step": 92
83
  },
84
  {
 
93
  "eval_combined_score": 0.2801317348206342,
94
  "eval_loss": 2.755934238433838,
95
  "eval_pearson": 0.27054585694075417,
96
+ "eval_runtime": 0.9542,
97
+ "eval_samples_per_second": 1572.032,
98
  "eval_spearmanr": 0.2897176127005142,
99
+ "eval_steps_per_second": 6.288,
100
  "step": 115
101
  },
102
  {
 
111
  "eval_combined_score": 0.30899397408639784,
112
  "eval_loss": 2.474417209625244,
113
  "eval_pearson": 0.3009404135965045,
114
+ "eval_runtime": 0.9638,
115
+ "eval_samples_per_second": 1556.332,
116
  "eval_spearmanr": 0.31704753457629126,
117
+ "eval_steps_per_second": 6.225,
118
  "step": 138
119
  },
120
  {
 
129
  "eval_combined_score": 0.2980624641456129,
130
  "eval_loss": 2.5624561309814453,
131
  "eval_pearson": 0.29500261904772385,
132
+ "eval_runtime": 0.9704,
133
+ "eval_samples_per_second": 1545.698,
134
  "eval_spearmanr": 0.301122309243502,
135
+ "eval_steps_per_second": 6.183,
136
  "step": 161
137
  },
138
  {
 
147
  "eval_combined_score": 0.31179634058913464,
148
  "eval_loss": 2.499624252319336,
149
  "eval_pearson": 0.3084881703414194,
150
+ "eval_runtime": 0.9545,
151
+ "eval_samples_per_second": 1571.47,
152
  "eval_spearmanr": 0.3151045108368498,
153
+ "eval_steps_per_second": 6.286,
154
  "step": 184
155
  },
156
  {
 
158
  "step": 184,
159
  "total_flos": 6050447504044032.0,
160
  "train_loss": 1.3817211441371753,
161
+ "train_runtime": 98.9888,
162
+ "train_samples_per_second": 2903.865,
163
+ "train_steps_per_second": 11.617
164
  }
165
  ],
166
  "logging_steps": 1,