Hartunka commited on
Commit
8309710
·
verified ·
1 Parent(s): 3a30e97

End of training

Browse files
README.md CHANGED
@@ -1,12 +1,27 @@
1
  ---
 
 
2
  base_model: Hartunka/tiny_bert_rand_100_v1
3
  tags:
4
  - generated_from_trainer
 
 
5
  metrics:
6
  - spearmanr
7
  model-index:
8
  - name: tiny_bert_rand_100_v1_stsb
9
- results: []
 
 
 
 
 
 
 
 
 
 
 
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -14,12 +29,12 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # tiny_bert_rand_100_v1_stsb
16
 
17
- This model is a fine-tuned version of [Hartunka/tiny_bert_rand_100_v1](https://huggingface.co/Hartunka/tiny_bert_rand_100_v1) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 2.8235
20
- - Pearson: 0.2652
21
- - Spearmanr: 0.2718
22
- - Combined Score: 0.2685
23
 
24
  ## Model description
25
 
 
1
  ---
2
+ language:
3
+ - en
4
  base_model: Hartunka/tiny_bert_rand_100_v1
5
  tags:
6
  - generated_from_trainer
7
+ datasets:
8
+ - glue
9
  metrics:
10
  - spearmanr
11
  model-index:
12
  - name: tiny_bert_rand_100_v1_stsb
13
+ results:
14
+ - task:
15
+ name: Text Classification
16
+ type: text-classification
17
+ dataset:
18
+ name: GLUE STSB
19
+ type: glue
20
+ args: stsb
21
+ metrics:
22
+ - name: Spearmanr
23
+ type: spearmanr
24
+ value: 0.2763743043991294
25
  ---
26
 
27
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
29
 
30
  # tiny_bert_rand_100_v1_stsb
31
 
32
+ This model is a fine-tuned version of [Hartunka/tiny_bert_rand_100_v1](https://huggingface.co/Hartunka/tiny_bert_rand_100_v1) on the GLUE STSB dataset.
33
  It achieves the following results on the evaluation set:
34
+ - Loss: 2.2768
35
+ - Pearson: 0.2798
36
+ - Spearmanr: 0.2764
37
+ - Combined Score: 0.2781
38
 
39
  ## Model description
40
 
all_results.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 10.0,
3
+ "eval_combined_score": 0.2780937290708646,
4
+ "eval_loss": 2.276757001876831,
5
+ "eval_pearson": 0.2798131537425998,
6
+ "eval_runtime": 0.5581,
7
+ "eval_samples": 1500,
8
+ "eval_samples_per_second": 2687.74,
9
+ "eval_spearmanr": 0.2763743043991294,
10
+ "eval_steps_per_second": 10.751,
11
+ "total_flos": 1507548205378560.0,
12
+ "train_loss": 1.4585458755493164,
13
+ "train_runtime": 51.1618,
14
+ "train_samples": 5749,
15
+ "train_samples_per_second": 5618.449,
16
+ "train_steps_per_second": 22.478
17
+ }
eval_results.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 10.0,
3
+ "eval_combined_score": 0.2780937290708646,
4
+ "eval_loss": 2.276757001876831,
5
+ "eval_pearson": 0.2798131537425998,
6
+ "eval_runtime": 0.5581,
7
+ "eval_samples": 1500,
8
+ "eval_samples_per_second": 2687.74,
9
+ "eval_spearmanr": 0.2763743043991294,
10
+ "eval_steps_per_second": 10.751
11
+ }
logs/events.out.tfevents.1744833239.s_004_m ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:489279eba01a3ca2931d0cdc9fbfd589a74dccb2447c546b1f39656d03e26ca6
3
+ size 40
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 10.0,
3
+ "total_flos": 1507548205378560.0,
4
+ "train_loss": 1.4585458755493164,
5
+ "train_runtime": 51.1618,
6
+ "train_samples": 5749,
7
+ "train_samples_per_second": 5618.449,
8
+ "train_steps_per_second": 22.478
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 2.276757001876831,
3
+ "best_model_checkpoint": "tiny_bert_rand_100_v1_stsb/checkpoint-115",
4
+ "epoch": 10.0,
5
+ "eval_steps": 500,
6
+ "global_step": 230,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "grad_norm": 4.961604595184326,
14
+ "learning_rate": 4.9e-05,
15
+ "loss": 3.4483,
16
+ "step": 23
17
+ },
18
+ {
19
+ "epoch": 1.0,
20
+ "eval_combined_score": 0.15705636987694221,
21
+ "eval_loss": 2.314561128616333,
22
+ "eval_pearson": 0.16774334792257545,
23
+ "eval_runtime": 0.5294,
24
+ "eval_samples_per_second": 2833.424,
25
+ "eval_spearmanr": 0.14636939183130895,
26
+ "eval_steps_per_second": 11.334,
27
+ "step": 23
28
+ },
29
+ {
30
+ "epoch": 2.0,
31
+ "grad_norm": 6.913020610809326,
32
+ "learning_rate": 4.8e-05,
33
+ "loss": 2.0255,
34
+ "step": 46
35
+ },
36
+ {
37
+ "epoch": 2.0,
38
+ "eval_combined_score": 0.1126483992024648,
39
+ "eval_loss": 2.5450494289398193,
40
+ "eval_pearson": 0.11683712415863884,
41
+ "eval_runtime": 0.5808,
42
+ "eval_samples_per_second": 2582.785,
43
+ "eval_spearmanr": 0.10845967424629079,
44
+ "eval_steps_per_second": 10.331,
45
+ "step": 46
46
+ },
47
+ {
48
+ "epoch": 3.0,
49
+ "grad_norm": 7.117321491241455,
50
+ "learning_rate": 4.7e-05,
51
+ "loss": 1.8523,
52
+ "step": 69
53
+ },
54
+ {
55
+ "epoch": 3.0,
56
+ "eval_combined_score": 0.21422209468092734,
57
+ "eval_loss": 2.3148019313812256,
58
+ "eval_pearson": 0.2201949812893677,
59
+ "eval_runtime": 0.6512,
60
+ "eval_samples_per_second": 2303.455,
61
+ "eval_spearmanr": 0.20824920807248695,
62
+ "eval_steps_per_second": 9.214,
63
+ "step": 69
64
+ },
65
+ {
66
+ "epoch": 4.0,
67
+ "grad_norm": 6.040245532989502,
68
+ "learning_rate": 4.600000000000001e-05,
69
+ "loss": 1.6156,
70
+ "step": 92
71
+ },
72
+ {
73
+ "epoch": 4.0,
74
+ "eval_combined_score": 0.2691090690581144,
75
+ "eval_loss": 2.3426806926727295,
76
+ "eval_pearson": 0.27034978222621336,
77
+ "eval_runtime": 0.548,
78
+ "eval_samples_per_second": 2737.269,
79
+ "eval_spearmanr": 0.26786835589001545,
80
+ "eval_steps_per_second": 10.949,
81
+ "step": 92
82
+ },
83
+ {
84
+ "epoch": 5.0,
85
+ "grad_norm": 9.559501647949219,
86
+ "learning_rate": 4.5e-05,
87
+ "loss": 1.3454,
88
+ "step": 115
89
+ },
90
+ {
91
+ "epoch": 5.0,
92
+ "eval_combined_score": 0.2780937290708646,
93
+ "eval_loss": 2.276757001876831,
94
+ "eval_pearson": 0.2798131537425998,
95
+ "eval_runtime": 0.5306,
96
+ "eval_samples_per_second": 2827.107,
97
+ "eval_spearmanr": 0.2763743043991294,
98
+ "eval_steps_per_second": 11.308,
99
+ "step": 115
100
+ },
101
+ {
102
+ "epoch": 6.0,
103
+ "grad_norm": 7.285312652587891,
104
+ "learning_rate": 4.4000000000000006e-05,
105
+ "loss": 1.1616,
106
+ "step": 138
107
+ },
108
+ {
109
+ "epoch": 6.0,
110
+ "eval_combined_score": 0.273431489839639,
111
+ "eval_loss": 2.6383583545684814,
112
+ "eval_pearson": 0.26860800836532117,
113
+ "eval_runtime": 0.5919,
114
+ "eval_samples_per_second": 2534.01,
115
+ "eval_spearmanr": 0.2782549713139568,
116
+ "eval_steps_per_second": 10.136,
117
+ "step": 138
118
+ },
119
+ {
120
+ "epoch": 7.0,
121
+ "grad_norm": 10.173330307006836,
122
+ "learning_rate": 4.3e-05,
123
+ "loss": 0.9734,
124
+ "step": 161
125
+ },
126
+ {
127
+ "epoch": 7.0,
128
+ "eval_combined_score": 0.2831437139838975,
129
+ "eval_loss": 2.477155923843384,
130
+ "eval_pearson": 0.28229446668730734,
131
+ "eval_runtime": 0.5315,
132
+ "eval_samples_per_second": 2822.41,
133
+ "eval_spearmanr": 0.2839929612804877,
134
+ "eval_steps_per_second": 11.29,
135
+ "step": 161
136
+ },
137
+ {
138
+ "epoch": 8.0,
139
+ "grad_norm": 9.35107707977295,
140
+ "learning_rate": 4.2e-05,
141
+ "loss": 0.8406,
142
+ "step": 184
143
+ },
144
+ {
145
+ "epoch": 8.0,
146
+ "eval_combined_score": 0.24872085472419414,
147
+ "eval_loss": 2.8825998306274414,
148
+ "eval_pearson": 0.2434565004022157,
149
+ "eval_runtime": 0.5698,
150
+ "eval_samples_per_second": 2632.351,
151
+ "eval_spearmanr": 0.25398520904617256,
152
+ "eval_steps_per_second": 10.529,
153
+ "step": 184
154
+ },
155
+ {
156
+ "epoch": 9.0,
157
+ "grad_norm": 17.66458511352539,
158
+ "learning_rate": 4.1e-05,
159
+ "loss": 0.7077,
160
+ "step": 207
161
+ },
162
+ {
163
+ "epoch": 9.0,
164
+ "eval_combined_score": 0.24927091603402246,
165
+ "eval_loss": 2.909149169921875,
166
+ "eval_pearson": 0.2461040064609805,
167
+ "eval_runtime": 0.5293,
168
+ "eval_samples_per_second": 2834.011,
169
+ "eval_spearmanr": 0.25243782560706446,
170
+ "eval_steps_per_second": 11.336,
171
+ "step": 207
172
+ },
173
+ {
174
+ "epoch": 10.0,
175
+ "grad_norm": 11.106776237487793,
176
+ "learning_rate": 4e-05,
177
+ "loss": 0.6149,
178
+ "step": 230
179
+ },
180
+ {
181
+ "epoch": 10.0,
182
+ "eval_combined_score": 0.2685087682776196,
183
+ "eval_loss": 2.823472023010254,
184
+ "eval_pearson": 0.26516905911909416,
185
+ "eval_runtime": 0.5621,
186
+ "eval_samples_per_second": 2668.438,
187
+ "eval_spearmanr": 0.2718484774361451,
188
+ "eval_steps_per_second": 10.674,
189
+ "step": 230
190
+ },
191
+ {
192
+ "epoch": 10.0,
193
+ "step": 230,
194
+ "total_flos": 1507548205378560.0,
195
+ "train_loss": 1.4585458755493164,
196
+ "train_runtime": 51.1618,
197
+ "train_samples_per_second": 5618.449,
198
+ "train_steps_per_second": 22.478
199
+ }
200
+ ],
201
+ "logging_steps": 1,
202
+ "max_steps": 1150,
203
+ "num_input_tokens_seen": 0,
204
+ "num_train_epochs": 50,
205
+ "save_steps": 500,
206
+ "total_flos": 1507548205378560.0,
207
+ "train_batch_size": 256,
208
+ "trial_name": null,
209
+ "trial_params": null
210
+ }