KotshinZ commited on
Commit
6998ab2
·
verified ·
1 Parent(s): 54cb710

Model save

Browse files
README.md CHANGED
@@ -1,11 +1,9 @@
1
  ---
2
  base_model: openai-community/gpt2
3
- datasets: HuggingFaceFW/fineweb-edu
4
  library_name: transformers
5
  model_name: gpt2-RMT-2
6
  tags:
7
  - generated_from_trainer
8
- - open-r1
9
  - trl
10
  - sft
11
  licence: license
@@ -13,7 +11,7 @@ licence: license
13
 
14
  # Model Card for gpt2-RMT-2
15
 
16
- This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on the [HuggingFaceFW/fineweb-edu](https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu) dataset.
17
  It has been trained using [TRL](https://github.com/huggingface/trl).
18
 
19
  ## Quick start
@@ -29,7 +27,7 @@ print(output["generated_text"])
29
 
30
  ## Training procedure
31
 
32
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/s18574s18574-/huggingface/runs/1xooe0sm)
33
 
34
 
35
  This model was trained with SFT.
 
1
  ---
2
  base_model: openai-community/gpt2
 
3
  library_name: transformers
4
  model_name: gpt2-RMT-2
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
  - sft
9
  licence: license
 
11
 
12
  # Model Card for gpt2-RMT-2
13
 
14
+ This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/s18574s18574-/huggingface/runs/mh05rq2b)
31
 
32
 
33
  This model was trained with SFT.
all_results.json CHANGED
@@ -3,10 +3,10 @@
3
  "eval_samples": 100,
4
  "eval_samples_per_second": 192.26,
5
  "eval_steps_per_second": 12.016,
6
- "total_flos": 541940076511232.0,
7
- "train_loss": 5.546875,
8
- "train_runtime": 36.3584,
9
  "train_samples": 19883,
10
- "train_samples_per_second": 29.074,
11
- "train_steps_per_second": 0.908
12
  }
 
3
  "eval_samples": 100,
4
  "eval_samples_per_second": 192.26,
5
  "eval_steps_per_second": 12.016,
6
+ "total_flos": 5419400756723712.0,
7
+ "train_loss": 3.6152698863636363,
8
+ "train_runtime": 337.2618,
9
  "train_samples": 19883,
10
+ "train_samples_per_second": 31.344,
11
+ "train_steps_per_second": 0.978
12
  }
config.json CHANGED
@@ -33,6 +33,6 @@
33
  },
34
  "torch_dtype": "bfloat16",
35
  "transformers_version": "4.50.0.dev0",
36
- "use_cache": true,
37
  "vocab_size": 50257
38
  }
 
33
  },
34
  "torch_dtype": "bfloat16",
35
  "transformers_version": "4.50.0.dev0",
36
+ "use_cache": false,
37
  "vocab_size": 50257
38
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ef6b24bad9283be8a3270f3898cb6e1d27404fa045313443fde712995a1cabc
3
  size 248912768
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b16e3936aa2c91fc567c45311766ae7a01909050a6b9a4ce29225967ebbf4bc
3
  size 248912768
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "total_flos": 541940076511232.0,
3
- "train_loss": 5.546875,
4
- "train_runtime": 36.3584,
5
  "train_samples": 19883,
6
- "train_samples_per_second": 29.074,
7
- "train_steps_per_second": 0.908
8
  }
 
1
  {
2
+ "total_flos": 5419400756723712.0,
3
+ "train_loss": 3.6152698863636363,
4
+ "train_runtime": 337.2618,
5
  "train_samples": 19883,
6
+ "train_samples_per_second": 31.344,
7
+ "train_steps_per_second": 0.978
8
  }
trainer_state.json CHANGED
@@ -1,50 +1,310 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.09984871406959153,
5
  "eval_steps": 100,
6
- "global_step": 33,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.030257186081694403,
13
- "grad_norm": 18.3522235886847,
14
- "learning_rate": 1.796093065705644e-05,
15
- "loss": 6.8937,
16
- "mean_token_accuracy": 0.1434693343937397,
17
  "step": 10
18
  },
19
  {
20
  "epoch": 0.060514372163388806,
21
- "grad_norm": 10.614870433655174,
22
- "learning_rate": 8.382180034472353e-06,
23
- "loss": 5.3195,
24
- "mean_token_accuracy": 0.1919048972427845,
25
  "step": 20
26
  },
27
  {
28
  "epoch": 0.0907715582450832,
29
- "grad_norm": 7.571339027683935,
30
- "learning_rate": 5.234682881719766e-07,
31
- "loss": 4.707,
32
- "mean_token_accuracy": 0.23758598268032075,
33
  "step": 30
34
  },
35
  {
36
- "epoch": 0.09984871406959153,
37
- "mean_token_accuracy": 0.24127203722794852,
38
- "step": 33,
39
- "total_flos": 541940076511232.0,
40
- "train_loss": 5.546875,
41
- "train_runtime": 36.3584,
42
- "train_samples_per_second": 29.074,
43
- "train_steps_per_second": 0.908
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
44
  }
45
  ],
46
  "logging_steps": 10,
47
- "max_steps": 33,
48
  "num_input_tokens_seen": 0,
49
  "num_train_epochs": 1,
50
  "save_steps": 500,
@@ -60,7 +320,7 @@
60
  "attributes": {}
61
  }
62
  },
63
- "total_flos": 541940076511232.0,
64
  "train_batch_size": 8,
65
  "trial_name": null,
66
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.9984871406959153,
5
  "eval_steps": 100,
6
+ "global_step": 330,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.030257186081694403,
13
+ "grad_norm": 51.5472786539882,
14
+ "learning_rate": 6.060606060606061e-06,
15
+ "loss": 7.3992,
16
+ "mean_token_accuracy": 0.13563802018761634,
17
  "step": 10
18
  },
19
  {
20
  "epoch": 0.060514372163388806,
21
+ "grad_norm": 27.69300301659005,
22
+ "learning_rate": 1.2121212121212122e-05,
23
+ "loss": 6.8242,
24
+ "mean_token_accuracy": 0.14449115768074988,
25
  "step": 20
26
  },
27
  {
28
  "epoch": 0.0907715582450832,
29
+ "grad_norm": 9.743705978929961,
30
+ "learning_rate": 1.8181818181818182e-05,
31
+ "loss": 5.6266,
32
+ "mean_token_accuracy": 0.1745789811015129,
33
  "step": 30
34
  },
35
  {
36
+ "epoch": 0.12102874432677761,
37
+ "grad_norm": 4.50812560927562,
38
+ "learning_rate": 1.9972599751485225e-05,
39
+ "loss": 4.6586,
40
+ "mean_token_accuracy": 0.24351915791630746,
41
+ "step": 40
42
+ },
43
+ {
44
+ "epoch": 0.15128593040847202,
45
+ "grad_norm": 3.1996689209062676,
46
+ "learning_rate": 1.9838755799290993e-05,
47
+ "loss": 4.0187,
48
+ "mean_token_accuracy": 0.3023824915289879,
49
+ "step": 50
50
+ },
51
+ {
52
+ "epoch": 0.1815431164901664,
53
+ "grad_norm": 1.8824084218377581,
54
+ "learning_rate": 1.9594929736144978e-05,
55
+ "loss": 3.6316,
56
+ "mean_token_accuracy": 0.3426647365093231,
57
+ "step": 60
58
+ },
59
+ {
60
+ "epoch": 0.2118003025718608,
61
+ "grad_norm": 1.1187763043308236,
62
+ "learning_rate": 1.9243847161266924e-05,
63
+ "loss": 3.3883,
64
+ "mean_token_accuracy": 0.3721240028738976,
65
+ "step": 70
66
+ },
67
+ {
68
+ "epoch": 0.24205748865355523,
69
+ "grad_norm": 1.0093619517882768,
70
+ "learning_rate": 1.8789432636206197e-05,
71
+ "loss": 3.3281,
72
+ "mean_token_accuracy": 0.3791012689471245,
73
+ "step": 80
74
+ },
75
+ {
76
+ "epoch": 0.2723146747352496,
77
+ "grad_norm": 0.8420416221184239,
78
+ "learning_rate": 1.8236765814298328e-05,
79
+ "loss": 3.2738,
80
+ "mean_token_accuracy": 0.3867510572075844,
81
+ "step": 90
82
+ },
83
+ {
84
+ "epoch": 0.30257186081694404,
85
+ "grad_norm": 0.8952752980575562,
86
+ "learning_rate": 1.7592024657977432e-05,
87
+ "loss": 3.2816,
88
+ "mean_token_accuracy": 0.3850349023938179,
89
+ "step": 100
90
+ },
91
+ {
92
+ "epoch": 0.30257186081694404,
93
+ "eval_runtime": 0.2249,
94
+ "eval_samples_per_second": 213.425,
95
+ "eval_steps_per_second": 13.339,
96
+ "step": 100
97
+ },
98
+ {
99
+ "epoch": 0.3328290468986384,
100
+ "grad_norm": 0.8146530137770093,
101
+ "learning_rate": 1.686241637868734e-05,
102
+ "loss": 3.2516,
103
+ "mean_token_accuracy": 0.3902030408382416,
104
+ "step": 110
105
+ },
106
+ {
107
+ "epoch": 0.3630862329803328,
108
+ "grad_norm": 0.7245362247896451,
109
+ "learning_rate": 1.6056096871376667e-05,
110
+ "loss": 3.2777,
111
+ "mean_token_accuracy": 0.3867401584982872,
112
+ "step": 120
113
+ },
114
+ {
115
+ "epoch": 0.39334341906202724,
116
+ "grad_norm": 0.7046871234836425,
117
+ "learning_rate": 1.5182079544175957e-05,
118
+ "loss": 3.25,
119
+ "mean_token_accuracy": 0.39042258113622663,
120
+ "step": 130
121
+ },
122
+ {
123
+ "epoch": 0.4236006051437216,
124
+ "grad_norm": 0.7889558903633703,
125
+ "learning_rate": 1.4250134562400301e-05,
126
+ "loss": 3.2254,
127
+ "mean_token_accuracy": 0.3929138645529747,
128
+ "step": 140
129
+ },
130
+ {
131
+ "epoch": 0.45385779122541603,
132
+ "grad_norm": 0.7526789075759445,
133
+ "learning_rate": 1.3270679633174219e-05,
134
+ "loss": 3.2168,
135
+ "mean_token_accuracy": 0.3945690140128136,
136
+ "step": 150
137
+ },
138
+ {
139
+ "epoch": 0.48411497730711045,
140
+ "grad_norm": 0.7037431923979744,
141
+ "learning_rate": 1.2254663551538047e-05,
142
+ "loss": 3.1898,
143
+ "mean_token_accuracy": 0.39688222110271454,
144
+ "step": 160
145
+ },
146
+ {
147
+ "epoch": 0.5143721633888049,
148
+ "grad_norm": 0.7269697505833069,
149
+ "learning_rate": 1.121344380981082e-05,
150
+ "loss": 3.2254,
151
+ "mean_token_accuracy": 0.39344793409109113,
152
+ "step": 170
153
+ },
154
+ {
155
+ "epoch": 0.5446293494704992,
156
+ "grad_norm": 0.71564273988106,
157
+ "learning_rate": 1.015865963834808e-05,
158
+ "loss": 3.198,
159
+ "mean_token_accuracy": 0.3971615001559258,
160
+ "step": 180
161
+ },
162
+ {
163
+ "epoch": 0.5748865355521936,
164
+ "grad_norm": 0.7272400173576803,
165
+ "learning_rate": 9.102101896903084e-06,
166
+ "loss": 3.2098,
167
+ "mean_token_accuracy": 0.3965013727545738,
168
+ "step": 190
169
+ },
170
+ {
171
+ "epoch": 0.6051437216338881,
172
+ "grad_norm": 0.7204839560593628,
173
+ "learning_rate": 8.055581271005292e-06,
174
+ "loss": 3.209,
175
+ "mean_token_accuracy": 0.394715978205204,
176
+ "step": 200
177
+ },
178
+ {
179
+ "epoch": 0.6051437216338881,
180
+ "eval_runtime": 0.2231,
181
+ "eval_samples_per_second": 215.144,
182
+ "eval_steps_per_second": 13.447,
183
+ "step": 200
184
+ },
185
+ {
186
+ "epoch": 0.6354009077155824,
187
+ "grad_norm": 0.6620216126492257,
188
+ "learning_rate": 7.0307962467172555e-06,
189
+ "loss": 3.1988,
190
+ "mean_token_accuracy": 0.39781390130519867,
191
+ "step": 210
192
+ },
193
+ {
194
+ "epoch": 0.6656580937972768,
195
+ "grad_norm": 0.7813459496424706,
196
+ "learning_rate": 6.039202339608432e-06,
197
+ "loss": 3.202,
198
+ "mean_token_accuracy": 0.3953568756580353,
199
+ "step": 220
200
+ },
201
+ {
202
+ "epoch": 0.6959152798789713,
203
+ "grad_norm": 0.6556046091666947,
204
+ "learning_rate": 5.091884039764321e-06,
205
+ "loss": 3.1969,
206
+ "mean_token_accuracy": 0.3961416259407997,
207
+ "step": 230
208
+ },
209
+ {
210
+ "epoch": 0.7261724659606656,
211
+ "grad_norm": 0.7243524981690901,
212
+ "learning_rate": 4.19943090428802e-06,
213
+ "loss": 3.2102,
214
+ "mean_token_accuracy": 0.3935989648103714,
215
+ "step": 240
216
+ },
217
+ {
218
+ "epoch": 0.75642965204236,
219
+ "grad_norm": 0.7423682208710032,
220
+ "learning_rate": 3.37181918238904e-06,
221
+ "loss": 3.2078,
222
+ "mean_token_accuracy": 0.39539735913276675,
223
+ "step": 250
224
+ },
225
+ {
226
+ "epoch": 0.7866868381240545,
227
+ "grad_norm": 0.684794452564992,
228
+ "learning_rate": 2.618300296308135e-06,
229
+ "loss": 3.2219,
230
+ "mean_token_accuracy": 0.3939197212457657,
231
+ "step": 260
232
+ },
233
+ {
234
+ "epoch": 0.8169440242057489,
235
+ "grad_norm": 0.6857522826014064,
236
+ "learning_rate": 1.947297424689414e-06,
237
+ "loss": 3.1992,
238
+ "mean_token_accuracy": 0.39573368430137634,
239
+ "step": 270
240
+ },
241
+ {
242
+ "epoch": 0.8472012102874432,
243
+ "grad_norm": 0.6966190236263456,
244
+ "learning_rate": 1.3663113444380905e-06,
245
+ "loss": 3.2031,
246
+ "mean_token_accuracy": 0.3933654069900513,
247
+ "step": 280
248
+ },
249
+ {
250
+ "epoch": 0.8774583963691377,
251
+ "grad_norm": 0.8226214667450941,
252
+ "learning_rate": 8.818365836066101e-07,
253
+ "loss": 3.1621,
254
+ "mean_token_accuracy": 0.4012774467468262,
255
+ "step": 290
256
+ },
257
+ {
258
+ "epoch": 0.9077155824508321,
259
+ "grad_norm": 0.6811554548251312,
260
+ "learning_rate": 4.992888225905467e-07,
261
+ "loss": 3.2254,
262
+ "mean_token_accuracy": 0.39436970353126527,
263
+ "step": 300
264
+ },
265
+ {
266
+ "epoch": 0.9077155824508321,
267
+ "eval_runtime": 0.223,
268
+ "eval_samples_per_second": 215.211,
269
+ "eval_steps_per_second": 13.451,
270
+ "step": 300
271
+ },
272
+ {
273
+ "epoch": 0.9379727685325264,
274
+ "grad_norm": 0.6793742171946109,
275
+ "learning_rate": 2.2294435517691504e-07,
276
+ "loss": 3.1926,
277
+ "mean_token_accuracy": 0.397060464322567,
278
+ "step": 310
279
+ },
280
+ {
281
+ "epoch": 0.9682299546142209,
282
+ "grad_norm": 0.6744479467043031,
283
+ "learning_rate": 5.5892286176932875e-08,
284
+ "loss": 3.1961,
285
+ "mean_token_accuracy": 0.3960541605949402,
286
+ "step": 320
287
+ },
288
+ {
289
+ "epoch": 0.9984871406959153,
290
+ "grad_norm": 0.6621754518709608,
291
+ "learning_rate": 0.0,
292
+ "loss": 3.2035,
293
+ "mean_token_accuracy": 0.3947854354977608,
294
+ "step": 330
295
+ },
296
+ {
297
+ "epoch": 0.9984871406959153,
298
+ "step": 330,
299
+ "total_flos": 5419400756723712.0,
300
+ "train_loss": 3.6152698863636363,
301
+ "train_runtime": 337.2618,
302
+ "train_samples_per_second": 31.344,
303
+ "train_steps_per_second": 0.978
304
  }
305
  ],
306
  "logging_steps": 10,
307
+ "max_steps": 330,
308
  "num_input_tokens_seen": 0,
309
  "num_train_epochs": 1,
310
  "save_steps": 500,
 
320
  "attributes": {}
321
  }
322
  },
323
+ "total_flos": 5419400756723712.0,
324
  "train_batch_size": 8,
325
  "trial_name": null,
326
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:baf06d605ca2ac794fb0cb09d28d4e25826d2c1b68fdaaefca458cb0a917d847
3
  size 7352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dba648ef7d2addc48fb53ca1ab6b041442778aef087f173aef4c1a8ab1b68d9
3
  size 7352