KotshinZ commited on
Commit
e1fc26c
·
verified ·
1 Parent(s): 9cba334

End of training

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  base_model: openai-community/gpt2
3
- datasets: HuggingFaceH4/Bespoke-Stratos-17k
4
  library_name: transformers
5
  model_name: gpt2-RMT-2
6
  tags:
@@ -11,7 +11,7 @@ licence: license
11
 
12
  # Model Card for gpt2-RMT-2
13
 
14
- This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on the [HuggingFaceH4/Bespoke-Stratos-17k](https://huggingface.co/datasets/HuggingFaceH4/Bespoke-Stratos-17k) dataset.
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/s18574s18574-/huggingface/runs/mbk8mzkf)
31
 
32
 
33
  This model was trained with SFT.
 
1
  ---
2
  base_model: openai-community/gpt2
3
+ datasets: HuggingFaceFW/fineweb-edu
4
  library_name: transformers
5
  model_name: gpt2-RMT-2
6
  tags:
 
11
 
12
  # Model Card for gpt2-RMT-2
13
 
14
+ This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on the [HuggingFaceFW/fineweb-edu](https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu) dataset.
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/s18574s18574-/huggingface/runs/t9renl1y)
31
 
32
 
33
  This model was trained with SFT.
all_results.json CHANGED
@@ -1,12 +1,12 @@
1
  {
2
- "eval_runtime": 2.7326,
3
  "eval_samples": 100,
4
- "eval_samples_per_second": 208.955,
5
- "eval_steps_per_second": 52.33,
6
  "total_flos": 0.0,
7
- "train_loss": 3.0202497909698995,
8
- "train_runtime": 60.0371,
9
- "train_samples": 16610,
10
- "train_samples_per_second": 79.507,
11
- "train_steps_per_second": 4.98
12
  }
 
1
  {
2
+ "eval_runtime": 0.3164,
3
  "eval_samples": 100,
4
+ "eval_samples_per_second": 151.686,
5
+ "eval_steps_per_second": 9.48,
6
  "total_flos": 0.0,
7
+ "train_loss": 5.556107954545454,
8
+ "train_runtime": 94.197,
9
+ "train_samples": 19883,
10
+ "train_samples_per_second": 11.222,
11
+ "train_steps_per_second": 0.35
12
  }
base_model/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "attn_pdrop": 0.1,
7
+ "bos_token_id": 50256,
8
+ "embd_pdrop": 0.1,
9
+ "eos_token_id": 50256,
10
+ "initializer_range": 0.02,
11
+ "layer_norm_epsilon": 1e-05,
12
+ "model_type": "gpt2",
13
+ "n_ctx": 1024,
14
+ "n_embd": 768,
15
+ "n_head": 12,
16
+ "n_inner": null,
17
+ "n_layer": 12,
18
+ "n_positions": 1024,
19
+ "reorder_and_upcast_attn": false,
20
+ "resid_pdrop": 0.1,
21
+ "scale_attn_by_inverse_layer_idx": false,
22
+ "scale_attn_weights": true,
23
+ "summary_activation": null,
24
+ "summary_first_dropout": 0.1,
25
+ "summary_proj_to_labels": true,
26
+ "summary_type": "cls_index",
27
+ "summary_use_proj": true,
28
+ "task_specific_params": {
29
+ "text-generation": {
30
+ "do_sample": true,
31
+ "max_length": 50
32
+ }
33
+ },
34
+ "torch_dtype": "bfloat16",
35
+ "transformers_version": "4.50.0.dev0",
36
+ "use_cache": false,
37
+ "vocab_size": 50257
38
+ }
base_model/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "transformers_version": "4.50.0.dev0"
6
+ }
base_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d237600614a23ab38e6e8626c1a3495f6abbc5bc35856fee8d2c293889dbd3f
3
+ size 248912768
config.json CHANGED
@@ -34,6 +34,6 @@
34
  },
35
  "torch_dtype": "bfloat16",
36
  "transformers_version": "4.50.0.dev0",
37
- "use_cache": true,
38
  "vocab_size": 50257
39
  }
 
34
  },
35
  "torch_dtype": "bfloat16",
36
  "transformers_version": "4.50.0.dev0",
37
+ "use_cache": false,
38
  "vocab_size": 50257
39
  }
eval_results.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "eval_runtime": 2.7326,
3
  "eval_samples": 100,
4
- "eval_samples_per_second": 208.955,
5
- "eval_steps_per_second": 52.33
6
  }
 
1
  {
2
+ "eval_runtime": 0.3164,
3
  "eval_samples": 100,
4
+ "eval_samples_per_second": 151.686,
5
+ "eval_steps_per_second": 9.48
6
  }
generation_config.json CHANGED
@@ -1,4 +1,6 @@
1
  {
2
  "_from_model_config": true,
 
 
3
  "transformers_version": "4.50.0.dev0"
4
  }
 
1
  {
2
  "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
  "transformers_version": "4.50.0.dev0"
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0b1cef2e205503013144a7bcad7ad852c288d9e54cc44a8aaf644d9d71ec6cbc
3
- size 326110968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddafb83cf1a59157af80d355275bd09239a37d622ab84d9c4df9fd1bb357cbce
3
+ size 326089656
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e07ea7caa7f567049c6ac949b18c3c222e52a15c44b05ed9bc048bb7eda52472
3
  size 248944286
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8e45365da2e62bd9b60b592a1fdd35a62ba9baf557b86a6caafeefd31fc75d5
3
  size 248944286
tokenizer_config.json CHANGED
@@ -11,7 +11,6 @@
11
  }
12
  },
13
  "bos_token": "<|endoftext|>",
14
- "chat_template": "{% for message in messages %}{{ message['role'] }}: {{ message['content'] }}\n{% endfor %}",
15
  "clean_up_tokenization_spaces": false,
16
  "eos_token": "<|endoftext|>",
17
  "extra_special_tokens": {},
 
11
  }
12
  },
13
  "bos_token": "<|endoftext|>",
 
14
  "clean_up_tokenization_spaces": false,
15
  "eos_token": "<|endoftext|>",
16
  "extra_special_tokens": {},
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 3.0202497909698995,
4
- "train_runtime": 60.0371,
5
- "train_samples": 16610,
6
- "train_samples_per_second": 79.507,
7
- "train_steps_per_second": 4.98
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 5.556107954545454,
4
+ "train_runtime": 94.197,
5
+ "train_samples": 19883,
6
+ "train_samples_per_second": 11.222,
7
+ "train_steps_per_second": 0.35
8
  }
trainer_state.json CHANGED
@@ -1,272 +1,50 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.05010893246187364,
5
  "eval_steps": 100,
6
- "global_step": 299,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.0016758840288252053,
13
- "grad_norm": 36.404178619384766,
14
- "learning_rate": 6.666666666666667e-06,
15
- "loss": 5.6828,
16
- "mean_token_accuracy": 0.27888479083776474,
17
  "step": 10
18
  },
19
  {
20
- "epoch": 0.0033517680576504107,
21
- "grad_norm": 42.79513931274414,
22
- "learning_rate": 1.3333333333333333e-05,
23
- "loss": 5.425,
24
- "mean_token_accuracy": 0.2772253334522247,
25
  "step": 20
26
  },
27
  {
28
- "epoch": 0.005027652086475616,
29
- "grad_norm": 17.52666664123535,
30
- "learning_rate": 2e-05,
31
- "loss": 4.6359,
32
- "mean_token_accuracy": 0.2905010163784027,
33
  "step": 30
34
  },
35
  {
36
- "epoch": 0.006703536115300821,
37
- "grad_norm": 39.22002410888672,
38
- "learning_rate": 1.9931880491250263e-05,
39
- "loss": 4.1836,
40
- "mean_token_accuracy": 0.29999364018440244,
41
- "step": 40
42
- },
43
- {
44
- "epoch": 0.008379420144126026,
45
- "grad_norm": 6.146772861480713,
46
- "learning_rate": 1.9728450018495506e-05,
47
- "loss": 3.9766,
48
- "mean_token_accuracy": 0.30167217552661896,
49
- "step": 50
50
- },
51
- {
52
- "epoch": 0.010055304172951232,
53
- "grad_norm": 6.285459995269775,
54
- "learning_rate": 1.9392480098509488e-05,
55
- "loss": 3.7109,
56
- "mean_token_accuracy": 0.32380467653274536,
57
- "step": 60
58
- },
59
- {
60
- "epoch": 0.011731188201776438,
61
- "grad_norm": 5.058553218841553,
62
- "learning_rate": 1.8928547952473037e-05,
63
- "loss": 3.5086,
64
- "mean_token_accuracy": 0.345212359726429,
65
- "step": 70
66
- },
67
- {
68
- "epoch": 0.013407072230601643,
69
- "grad_norm": 15.02955150604248,
70
- "learning_rate": 1.8342974146362397e-05,
71
- "loss": 3.3859,
72
- "mean_token_accuracy": 0.3557349964976311,
73
- "step": 80
74
- },
75
- {
76
- "epoch": 0.015082956259426848,
77
- "grad_norm": 4.372200965881348,
78
- "learning_rate": 1.7643736480179353e-05,
79
- "loss": 3.2352,
80
- "mean_token_accuracy": 0.3736902341246605,
81
- "step": 90
82
- },
83
- {
84
- "epoch": 0.01675884028825205,
85
- "grad_norm": 3.283221960067749,
86
- "learning_rate": 1.684036129918786e-05,
87
- "loss": 3.0875,
88
- "mean_token_accuracy": 0.3915310278534889,
89
- "step": 100
90
- },
91
- {
92
- "epoch": 0.01675884028825205,
93
- "eval_runtime": 2.7416,
94
- "eval_samples_per_second": 208.276,
95
- "eval_steps_per_second": 52.16,
96
- "step": 100
97
- },
98
- {
99
- "epoch": 0.01843472431707726,
100
- "grad_norm": 17.207765579223633,
101
- "learning_rate": 1.5943793707922086e-05,
102
- "loss": 2.9906,
103
- "mean_token_accuracy": 0.40075025260448455,
104
- "step": 110
105
- },
106
- {
107
- "epoch": 0.020110608345902465,
108
- "grad_norm": 3.305183172225952,
109
- "learning_rate": 1.4966248455157622e-05,
110
- "loss": 2.8484,
111
- "mean_token_accuracy": 0.42432604134082796,
112
- "step": 120
113
- },
114
- {
115
- "epoch": 0.02178649237472767,
116
- "grad_norm": 3.639101266860962,
117
- "learning_rate": 1.392104352137426e-05,
118
- "loss": 2.7719,
119
- "mean_token_accuracy": 0.4330811306834221,
120
- "step": 130
121
- },
122
- {
123
- "epoch": 0.023462376403552875,
124
- "grad_norm": 2.6281790733337402,
125
- "learning_rate": 1.2822418675898428e-05,
126
- "loss": 2.6055,
127
- "mean_token_accuracy": 0.46078331768512726,
128
- "step": 140
129
- },
130
- {
131
- "epoch": 0.02513826043237808,
132
- "grad_norm": 2.1558072566986084,
133
- "learning_rate": 1.1685341475684935e-05,
134
- "loss": 2.5195,
135
- "mean_token_accuracy": 0.4750635787844658,
136
- "step": 150
137
- },
138
- {
139
- "epoch": 0.026814144461203285,
140
- "grad_norm": 3.083050489425659,
141
- "learning_rate": 1.0525303348791599e-05,
142
- "loss": 2.5078,
143
- "mean_token_accuracy": 0.4781091034412384,
144
- "step": 160
145
- },
146
- {
147
- "epoch": 0.02849002849002849,
148
- "grad_norm": 1.9051162004470825,
149
- "learning_rate": 9.358108540685406e-06,
150
- "loss": 2.4273,
151
- "mean_token_accuracy": 0.4916709065437317,
152
- "step": 170
153
- },
154
- {
155
- "epoch": 0.030165912518853696,
156
- "grad_norm": 1.8084359169006348,
157
- "learning_rate": 8.199658798755048e-06,
158
- "loss": 2.4664,
159
- "mean_token_accuracy": 0.4778802111744881,
160
- "step": 180
161
- },
162
- {
163
- "epoch": 0.0318417965476789,
164
- "grad_norm": 1.7160115242004395,
165
- "learning_rate": 7.065736728466832e-06,
166
- "loss": 2.3969,
167
- "mean_token_accuracy": 0.49128942042589185,
168
- "step": 190
169
- },
170
- {
171
- "epoch": 0.0335176805765041,
172
- "grad_norm": 1.6504725217819214,
173
- "learning_rate": 5.971790772698467e-06,
174
- "loss": 2.4383,
175
- "mean_token_accuracy": 0.4894074246287346,
176
- "step": 200
177
- },
178
- {
179
- "epoch": 0.0335176805765041,
180
- "eval_runtime": 2.7088,
181
- "eval_samples_per_second": 210.793,
182
- "eval_steps_per_second": 52.79,
183
- "step": 200
184
- },
185
- {
186
- "epoch": 0.03519356460532931,
187
- "grad_norm": 1.851637363433838,
188
- "learning_rate": 4.932724743671089e-06,
189
- "loss": 2.3555,
190
- "mean_token_accuracy": 0.5041772618889808,
191
- "step": 210
192
- },
193
- {
194
- "epoch": 0.03686944863415452,
195
- "grad_norm": 1.8771989345550537,
196
- "learning_rate": 3.96269477487588e-06,
197
- "loss": 2.3977,
198
- "mean_token_accuracy": 0.4926373302936554,
199
- "step": 220
200
- },
201
- {
202
- "epoch": 0.03854533266297972,
203
- "grad_norm": 4.372796535491943,
204
- "learning_rate": 3.074916459302211e-06,
205
- "loss": 2.3508,
206
- "mean_token_accuracy": 0.5035096645355225,
207
- "step": 230
208
- },
209
- {
210
- "epoch": 0.04022121669180493,
211
- "grad_norm": 1.8549838066101074,
212
- "learning_rate": 2.281484801497186e-06,
213
- "loss": 2.4258,
214
- "mean_token_accuracy": 0.48886062651872636,
215
- "step": 240
216
- },
217
- {
218
- "epoch": 0.04189710072063013,
219
- "grad_norm": 1.7765941619873047,
220
- "learning_rate": 1.5932094364120453e-06,
221
- "loss": 2.3656,
222
- "mean_token_accuracy": 0.4967192322015762,
223
- "step": 250
224
- },
225
- {
226
- "epoch": 0.04357298474945534,
227
- "grad_norm": 1.8978796005249023,
228
- "learning_rate": 1.0194673599976134e-06,
229
- "loss": 2.3398,
230
- "mean_token_accuracy": 0.5044379487633706,
231
- "step": 260
232
- },
233
- {
234
- "epoch": 0.04524886877828054,
235
- "grad_norm": 2.0636632442474365,
236
- "learning_rate": 5.680751779327742e-07,
237
- "loss": 2.3687,
238
- "mean_token_accuracy": 0.49782553166151045,
239
- "step": 270
240
- },
241
- {
242
- "epoch": 0.04692475280710575,
243
- "grad_norm": 1.694811224937439,
244
- "learning_rate": 2.4518261295667255e-07,
245
- "loss": 2.3867,
246
- "mean_token_accuracy": 0.4927390620112419,
247
- "step": 280
248
- },
249
- {
250
- "epoch": 0.04860063683593095,
251
- "grad_norm": 2.2234456539154053,
252
- "learning_rate": 5.518872165033329e-08,
253
- "loss": 2.3953,
254
- "mean_token_accuracy": 0.49295524060726165,
255
- "step": 290
256
- },
257
- {
258
- "epoch": 0.05010893246187364,
259
- "mean_token_accuracy": 0.4976404458284378,
260
- "step": 299,
261
  "total_flos": 0.0,
262
- "train_loss": 3.0202497909698995,
263
- "train_runtime": 60.0371,
264
- "train_samples_per_second": 79.507,
265
- "train_steps_per_second": 4.98
266
  }
267
  ],
268
  "logging_steps": 10,
269
- "max_steps": 299,
270
  "num_input_tokens_seen": 0,
271
  "num_train_epochs": 1,
272
  "save_steps": 500,
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.09984871406959153,
5
  "eval_steps": 100,
6
+ "global_step": 33,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.030257186081694403,
13
+ "grad_norm": 18.575438958798006,
14
+ "learning_rate": 1.796093065705644e-05,
15
+ "loss": 6.9031,
16
+ "mean_token_accuracy": 0.1433999717235565,
17
  "step": 10
18
  },
19
  {
20
+ "epoch": 0.060514372163388806,
21
+ "grad_norm": 11.22312925711848,
22
+ "learning_rate": 8.382180034472353e-06,
23
+ "loss": 5.325,
24
+ "mean_token_accuracy": 0.19182548448443412,
25
  "step": 20
26
  },
27
  {
28
+ "epoch": 0.0907715582450832,
29
+ "grad_norm": 7.8614989055585625,
30
+ "learning_rate": 5.234682881719766e-07,
31
+ "loss": 4.7195,
32
+ "mean_token_accuracy": 0.23640244454145432,
33
  "step": 30
34
  },
35
  {
36
+ "epoch": 0.09984871406959153,
37
+ "mean_token_accuracy": 0.24069985498984656,
38
+ "step": 33,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
39
  "total_flos": 0.0,
40
+ "train_loss": 5.556107954545454,
41
+ "train_runtime": 94.197,
42
+ "train_samples_per_second": 11.222,
43
+ "train_steps_per_second": 0.35
44
  }
45
  ],
46
  "logging_steps": 10,
47
+ "max_steps": 33,
48
  "num_input_tokens_seen": 0,
49
  "num_train_epochs": 1,
50
  "save_steps": 500,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bdd0e83665f82cbff34945a1dc5f700141602e90d62707ec1ecd8225c96d8e93
3
  size 7352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:868f332cb0455cb6b6311de1634eb90272c59e27c7f7a526d31c36d9ac33732f
3
  size 7352