MHGanainy commited on
Commit
1f3bca9
·
verified ·
1 Parent(s): 7518816

MHGanainy/gpt2-xl-lora-multi-7

Browse files
Files changed (5) hide show
  1. README.md +2 -0
  2. all_results.json +10 -10
  3. eval_results.json +5 -5
  4. train_results.json +5 -5
  5. trainer_state.json +334 -117
README.md CHANGED
@@ -15,6 +15,8 @@ should probably proofread and complete it, then remove this comment. -->
15
  # gpt2-xl-lora-multi-7
16
 
17
  This model is a fine-tuned version of [openai-community/gpt2-xl](https://huggingface.co/openai-community/gpt2-xl) on an unknown dataset.
 
 
18
 
19
  ## Model description
20
 
 
15
  # gpt2-xl-lora-multi-7
16
 
17
  This model is a fine-tuned version of [openai-community/gpt2-xl](https://huggingface.co/openai-community/gpt2-xl) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 2.5070
20
 
21
  ## Model description
22
 
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_loss": 2.5500738620758057,
4
- "eval_runtime": 333.0501,
5
- "eval_samples_per_second": 7.786,
6
- "eval_steps_per_second": 0.489,
7
- "perplexity": 12.808049776869597,
8
- "total_flos": 3.940214344788214e+17,
9
- "train_loss": 2.675769654604105,
10
- "train_runtime": 864.3634,
11
- "train_samples_per_second": 50.048,
12
- "train_steps_per_second": 3.128
13
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_loss": 2.506993293762207,
4
+ "eval_runtime": 354.5986,
5
+ "eval_samples_per_second": 8.503,
6
+ "eval_steps_per_second": 0.533,
7
+ "perplexity": 12.267988314860721,
8
+ "total_flos": 8.454557554683412e+17,
9
+ "train_loss": 2.7013893893405267,
10
+ "train_runtime": 1640.3942,
11
+ "train_samples_per_second": 56.586,
12
+ "train_steps_per_second": 3.537
13
  }
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_loss": 2.5500738620758057,
4
- "eval_runtime": 333.0501,
5
- "eval_samples_per_second": 7.786,
6
- "eval_steps_per_second": 0.489,
7
- "perplexity": 12.808049776869597
8
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_loss": 2.506993293762207,
4
+ "eval_runtime": 354.5986,
5
+ "eval_samples_per_second": 8.503,
6
+ "eval_steps_per_second": 0.533,
7
+ "perplexity": 12.267988314860721
8
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "total_flos": 3.940214344788214e+17,
4
- "train_loss": 2.675769654604105,
5
- "train_runtime": 864.3634,
6
- "train_samples_per_second": 50.048,
7
- "train_steps_per_second": 3.128
8
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "total_flos": 8.454557554683412e+17,
4
+ "train_loss": 2.7013893893405267,
5
+ "train_runtime": 1640.3942,
6
+ "train_samples_per_second": 56.586,
7
+ "train_steps_per_second": 3.537
8
  }
trainer_state.json CHANGED
@@ -3,212 +3,429 @@
3
  "best_model_checkpoint": null,
4
  "epoch": 1.0,
5
  "eval_steps": 500,
6
- "global_step": 2704,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.03698224852071006,
13
- "grad_norm": 0.14849229156970978,
14
- "learning_rate": 9.06148867313916e-07,
15
- "loss": 2.9299,
16
  "step": 100
17
  },
18
  {
19
- "epoch": 0.07396449704142012,
20
- "grad_norm": 0.18462003767490387,
21
- "learning_rate": 1.8307905686546466e-06,
22
- "loss": 2.9277,
23
  "step": 200
24
  },
25
  {
26
- "epoch": 0.11094674556213018,
27
- "grad_norm": 0.2244928628206253,
28
- "learning_rate": 2.7461858529819695e-06,
29
- "loss": 2.9367,
30
  "step": 300
31
  },
32
  {
33
- "epoch": 0.14792899408284024,
34
- "grad_norm": 0.11301285773515701,
35
- "learning_rate": 3.6708275543227e-06,
36
- "loss": 2.865,
37
  "step": 400
38
  },
39
  {
40
- "epoch": 0.1849112426035503,
41
- "grad_norm": 0.16168534755706787,
42
- "learning_rate": 4.59546925566343e-06,
43
- "loss": 2.834,
44
  "step": 500
45
  },
46
  {
47
- "epoch": 0.22189349112426035,
48
- "grad_norm": 0.23730890452861786,
49
- "learning_rate": 5.520110957004161e-06,
50
- "loss": 2.8,
51
  "step": 600
52
  },
53
  {
54
- "epoch": 0.2588757396449704,
55
- "grad_norm": 0.150528684258461,
56
- "learning_rate": 6.444752658344892e-06,
57
- "loss": 2.7822,
58
  "step": 700
59
  },
60
  {
61
- "epoch": 0.2958579881656805,
62
- "grad_norm": 0.1429228037595749,
63
- "learning_rate": 7.369394359685622e-06,
64
- "loss": 2.7178,
65
  "step": 800
66
  },
67
  {
68
- "epoch": 0.3328402366863905,
69
- "grad_norm": 0.30748799443244934,
70
- "learning_rate": 8.294036061026353e-06,
71
- "loss": 2.7005,
72
  "step": 900
73
  },
74
  {
75
- "epoch": 0.3698224852071006,
76
- "grad_norm": 0.163905069231987,
77
- "learning_rate": 9.218677762367084e-06,
78
- "loss": 2.6718,
79
  "step": 1000
80
  },
81
  {
82
- "epoch": 0.4068047337278107,
83
- "grad_norm": 0.2939082384109497,
84
- "learning_rate": 1.0143319463707815e-05,
85
- "loss": 2.6669,
86
  "step": 1100
87
  },
88
  {
89
- "epoch": 0.4437869822485207,
90
- "grad_norm": 0.22948721051216125,
91
- "learning_rate": 1.1067961165048544e-05,
92
- "loss": 2.6516,
93
  "step": 1200
94
  },
95
  {
96
- "epoch": 0.4807692307692308,
97
- "grad_norm": 0.19319839775562286,
98
- "learning_rate": 1.1992602866389276e-05,
99
- "loss": 2.647,
100
  "step": 1300
101
  },
102
  {
103
- "epoch": 0.5177514792899408,
104
- "grad_norm": 0.21269312500953674,
105
- "learning_rate": 1.2917244567730005e-05,
106
- "loss": 2.6304,
107
  "step": 1400
108
  },
109
  {
110
- "epoch": 0.5547337278106509,
111
- "grad_norm": 0.2854045331478119,
112
- "learning_rate": 1.3841886269070736e-05,
113
- "loss": 2.6148,
114
  "step": 1500
115
  },
116
  {
117
- "epoch": 0.591715976331361,
118
- "grad_norm": 0.21025152504444122,
119
- "learning_rate": 1.4766527970411465e-05,
120
- "loss": 2.6068,
121
  "step": 1600
122
  },
123
  {
124
- "epoch": 0.628698224852071,
125
- "grad_norm": 0.2919473350048065,
126
- "learning_rate": 1.5691169671752196e-05,
127
- "loss": 2.6001,
128
  "step": 1700
129
  },
130
  {
131
- "epoch": 0.665680473372781,
132
- "grad_norm": 0.21243709325790405,
133
- "learning_rate": 1.661581137309293e-05,
134
- "loss": 2.5868,
135
  "step": 1800
136
  },
137
  {
138
- "epoch": 0.7026627218934911,
139
- "grad_norm": 0.2556278109550476,
140
- "learning_rate": 1.7540453074433658e-05,
141
- "loss": 2.5812,
142
  "step": 1900
143
  },
144
  {
145
- "epoch": 0.7396449704142012,
146
- "grad_norm": 0.2942062318325043,
147
- "learning_rate": 1.8465094775774387e-05,
148
- "loss": 2.5816,
149
  "step": 2000
150
  },
151
  {
152
- "epoch": 0.7766272189349113,
153
- "grad_norm": 0.23691928386688232,
154
- "learning_rate": 1.938973647711512e-05,
155
- "loss": 2.5727,
156
  "step": 2100
157
  },
158
  {
159
- "epoch": 0.8136094674556213,
160
- "grad_norm": 0.2230062484741211,
161
- "learning_rate": 1.980572281912978e-05,
162
- "loss": 2.5698,
163
  "step": 2200
164
  },
165
  {
166
- "epoch": 0.8505917159763313,
167
- "grad_norm": 0.23846900463104248,
168
- "learning_rate": 1.7122208281783788e-05,
169
- "loss": 2.552,
170
  "step": 2300
171
  },
172
  {
173
- "epoch": 0.8875739644970414,
174
- "grad_norm": 0.24352632462978363,
175
- "learning_rate": 1.2103724124062295e-05,
176
- "loss": 2.565,
177
  "step": 2400
178
  },
179
  {
180
- "epoch": 0.9245562130177515,
181
- "grad_norm": 0.2532407343387604,
182
- "learning_rate": 6.395547707284e-06,
183
- "loss": 2.5503,
184
  "step": 2500
185
  },
186
  {
187
- "epoch": 0.9615384615384616,
188
- "grad_norm": 0.2501896619796753,
189
- "learning_rate": 1.869067510077197e-06,
190
- "loss": 2.5497,
191
  "step": 2600
192
  },
193
  {
194
- "epoch": 0.9985207100591716,
195
- "grad_norm": 0.21952682733535767,
196
- "learning_rate": 8.260598198537306e-09,
197
- "loss": 2.5555,
198
  "step": 2700
199
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
200
  {
201
  "epoch": 1.0,
202
- "step": 2704,
203
- "total_flos": 3.940214344788214e+17,
204
- "train_loss": 2.675769654604105,
205
- "train_runtime": 864.3634,
206
- "train_samples_per_second": 50.048,
207
- "train_steps_per_second": 3.128
208
  }
209
  ],
210
  "logging_steps": 100,
211
- "max_steps": 2704,
212
  "num_input_tokens_seen": 0,
213
  "num_train_epochs": 1,
214
  "save_steps": 500,
@@ -224,7 +441,7 @@
224
  "attributes": {}
225
  }
226
  },
227
- "total_flos": 3.940214344788214e+17,
228
  "train_batch_size": 2,
229
  "trial_name": null,
230
  "trial_params": null
 
3
  "best_model_checkpoint": null,
4
  "epoch": 1.0,
5
  "eval_steps": 500,
6
+ "global_step": 5802,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.01723543605653223,
13
+ "grad_norm": 0.04860854893922806,
14
+ "learning_rate": 4.309416074121957e-07,
15
+ "loss": 2.8479,
16
  "step": 100
17
  },
18
  {
19
+ "epoch": 0.03447087211306446,
20
+ "grad_norm": 0.04985743388533592,
21
+ "learning_rate": 8.618832148243914e-07,
22
+ "loss": 2.8411,
23
  "step": 200
24
  },
25
  {
26
+ "epoch": 0.05170630816959669,
27
+ "grad_norm": 0.05735534057021141,
28
+ "learning_rate": 1.292824822236587e-06,
29
+ "loss": 2.8386,
30
  "step": 300
31
  },
32
  {
33
+ "epoch": 0.06894174422612892,
34
+ "grad_norm": 0.06343039870262146,
35
+ "learning_rate": 1.7237664296487828e-06,
36
+ "loss": 2.8455,
37
  "step": 400
38
  },
39
  {
40
+ "epoch": 0.08617718028266115,
41
+ "grad_norm": 0.08376861363649368,
42
+ "learning_rate": 2.1547080370609785e-06,
43
+ "loss": 2.8307,
44
  "step": 500
45
  },
46
  {
47
+ "epoch": 0.10341261633919338,
48
+ "grad_norm": 0.09948327392339706,
49
+ "learning_rate": 2.585649644473174e-06,
50
+ "loss": 2.8059,
51
  "step": 600
52
  },
53
  {
54
+ "epoch": 0.12064805239572561,
55
+ "grad_norm": 0.12135949730873108,
56
+ "learning_rate": 3.0165912518853694e-06,
57
+ "loss": 2.8064,
58
  "step": 700
59
  },
60
  {
61
+ "epoch": 0.13788348845225784,
62
+ "grad_norm": 0.11192767322063446,
63
+ "learning_rate": 3.4475328592975655e-06,
64
+ "loss": 2.7798,
65
  "step": 800
66
  },
67
  {
68
+ "epoch": 0.15511892450879008,
69
+ "grad_norm": 0.1301010698080063,
70
+ "learning_rate": 3.878474466709762e-06,
71
+ "loss": 2.8018,
72
  "step": 900
73
  },
74
  {
75
+ "epoch": 0.1723543605653223,
76
+ "grad_norm": 0.17249374091625214,
77
+ "learning_rate": 4.309416074121957e-06,
78
+ "loss": 2.7911,
79
  "step": 1000
80
  },
81
  {
82
+ "epoch": 0.18958979662185454,
83
+ "grad_norm": 0.14794030785560608,
84
+ "learning_rate": 4.740357681534152e-06,
85
+ "loss": 2.7674,
86
  "step": 1100
87
  },
88
  {
89
+ "epoch": 0.20682523267838676,
90
+ "grad_norm": 0.15302610397338867,
91
+ "learning_rate": 5.171299288946348e-06,
92
+ "loss": 2.7775,
93
  "step": 1200
94
  },
95
  {
96
+ "epoch": 0.224060668734919,
97
+ "grad_norm": 0.17765101790428162,
98
+ "learning_rate": 5.6022408963585436e-06,
99
+ "loss": 2.7465,
100
  "step": 1300
101
  },
102
  {
103
+ "epoch": 0.24129610479145122,
104
+ "grad_norm": 0.1747065782546997,
105
+ "learning_rate": 6.033182503770739e-06,
106
+ "loss": 2.7453,
107
  "step": 1400
108
  },
109
  {
110
+ "epoch": 0.25853154084798347,
111
+ "grad_norm": 0.19892176985740662,
112
+ "learning_rate": 6.464124111182936e-06,
113
+ "loss": 2.74,
114
  "step": 1500
115
  },
116
  {
117
+ "epoch": 0.2757669769045157,
118
+ "grad_norm": 0.2117730677127838,
119
+ "learning_rate": 6.895065718595131e-06,
120
+ "loss": 2.7307,
121
  "step": 1600
122
  },
123
  {
124
+ "epoch": 0.2930024129610479,
125
+ "grad_norm": 0.198543980717659,
126
+ "learning_rate": 7.326007326007326e-06,
127
+ "loss": 2.7261,
128
  "step": 1700
129
  },
130
  {
131
+ "epoch": 0.31023784901758017,
132
+ "grad_norm": 0.2144545465707779,
133
+ "learning_rate": 7.756948933419523e-06,
134
+ "loss": 2.7167,
135
  "step": 1800
136
  },
137
  {
138
+ "epoch": 0.3274732850741124,
139
+ "grad_norm": 0.20613212883472443,
140
+ "learning_rate": 8.187890540831719e-06,
141
+ "loss": 2.7269,
142
  "step": 1900
143
  },
144
  {
145
+ "epoch": 0.3447087211306446,
146
+ "grad_norm": 0.21626847982406616,
147
+ "learning_rate": 8.618832148243914e-06,
148
+ "loss": 2.7157,
149
  "step": 2000
150
  },
151
  {
152
+ "epoch": 0.3619441571871768,
153
+ "grad_norm": 0.2106010764837265,
154
+ "learning_rate": 9.049773755656109e-06,
155
+ "loss": 2.7148,
156
  "step": 2100
157
  },
158
  {
159
+ "epoch": 0.3791795932437091,
160
+ "grad_norm": 0.2537670433521271,
161
+ "learning_rate": 9.480715363068304e-06,
162
+ "loss": 2.7181,
163
  "step": 2200
164
  },
165
  {
166
+ "epoch": 0.3964150293002413,
167
+ "grad_norm": 0.23555806279182434,
168
+ "learning_rate": 9.911656970480501e-06,
169
+ "loss": 2.6989,
170
  "step": 2300
171
  },
172
  {
173
+ "epoch": 0.4136504653567735,
174
+ "grad_norm": 0.2480691522359848,
175
+ "learning_rate": 1.0342598577892697e-05,
176
+ "loss": 2.6951,
177
  "step": 2400
178
  },
179
  {
180
+ "epoch": 0.43088590141330574,
181
+ "grad_norm": 0.2250298112630844,
182
+ "learning_rate": 1.0773540185304892e-05,
183
+ "loss": 2.705,
184
  "step": 2500
185
  },
186
  {
187
+ "epoch": 0.448121337469838,
188
+ "grad_norm": 0.2429608255624771,
189
+ "learning_rate": 1.1204481792717087e-05,
190
+ "loss": 2.6872,
191
  "step": 2600
192
  },
193
  {
194
+ "epoch": 0.4653567735263702,
195
+ "grad_norm": 0.251676082611084,
196
+ "learning_rate": 1.1635423400129284e-05,
197
+ "loss": 2.6947,
198
  "step": 2700
199
  },
200
+ {
201
+ "epoch": 0.48259220958290244,
202
+ "grad_norm": 0.2518726587295532,
203
+ "learning_rate": 1.2066365007541478e-05,
204
+ "loss": 2.6895,
205
+ "step": 2800
206
+ },
207
+ {
208
+ "epoch": 0.49982764563943466,
209
+ "grad_norm": 0.24320507049560547,
210
+ "learning_rate": 1.2497306614953675e-05,
211
+ "loss": 2.6864,
212
+ "step": 2900
213
+ },
214
+ {
215
+ "epoch": 0.5170630816959669,
216
+ "grad_norm": 0.2310476005077362,
217
+ "learning_rate": 1.2928248222365872e-05,
218
+ "loss": 2.6808,
219
+ "step": 3000
220
+ },
221
+ {
222
+ "epoch": 0.5342985177524991,
223
+ "grad_norm": 0.22928917407989502,
224
+ "learning_rate": 1.3359189829778065e-05,
225
+ "loss": 2.6752,
226
+ "step": 3100
227
+ },
228
+ {
229
+ "epoch": 0.5515339538090314,
230
+ "grad_norm": 0.24209704995155334,
231
+ "learning_rate": 1.3790131437190262e-05,
232
+ "loss": 2.6736,
233
+ "step": 3200
234
+ },
235
+ {
236
+ "epoch": 0.5687693898655636,
237
+ "grad_norm": 0.23778828978538513,
238
+ "learning_rate": 1.4221073044602457e-05,
239
+ "loss": 2.6668,
240
+ "step": 3300
241
+ },
242
+ {
243
+ "epoch": 0.5860048259220958,
244
+ "grad_norm": 0.2457701861858368,
245
+ "learning_rate": 1.4652014652014653e-05,
246
+ "loss": 2.6593,
247
+ "step": 3400
248
+ },
249
+ {
250
+ "epoch": 0.603240261978628,
251
+ "grad_norm": 0.2431216984987259,
252
+ "learning_rate": 1.5082956259426848e-05,
253
+ "loss": 2.6671,
254
+ "step": 3500
255
+ },
256
+ {
257
+ "epoch": 0.6204756980351603,
258
+ "grad_norm": 0.23864516615867615,
259
+ "learning_rate": 1.5513897866839047e-05,
260
+ "loss": 2.6496,
261
+ "step": 3600
262
+ },
263
+ {
264
+ "epoch": 0.6377111340916926,
265
+ "grad_norm": 0.2714649438858032,
266
+ "learning_rate": 1.594483947425124e-05,
267
+ "loss": 2.6536,
268
+ "step": 3700
269
+ },
270
+ {
271
+ "epoch": 0.6549465701482248,
272
+ "grad_norm": 0.2654697895050049,
273
+ "learning_rate": 1.6375781081663437e-05,
274
+ "loss": 2.6632,
275
+ "step": 3800
276
+ },
277
+ {
278
+ "epoch": 0.672182006204757,
279
+ "grad_norm": 0.27372944355010986,
280
+ "learning_rate": 1.680672268907563e-05,
281
+ "loss": 2.6646,
282
+ "step": 3900
283
+ },
284
+ {
285
+ "epoch": 0.6894174422612892,
286
+ "grad_norm": 0.265690416097641,
287
+ "learning_rate": 1.7237664296487828e-05,
288
+ "loss": 2.6554,
289
+ "step": 4000
290
+ },
291
+ {
292
+ "epoch": 0.7066528783178214,
293
+ "grad_norm": 0.27707237005233765,
294
+ "learning_rate": 1.766860590390002e-05,
295
+ "loss": 2.6552,
296
+ "step": 4100
297
+ },
298
+ {
299
+ "epoch": 0.7238883143743536,
300
+ "grad_norm": 0.2691991627216339,
301
+ "learning_rate": 1.8099547511312218e-05,
302
+ "loss": 2.6336,
303
+ "step": 4200
304
+ },
305
+ {
306
+ "epoch": 0.7411237504308859,
307
+ "grad_norm": 0.2783598303794861,
308
+ "learning_rate": 1.8530489118724415e-05,
309
+ "loss": 2.6479,
310
+ "step": 4300
311
+ },
312
+ {
313
+ "epoch": 0.7583591864874182,
314
+ "grad_norm": 0.27648258209228516,
315
+ "learning_rate": 1.896143072613661e-05,
316
+ "loss": 2.6394,
317
+ "step": 4400
318
+ },
319
+ {
320
+ "epoch": 0.7755946225439504,
321
+ "grad_norm": 0.28013989329338074,
322
+ "learning_rate": 1.9392372333548806e-05,
323
+ "loss": 2.6528,
324
+ "step": 4500
325
+ },
326
+ {
327
+ "epoch": 0.7928300586004826,
328
+ "grad_norm": 0.28954795002937317,
329
+ "learning_rate": 1.9823313940961003e-05,
330
+ "loss": 2.6485,
331
+ "step": 4600
332
+ },
333
+ {
334
+ "epoch": 0.8100654946570148,
335
+ "grad_norm": 0.2726477384567261,
336
+ "learning_rate": 1.9872829409331243e-05,
337
+ "loss": 2.6422,
338
+ "step": 4700
339
+ },
340
+ {
341
+ "epoch": 0.827300930713547,
342
+ "grad_norm": 0.27916714549064636,
343
+ "learning_rate": 1.9088640248516185e-05,
344
+ "loss": 2.6409,
345
+ "step": 4800
346
+ },
347
+ {
348
+ "epoch": 0.8445363667700793,
349
+ "grad_norm": 0.28933390974998474,
350
+ "learning_rate": 1.7643022980549708e-05,
351
+ "loss": 2.6298,
352
+ "step": 4900
353
+ },
354
+ {
355
+ "epoch": 0.8617718028266115,
356
+ "grad_norm": 0.27602875232696533,
357
+ "learning_rate": 1.5641182770018732e-05,
358
+ "loss": 2.6445,
359
+ "step": 5000
360
+ },
361
+ {
362
+ "epoch": 0.8790072388831437,
363
+ "grad_norm": 0.32238247990608215,
364
+ "learning_rate": 1.3228804047714462e-05,
365
+ "loss": 2.6318,
366
+ "step": 5100
367
+ },
368
+ {
369
+ "epoch": 0.896242674939676,
370
+ "grad_norm": 0.32085633277893066,
371
+ "learning_rate": 1.0581448289104759e-05,
372
+ "loss": 2.6317,
373
+ "step": 5200
374
+ },
375
+ {
376
+ "epoch": 0.9134781109962082,
377
+ "grad_norm": 0.2872781455516815,
378
+ "learning_rate": 7.891777483276439e-06,
379
+ "loss": 2.6305,
380
+ "step": 5300
381
+ },
382
+ {
383
+ "epoch": 0.9307135470527405,
384
+ "grad_norm": 0.29465451836586,
385
+ "learning_rate": 5.3555331076468585e-06,
386
+ "loss": 2.6339,
387
+ "step": 5400
388
+ },
389
+ {
390
+ "epoch": 0.9479489831092727,
391
+ "grad_norm": 0.27886396646499634,
392
+ "learning_rate": 3.1572909921285387e-06,
393
+ "loss": 2.6381,
394
+ "step": 5500
395
+ },
396
+ {
397
+ "epoch": 0.9651844191658049,
398
+ "grad_norm": 0.3047686219215393,
399
+ "learning_rate": 1.4570287661325555e-06,
400
+ "loss": 2.6391,
401
+ "step": 5600
402
+ },
403
+ {
404
+ "epoch": 0.9824198552223371,
405
+ "grad_norm": 0.29873353242874146,
406
+ "learning_rate": 3.7848344587376297e-07,
407
+ "loss": 2.6345,
408
+ "step": 5700
409
+ },
410
+ {
411
+ "epoch": 0.9996552912788693,
412
+ "grad_norm": 0.2810242474079132,
413
+ "learning_rate": 1.4644152457554683e-10,
414
+ "loss": 2.6207,
415
+ "step": 5800
416
+ },
417
  {
418
  "epoch": 1.0,
419
+ "step": 5802,
420
+ "total_flos": 8.454557554683412e+17,
421
+ "train_loss": 2.7013893893405267,
422
+ "train_runtime": 1640.3942,
423
+ "train_samples_per_second": 56.586,
424
+ "train_steps_per_second": 3.537
425
  }
426
  ],
427
  "logging_steps": 100,
428
+ "max_steps": 5802,
429
  "num_input_tokens_seen": 0,
430
  "num_train_epochs": 1,
431
  "save_steps": 500,
 
441
  "attributes": {}
442
  }
443
  },
444
+ "total_flos": 8.454557554683412e+17,
445
  "train_batch_size": 2,
446
  "trial_name": null,
447
  "trial_params": null