rbelanec commited on
Commit
3c1c82a
·
verified ·
1 Parent(s): 7ea29a3

End of training

Browse files
README.md CHANGED
@@ -4,6 +4,7 @@ license: llama3
4
  base_model: meta-llama/Meta-Llama-3-8B-Instruct
5
  tags:
6
  - llama-factory
 
7
  - generated_from_trainer
8
  model-index:
9
  - name: train_copa_1754652160
@@ -15,9 +16,9 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # train_copa_1754652160
17
 
18
- This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.2497
21
  - Num Input Tokens Seen: 281856
22
 
23
  ## Model description
 
4
  base_model: meta-llama/Meta-Llama-3-8B-Instruct
5
  tags:
6
  - llama-factory
7
+ - p-tuning
8
  - generated_from_trainer
9
  model-index:
10
  - name: train_copa_1754652160
 
16
 
17
  # train_copa_1754652160
18
 
19
+ This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on the copa dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.1383
22
  - Num Input Tokens Seen: 281856
23
 
24
  ## Model description
all_results.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 10.0,
3
+ "eval_loss": 0.13831530511379242,
4
+ "eval_runtime": 0.9699,
5
+ "eval_samples_per_second": 41.242,
6
+ "eval_steps_per_second": 10.31,
7
+ "num_input_tokens_seen": 281856,
8
+ "total_flos": 1.2691848290107392e+16,
9
+ "train_loss": 0.195216263138508,
10
+ "train_runtime": 204.8687,
11
+ "train_samples_per_second": 17.572,
12
+ "train_steps_per_second": 4.393
13
+ }
eval_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 10.0,
3
+ "eval_loss": 0.13831530511379242,
4
+ "eval_runtime": 0.9699,
5
+ "eval_samples_per_second": 41.242,
6
+ "eval_steps_per_second": 10.31,
7
+ "num_input_tokens_seen": 281856
8
+ }
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 10.0,
3
+ "num_input_tokens_seen": 281856,
4
+ "total_flos": 1.2691848290107392e+16,
5
+ "train_loss": 0.195216263138508,
6
+ "train_runtime": 204.8687,
7
+ "train_samples_per_second": 17.572,
8
+ "train_steps_per_second": 4.393
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,1664 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 360,
3
+ "best_metric": 0.1140967383980751,
4
+ "best_model_checkpoint": "saves/p-tuning/llama-3-8b-instruct/train_copa_1754652160/checkpoint-360",
5
+ "epoch": 10.0,
6
+ "eval_steps": 45,
7
+ "global_step": 900,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.05555555555555555,
14
+ "grad_norm": 211.93502807617188,
15
+ "learning_rate": 2.2222222222222225e-06,
16
+ "loss": 4.5242,
17
+ "num_input_tokens_seen": 1536,
18
+ "step": 5
19
+ },
20
+ {
21
+ "epoch": 0.1111111111111111,
22
+ "grad_norm": 55.84653854370117,
23
+ "learning_rate": 5e-06,
24
+ "loss": 3.1759,
25
+ "num_input_tokens_seen": 3168,
26
+ "step": 10
27
+ },
28
+ {
29
+ "epoch": 0.16666666666666666,
30
+ "grad_norm": 77.70538330078125,
31
+ "learning_rate": 7.777777777777777e-06,
32
+ "loss": 2.3772,
33
+ "num_input_tokens_seen": 4736,
34
+ "step": 15
35
+ },
36
+ {
37
+ "epoch": 0.2222222222222222,
38
+ "grad_norm": 88.52490997314453,
39
+ "learning_rate": 1.0555555555555555e-05,
40
+ "loss": 1.808,
41
+ "num_input_tokens_seen": 6304,
42
+ "step": 20
43
+ },
44
+ {
45
+ "epoch": 0.2777777777777778,
46
+ "grad_norm": 171.03692626953125,
47
+ "learning_rate": 1.3333333333333333e-05,
48
+ "loss": 1.0693,
49
+ "num_input_tokens_seen": 7840,
50
+ "step": 25
51
+ },
52
+ {
53
+ "epoch": 0.3333333333333333,
54
+ "grad_norm": 217.7040557861328,
55
+ "learning_rate": 1.6111111111111115e-05,
56
+ "loss": 0.7198,
57
+ "num_input_tokens_seen": 9408,
58
+ "step": 30
59
+ },
60
+ {
61
+ "epoch": 0.3888888888888889,
62
+ "grad_norm": 158.4123077392578,
63
+ "learning_rate": 1.888888888888889e-05,
64
+ "loss": 1.0175,
65
+ "num_input_tokens_seen": 10912,
66
+ "step": 35
67
+ },
68
+ {
69
+ "epoch": 0.4444444444444444,
70
+ "grad_norm": 36.769805908203125,
71
+ "learning_rate": 2.1666666666666667e-05,
72
+ "loss": 1.0818,
73
+ "num_input_tokens_seen": 12448,
74
+ "step": 40
75
+ },
76
+ {
77
+ "epoch": 0.5,
78
+ "grad_norm": 28.32155418395996,
79
+ "learning_rate": 2.4444444444444445e-05,
80
+ "loss": 0.5598,
81
+ "num_input_tokens_seen": 14016,
82
+ "step": 45
83
+ },
84
+ {
85
+ "epoch": 0.5,
86
+ "eval_loss": 0.3143084943294525,
87
+ "eval_runtime": 0.9332,
88
+ "eval_samples_per_second": 42.863,
89
+ "eval_steps_per_second": 10.716,
90
+ "num_input_tokens_seen": 14016,
91
+ "step": 45
92
+ },
93
+ {
94
+ "epoch": 0.5555555555555556,
95
+ "grad_norm": 243.9434356689453,
96
+ "learning_rate": 2.7222222222222223e-05,
97
+ "loss": 0.4034,
98
+ "num_input_tokens_seen": 15584,
99
+ "step": 50
100
+ },
101
+ {
102
+ "epoch": 0.6111111111111112,
103
+ "grad_norm": 21.132078170776367,
104
+ "learning_rate": 3e-05,
105
+ "loss": 0.4113,
106
+ "num_input_tokens_seen": 17184,
107
+ "step": 55
108
+ },
109
+ {
110
+ "epoch": 0.6666666666666666,
111
+ "grad_norm": 11.026226043701172,
112
+ "learning_rate": 3.277777777777778e-05,
113
+ "loss": 0.2833,
114
+ "num_input_tokens_seen": 18752,
115
+ "step": 60
116
+ },
117
+ {
118
+ "epoch": 0.7222222222222222,
119
+ "grad_norm": 13.724584579467773,
120
+ "learning_rate": 3.555555555555556e-05,
121
+ "loss": 0.2309,
122
+ "num_input_tokens_seen": 20352,
123
+ "step": 65
124
+ },
125
+ {
126
+ "epoch": 0.7777777777777778,
127
+ "grad_norm": 78.358642578125,
128
+ "learning_rate": 3.8333333333333334e-05,
129
+ "loss": 0.741,
130
+ "num_input_tokens_seen": 21952,
131
+ "step": 70
132
+ },
133
+ {
134
+ "epoch": 0.8333333333333334,
135
+ "grad_norm": 8.49279499053955,
136
+ "learning_rate": 4.111111111111111e-05,
137
+ "loss": 0.2628,
138
+ "num_input_tokens_seen": 23456,
139
+ "step": 75
140
+ },
141
+ {
142
+ "epoch": 0.8888888888888888,
143
+ "grad_norm": 5.491214275360107,
144
+ "learning_rate": 4.388888888888889e-05,
145
+ "loss": 0.3305,
146
+ "num_input_tokens_seen": 25056,
147
+ "step": 80
148
+ },
149
+ {
150
+ "epoch": 0.9444444444444444,
151
+ "grad_norm": 229.71315002441406,
152
+ "learning_rate": 4.666666666666667e-05,
153
+ "loss": 4.0693,
154
+ "num_input_tokens_seen": 26560,
155
+ "step": 85
156
+ },
157
+ {
158
+ "epoch": 1.0,
159
+ "grad_norm": 65.04704284667969,
160
+ "learning_rate": 4.9444444444444446e-05,
161
+ "loss": 1.6129,
162
+ "num_input_tokens_seen": 28096,
163
+ "step": 90
164
+ },
165
+ {
166
+ "epoch": 1.0,
167
+ "eval_loss": 0.28230223059654236,
168
+ "eval_runtime": 0.9341,
169
+ "eval_samples_per_second": 42.823,
170
+ "eval_steps_per_second": 10.706,
171
+ "num_input_tokens_seen": 28096,
172
+ "step": 90
173
+ },
174
+ {
175
+ "epoch": 1.0555555555555556,
176
+ "grad_norm": 2.8092379570007324,
177
+ "learning_rate": 4.9996991493233693e-05,
178
+ "loss": 0.2852,
179
+ "num_input_tokens_seen": 29696,
180
+ "step": 95
181
+ },
182
+ {
183
+ "epoch": 1.1111111111111112,
184
+ "grad_norm": 3.550919532775879,
185
+ "learning_rate": 4.99847706754774e-05,
186
+ "loss": 0.2552,
187
+ "num_input_tokens_seen": 31232,
188
+ "step": 100
189
+ },
190
+ {
191
+ "epoch": 1.1666666666666667,
192
+ "grad_norm": 4.582396507263184,
193
+ "learning_rate": 4.9963154107272295e-05,
194
+ "loss": 0.279,
195
+ "num_input_tokens_seen": 32768,
196
+ "step": 105
197
+ },
198
+ {
199
+ "epoch": 1.2222222222222223,
200
+ "grad_norm": 3.2737157344818115,
201
+ "learning_rate": 4.993214991772563e-05,
202
+ "loss": 0.2366,
203
+ "num_input_tokens_seen": 34304,
204
+ "step": 110
205
+ },
206
+ {
207
+ "epoch": 1.2777777777777777,
208
+ "grad_norm": 5.495439529418945,
209
+ "learning_rate": 4.989176976624511e-05,
210
+ "loss": 0.6461,
211
+ "num_input_tokens_seen": 35872,
212
+ "step": 115
213
+ },
214
+ {
215
+ "epoch": 1.3333333333333333,
216
+ "grad_norm": 14.088470458984375,
217
+ "learning_rate": 4.9842028838154285e-05,
218
+ "loss": 0.3324,
219
+ "num_input_tokens_seen": 37408,
220
+ "step": 120
221
+ },
222
+ {
223
+ "epoch": 1.3888888888888888,
224
+ "grad_norm": 18.816679000854492,
225
+ "learning_rate": 4.978294583898196e-05,
226
+ "loss": 0.2167,
227
+ "num_input_tokens_seen": 38976,
228
+ "step": 125
229
+ },
230
+ {
231
+ "epoch": 1.4444444444444444,
232
+ "grad_norm": 10.976043701171875,
233
+ "learning_rate": 4.971454298742779e-05,
234
+ "loss": 0.2518,
235
+ "num_input_tokens_seen": 40576,
236
+ "step": 130
237
+ },
238
+ {
239
+ "epoch": 1.5,
240
+ "grad_norm": 136.6255340576172,
241
+ "learning_rate": 4.963684600700679e-05,
242
+ "loss": 0.8986,
243
+ "num_input_tokens_seen": 42144,
244
+ "step": 135
245
+ },
246
+ {
247
+ "epoch": 1.5,
248
+ "eval_loss": 0.3364465534687042,
249
+ "eval_runtime": 0.9378,
250
+ "eval_samples_per_second": 42.655,
251
+ "eval_steps_per_second": 10.664,
252
+ "num_input_tokens_seen": 42144,
253
+ "step": 135
254
+ },
255
+ {
256
+ "epoch": 1.5555555555555556,
257
+ "grad_norm": 6.133034706115723,
258
+ "learning_rate": 4.9549884116375714e-05,
259
+ "loss": 0.4373,
260
+ "num_input_tokens_seen": 43680,
261
+ "step": 140
262
+ },
263
+ {
264
+ "epoch": 1.6111111111111112,
265
+ "grad_norm": 3.9126594066619873,
266
+ "learning_rate": 4.9453690018345144e-05,
267
+ "loss": 1.3978,
268
+ "num_input_tokens_seen": 45248,
269
+ "step": 145
270
+ },
271
+ {
272
+ "epoch": 1.6666666666666665,
273
+ "grad_norm": 6.755434513092041,
274
+ "learning_rate": 4.934829988758131e-05,
275
+ "loss": 0.2847,
276
+ "num_input_tokens_seen": 46816,
277
+ "step": 150
278
+ },
279
+ {
280
+ "epoch": 1.7222222222222223,
281
+ "grad_norm": 18.98358154296875,
282
+ "learning_rate": 4.923375335700223e-05,
283
+ "loss": 0.2777,
284
+ "num_input_tokens_seen": 48384,
285
+ "step": 155
286
+ },
287
+ {
288
+ "epoch": 1.7777777777777777,
289
+ "grad_norm": 2.0623528957366943,
290
+ "learning_rate": 4.9110093502873476e-05,
291
+ "loss": 0.1408,
292
+ "num_input_tokens_seen": 49952,
293
+ "step": 160
294
+ },
295
+ {
296
+ "epoch": 1.8333333333333335,
297
+ "grad_norm": 17.358171463012695,
298
+ "learning_rate": 4.897736682860885e-05,
299
+ "loss": 0.0471,
300
+ "num_input_tokens_seen": 51520,
301
+ "step": 165
302
+ },
303
+ {
304
+ "epoch": 1.8888888888888888,
305
+ "grad_norm": 22.74451446533203,
306
+ "learning_rate": 4.883562324728241e-05,
307
+ "loss": 0.1595,
308
+ "num_input_tokens_seen": 53024,
309
+ "step": 170
310
+ },
311
+ {
312
+ "epoch": 1.9444444444444444,
313
+ "grad_norm": 7.793264865875244,
314
+ "learning_rate": 4.868491606285823e-05,
315
+ "loss": 0.2312,
316
+ "num_input_tokens_seen": 54592,
317
+ "step": 175
318
+ },
319
+ {
320
+ "epoch": 2.0,
321
+ "grad_norm": 3.0166492462158203,
322
+ "learning_rate": 4.8525301950144894e-05,
323
+ "loss": 0.162,
324
+ "num_input_tokens_seen": 56128,
325
+ "step": 180
326
+ },
327
+ {
328
+ "epoch": 2.0,
329
+ "eval_loss": 0.12522773444652557,
330
+ "eval_runtime": 0.9365,
331
+ "eval_samples_per_second": 42.713,
332
+ "eval_steps_per_second": 10.678,
333
+ "num_input_tokens_seen": 56128,
334
+ "step": 180
335
+ },
336
+ {
337
+ "epoch": 2.0555555555555554,
338
+ "grad_norm": 3.614006519317627,
339
+ "learning_rate": 4.835684093348244e-05,
340
+ "loss": 0.0719,
341
+ "num_input_tokens_seen": 57696,
342
+ "step": 185
343
+ },
344
+ {
345
+ "epoch": 2.111111111111111,
346
+ "grad_norm": 11.736072540283203,
347
+ "learning_rate": 4.817959636416969e-05,
348
+ "loss": 0.3451,
349
+ "num_input_tokens_seen": 59264,
350
+ "step": 190
351
+ },
352
+ {
353
+ "epoch": 2.1666666666666665,
354
+ "grad_norm": 0.017602359876036644,
355
+ "learning_rate": 4.7993634896640394e-05,
356
+ "loss": 0.0367,
357
+ "num_input_tokens_seen": 60864,
358
+ "step": 195
359
+ },
360
+ {
361
+ "epoch": 2.2222222222222223,
362
+ "grad_norm": 128.3651885986328,
363
+ "learning_rate": 4.779902646339722e-05,
364
+ "loss": 0.1548,
365
+ "num_input_tokens_seen": 62464,
366
+ "step": 200
367
+ },
368
+ {
369
+ "epoch": 2.2777777777777777,
370
+ "grad_norm": 4.051515579223633,
371
+ "learning_rate": 4.759584424871302e-05,
372
+ "loss": 0.1765,
373
+ "num_input_tokens_seen": 64032,
374
+ "step": 205
375
+ },
376
+ {
377
+ "epoch": 2.3333333333333335,
378
+ "grad_norm": 0.603679895401001,
379
+ "learning_rate": 4.7384164661109176e-05,
380
+ "loss": 0.1182,
381
+ "num_input_tokens_seen": 65568,
382
+ "step": 210
383
+ },
384
+ {
385
+ "epoch": 2.388888888888889,
386
+ "grad_norm": 0.7103855609893799,
387
+ "learning_rate": 4.7164067304621536e-05,
388
+ "loss": 0.0868,
389
+ "num_input_tokens_seen": 67104,
390
+ "step": 215
391
+ },
392
+ {
393
+ "epoch": 2.4444444444444446,
394
+ "grad_norm": 3.1308445930480957,
395
+ "learning_rate": 4.693563494886455e-05,
396
+ "loss": 0.2294,
397
+ "num_input_tokens_seen": 68704,
398
+ "step": 220
399
+ },
400
+ {
401
+ "epoch": 2.5,
402
+ "grad_norm": 3.5392675399780273,
403
+ "learning_rate": 4.669895349790502e-05,
404
+ "loss": 0.0545,
405
+ "num_input_tokens_seen": 70272,
406
+ "step": 225
407
+ },
408
+ {
409
+ "epoch": 2.5,
410
+ "eval_loss": 0.1658935248851776,
411
+ "eval_runtime": 0.9481,
412
+ "eval_samples_per_second": 42.19,
413
+ "eval_steps_per_second": 10.548,
414
+ "num_input_tokens_seen": 70272,
415
+ "step": 225
416
+ },
417
+ {
418
+ "epoch": 2.5555555555555554,
419
+ "grad_norm": 1.4086837768554688,
420
+ "learning_rate": 4.645411195795709e-05,
421
+ "loss": 0.0091,
422
+ "num_input_tokens_seen": 71808,
423
+ "step": 230
424
+ },
425
+ {
426
+ "epoch": 2.611111111111111,
427
+ "grad_norm": 6.530168056488037,
428
+ "learning_rate": 4.620120240391065e-05,
429
+ "loss": 0.1379,
430
+ "num_input_tokens_seen": 73408,
431
+ "step": 235
432
+ },
433
+ {
434
+ "epoch": 2.6666666666666665,
435
+ "grad_norm": 4.456235885620117,
436
+ "learning_rate": 4.5940319944705736e-05,
437
+ "loss": 0.1022,
438
+ "num_input_tokens_seen": 74912,
439
+ "step": 240
440
+ },
441
+ {
442
+ "epoch": 2.7222222222222223,
443
+ "grad_norm": 4.331588268280029,
444
+ "learning_rate": 4.567156268756594e-05,
445
+ "loss": 0.1812,
446
+ "num_input_tokens_seen": 76544,
447
+ "step": 245
448
+ },
449
+ {
450
+ "epoch": 2.7777777777777777,
451
+ "grad_norm": 0.16590723395347595,
452
+ "learning_rate": 4.539503170110431e-05,
453
+ "loss": 0.0448,
454
+ "num_input_tokens_seen": 78112,
455
+ "step": 250
456
+ },
457
+ {
458
+ "epoch": 2.8333333333333335,
459
+ "grad_norm": 0.05489581450819969,
460
+ "learning_rate": 4.5110830977315556e-05,
461
+ "loss": 0.004,
462
+ "num_input_tokens_seen": 79712,
463
+ "step": 255
464
+ },
465
+ {
466
+ "epoch": 2.888888888888889,
467
+ "grad_norm": 0.06641557812690735,
468
+ "learning_rate": 4.4819067392468944e-05,
469
+ "loss": 0.076,
470
+ "num_input_tokens_seen": 81280,
471
+ "step": 260
472
+ },
473
+ {
474
+ "epoch": 2.9444444444444446,
475
+ "grad_norm": 0.9573344588279724,
476
+ "learning_rate": 4.4519850666916484e-05,
477
+ "loss": 0.0639,
478
+ "num_input_tokens_seen": 82848,
479
+ "step": 265
480
+ },
481
+ {
482
+ "epoch": 3.0,
483
+ "grad_norm": 0.09293336421251297,
484
+ "learning_rate": 4.4213293323831585e-05,
485
+ "loss": 0.0493,
486
+ "num_input_tokens_seen": 84352,
487
+ "step": 270
488
+ },
489
+ {
490
+ "epoch": 3.0,
491
+ "eval_loss": 0.11682591587305069,
492
+ "eval_runtime": 0.9379,
493
+ "eval_samples_per_second": 42.648,
494
+ "eval_steps_per_second": 10.662,
495
+ "num_input_tokens_seen": 84352,
496
+ "step": 270
497
+ },
498
+ {
499
+ "epoch": 3.0555555555555554,
500
+ "grad_norm": 2.5821824073791504,
501
+ "learning_rate": 4.38995106468937e-05,
502
+ "loss": 0.011,
503
+ "num_input_tokens_seen": 85920,
504
+ "step": 275
505
+ },
506
+ {
507
+ "epoch": 3.111111111111111,
508
+ "grad_norm": 0.03350764885544777,
509
+ "learning_rate": 4.357862063693486e-05,
510
+ "loss": 0.0094,
511
+ "num_input_tokens_seen": 87520,
512
+ "step": 280
513
+ },
514
+ {
515
+ "epoch": 3.1666666666666665,
516
+ "grad_norm": 0.02359146997332573,
517
+ "learning_rate": 4.325074396756437e-05,
518
+ "loss": 0.0009,
519
+ "num_input_tokens_seen": 89088,
520
+ "step": 285
521
+ },
522
+ {
523
+ "epoch": 3.2222222222222223,
524
+ "grad_norm": 0.9292595386505127,
525
+ "learning_rate": 4.2916003939788403e-05,
526
+ "loss": 0.0859,
527
+ "num_input_tokens_seen": 90688,
528
+ "step": 290
529
+ },
530
+ {
531
+ "epoch": 3.2777777777777777,
532
+ "grad_norm": 0.3055465817451477,
533
+ "learning_rate": 4.257452643564155e-05,
534
+ "loss": 0.1095,
535
+ "num_input_tokens_seen": 92160,
536
+ "step": 295
537
+ },
538
+ {
539
+ "epoch": 3.3333333333333335,
540
+ "grad_norm": 8.65383243560791,
541
+ "learning_rate": 4.22264398708477e-05,
542
+ "loss": 0.1926,
543
+ "num_input_tokens_seen": 93760,
544
+ "step": 300
545
+ },
546
+ {
547
+ "epoch": 3.388888888888889,
548
+ "grad_norm": 5.946500301361084,
549
+ "learning_rate": 4.1871875146528195e-05,
550
+ "loss": 0.1205,
551
+ "num_input_tokens_seen": 95360,
552
+ "step": 305
553
+ },
554
+ {
555
+ "epoch": 3.4444444444444446,
556
+ "grad_norm": 0.05868682637810707,
557
+ "learning_rate": 4.1510965599975196e-05,
558
+ "loss": 0.0432,
559
+ "num_input_tokens_seen": 96928,
560
+ "step": 310
561
+ },
562
+ {
563
+ "epoch": 3.5,
564
+ "grad_norm": 0.07760920375585556,
565
+ "learning_rate": 4.114384695450906e-05,
566
+ "loss": 0.0166,
567
+ "num_input_tokens_seen": 98464,
568
+ "step": 315
569
+ },
570
+ {
571
+ "epoch": 3.5,
572
+ "eval_loss": 0.16605496406555176,
573
+ "eval_runtime": 0.9477,
574
+ "eval_samples_per_second": 42.207,
575
+ "eval_steps_per_second": 10.552,
576
+ "num_input_tokens_seen": 98464,
577
+ "step": 315
578
+ },
579
+ {
580
+ "epoch": 3.5555555555555554,
581
+ "grad_norm": 4.408195972442627,
582
+ "learning_rate": 4.077065726843828e-05,
583
+ "loss": 0.0249,
584
+ "num_input_tokens_seen": 100064,
585
+ "step": 320
586
+ },
587
+ {
588
+ "epoch": 3.611111111111111,
589
+ "grad_norm": 0.08003110438585281,
590
+ "learning_rate": 4.039153688314145e-05,
591
+ "loss": 0.056,
592
+ "num_input_tokens_seen": 101600,
593
+ "step": 325
594
+ },
595
+ {
596
+ "epoch": 3.6666666666666665,
597
+ "grad_norm": 0.014739356003701687,
598
+ "learning_rate": 4.000662837029062e-05,
599
+ "loss": 0.01,
600
+ "num_input_tokens_seen": 103200,
601
+ "step": 330
602
+ },
603
+ {
604
+ "epoch": 3.7222222222222223,
605
+ "grad_norm": 4.2343668937683105,
606
+ "learning_rate": 3.961607647823583e-05,
607
+ "loss": 0.1279,
608
+ "num_input_tokens_seen": 104768,
609
+ "step": 335
610
+ },
611
+ {
612
+ "epoch": 3.7777777777777777,
613
+ "grad_norm": 0.9541974663734436,
614
+ "learning_rate": 3.9220028077571295e-05,
615
+ "loss": 0.043,
616
+ "num_input_tokens_seen": 106304,
617
+ "step": 340
618
+ },
619
+ {
620
+ "epoch": 3.8333333333333335,
621
+ "grad_norm": 0.029086820781230927,
622
+ "learning_rate": 3.881863210590332e-05,
623
+ "loss": 0.0006,
624
+ "num_input_tokens_seen": 107904,
625
+ "step": 345
626
+ },
627
+ {
628
+ "epoch": 3.888888888888889,
629
+ "grad_norm": 4.72139835357666,
630
+ "learning_rate": 3.841203951184095e-05,
631
+ "loss": 0.1241,
632
+ "num_input_tokens_seen": 109408,
633
+ "step": 350
634
+ },
635
+ {
636
+ "epoch": 3.9444444444444446,
637
+ "grad_norm": 7.2094197273254395,
638
+ "learning_rate": 3.8000403198230387e-05,
639
+ "loss": 0.0363,
640
+ "num_input_tokens_seen": 111008,
641
+ "step": 355
642
+ },
643
+ {
644
+ "epoch": 4.0,
645
+ "grad_norm": 0.01099959947168827,
646
+ "learning_rate": 3.75838779646545e-05,
647
+ "loss": 0.0146,
648
+ "num_input_tokens_seen": 112576,
649
+ "step": 360
650
+ },
651
+ {
652
+ "epoch": 4.0,
653
+ "eval_loss": 0.1140967383980751,
654
+ "eval_runtime": 0.943,
655
+ "eval_samples_per_second": 42.417,
656
+ "eval_steps_per_second": 10.604,
657
+ "num_input_tokens_seen": 112576,
658
+ "step": 360
659
+ },
660
+ {
661
+ "epoch": 4.055555555555555,
662
+ "grad_norm": 0.06453730165958405,
663
+ "learning_rate": 3.7162620449219e-05,
664
+ "loss": 0.1621,
665
+ "num_input_tokens_seen": 114144,
666
+ "step": 365
667
+ },
668
+ {
669
+ "epoch": 4.111111111111111,
670
+ "grad_norm": 1.2406953573226929,
671
+ "learning_rate": 3.673678906964727e-05,
672
+ "loss": 0.0137,
673
+ "num_input_tokens_seen": 115712,
674
+ "step": 370
675
+ },
676
+ {
677
+ "epoch": 4.166666666666667,
678
+ "grad_norm": 0.040748681873083115,
679
+ "learning_rate": 3.630654396370594e-05,
680
+ "loss": 0.0514,
681
+ "num_input_tokens_seen": 117216,
682
+ "step": 375
683
+ },
684
+ {
685
+ "epoch": 4.222222222222222,
686
+ "grad_norm": 0.04640405625104904,
687
+ "learning_rate": 3.5872046928983626e-05,
688
+ "loss": 0.0012,
689
+ "num_input_tokens_seen": 118816,
690
+ "step": 380
691
+ },
692
+ {
693
+ "epoch": 4.277777777777778,
694
+ "grad_norm": 0.4107241928577423,
695
+ "learning_rate": 3.543346136204545e-05,
696
+ "loss": 0.0132,
697
+ "num_input_tokens_seen": 120352,
698
+ "step": 385
699
+ },
700
+ {
701
+ "epoch": 4.333333333333333,
702
+ "grad_norm": 0.03547825664281845,
703
+ "learning_rate": 3.499095219698631e-05,
704
+ "loss": 0.0107,
705
+ "num_input_tokens_seen": 121920,
706
+ "step": 390
707
+ },
708
+ {
709
+ "epoch": 4.388888888888889,
710
+ "grad_norm": 4.26585578918457,
711
+ "learning_rate": 3.454468584340588e-05,
712
+ "loss": 0.0743,
713
+ "num_input_tokens_seen": 123456,
714
+ "step": 395
715
+ },
716
+ {
717
+ "epoch": 4.444444444444445,
718
+ "grad_norm": 0.04872285574674606,
719
+ "learning_rate": 3.409483012382879e-05,
720
+ "loss": 0.0116,
721
+ "num_input_tokens_seen": 125056,
722
+ "step": 400
723
+ },
724
+ {
725
+ "epoch": 4.5,
726
+ "grad_norm": 4.138394355773926,
727
+ "learning_rate": 3.364155421059342e-05,
728
+ "loss": 0.1392,
729
+ "num_input_tokens_seen": 126624,
730
+ "step": 405
731
+ },
732
+ {
733
+ "epoch": 4.5,
734
+ "eval_loss": 0.12619957327842712,
735
+ "eval_runtime": 0.9498,
736
+ "eval_samples_per_second": 42.114,
737
+ "eval_steps_per_second": 10.528,
738
+ "num_input_tokens_seen": 126624,
739
+ "step": 405
740
+ },
741
+ {
742
+ "epoch": 4.555555555555555,
743
+ "grad_norm": 0.07925529032945633,
744
+ "learning_rate": 3.318502856223311e-05,
745
+ "loss": 0.0013,
746
+ "num_input_tokens_seen": 128224,
747
+ "step": 410
748
+ },
749
+ {
750
+ "epoch": 4.611111111111111,
751
+ "grad_norm": 0.07759137451648712,
752
+ "learning_rate": 3.272542485937369e-05,
753
+ "loss": 0.0181,
754
+ "num_input_tokens_seen": 129728,
755
+ "step": 415
756
+ },
757
+ {
758
+ "epoch": 4.666666666666667,
759
+ "grad_norm": 0.020686015486717224,
760
+ "learning_rate": 3.2262915940171376e-05,
761
+ "loss": 0.002,
762
+ "num_input_tokens_seen": 131328,
763
+ "step": 420
764
+ },
765
+ {
766
+ "epoch": 4.722222222222222,
767
+ "grad_norm": 0.10983490198850632,
768
+ "learning_rate": 3.1797675735315455e-05,
769
+ "loss": 0.0026,
770
+ "num_input_tokens_seen": 132896,
771
+ "step": 425
772
+ },
773
+ {
774
+ "epoch": 4.777777777777778,
775
+ "grad_norm": 0.06394976377487183,
776
+ "learning_rate": 3.132987920262005e-05,
777
+ "loss": 0.0024,
778
+ "num_input_tokens_seen": 134496,
779
+ "step": 430
780
+ },
781
+ {
782
+ "epoch": 4.833333333333333,
783
+ "grad_norm": 0.006211922504007816,
784
+ "learning_rate": 3.085970226122962e-05,
785
+ "loss": 0.0031,
786
+ "num_input_tokens_seen": 136064,
787
+ "step": 435
788
+ },
789
+ {
790
+ "epoch": 4.888888888888889,
791
+ "grad_norm": 0.00576823391020298,
792
+ "learning_rate": 3.0387321725463e-05,
793
+ "loss": 0.0282,
794
+ "num_input_tokens_seen": 137664,
795
+ "step": 440
796
+ },
797
+ {
798
+ "epoch": 4.944444444444445,
799
+ "grad_norm": 0.004720824770629406,
800
+ "learning_rate": 2.9912915238320754e-05,
801
+ "loss": 0.0188,
802
+ "num_input_tokens_seen": 139232,
803
+ "step": 445
804
+ },
805
+ {
806
+ "epoch": 5.0,
807
+ "grad_norm": 0.12190665304660797,
808
+ "learning_rate": 2.9436661204680882e-05,
809
+ "loss": 0.0007,
810
+ "num_input_tokens_seen": 140832,
811
+ "step": 450
812
+ },
813
+ {
814
+ "epoch": 5.0,
815
+ "eval_loss": 0.16102315485477448,
816
+ "eval_runtime": 0.9518,
817
+ "eval_samples_per_second": 42.027,
818
+ "eval_steps_per_second": 10.507,
819
+ "num_input_tokens_seen": 140832,
820
+ "step": 450
821
+ },
822
+ {
823
+ "epoch": 5.055555555555555,
824
+ "grad_norm": 0.12717540562152863,
825
+ "learning_rate": 2.8958738724208072e-05,
826
+ "loss": 0.0051,
827
+ "num_input_tokens_seen": 142368,
828
+ "step": 455
829
+ },
830
+ {
831
+ "epoch": 5.111111111111111,
832
+ "grad_norm": 0.004431543871760368,
833
+ "learning_rate": 2.8479327524001636e-05,
834
+ "loss": 0.0012,
835
+ "num_input_tokens_seen": 144000,
836
+ "step": 460
837
+ },
838
+ {
839
+ "epoch": 5.166666666666667,
840
+ "grad_norm": 0.4021238684654236,
841
+ "learning_rate": 2.7998607891007495e-05,
842
+ "loss": 0.0018,
843
+ "num_input_tokens_seen": 145632,
844
+ "step": 465
845
+ },
846
+ {
847
+ "epoch": 5.222222222222222,
848
+ "grad_norm": 0.034988883882761,
849
+ "learning_rate": 2.7516760604219617e-05,
850
+ "loss": 0.0024,
851
+ "num_input_tokens_seen": 147168,
852
+ "step": 470
853
+ },
854
+ {
855
+ "epoch": 5.277777777777778,
856
+ "grad_norm": 0.008324529975652695,
857
+ "learning_rate": 2.7033966866696457e-05,
858
+ "loss": 0.0002,
859
+ "num_input_tokens_seen": 148736,
860
+ "step": 475
861
+ },
862
+ {
863
+ "epoch": 5.333333333333333,
864
+ "grad_norm": 0.012731038965284824,
865
+ "learning_rate": 2.6550408237417885e-05,
866
+ "loss": 0.0002,
867
+ "num_input_tokens_seen": 150304,
868
+ "step": 480
869
+ },
870
+ {
871
+ "epoch": 5.388888888888889,
872
+ "grad_norm": 0.09622369706630707,
873
+ "learning_rate": 2.6066266563008267e-05,
874
+ "loss": 0.0004,
875
+ "num_input_tokens_seen": 151872,
876
+ "step": 485
877
+ },
878
+ {
879
+ "epoch": 5.444444444444445,
880
+ "grad_norm": 11.24488639831543,
881
+ "learning_rate": 2.5581723909351406e-05,
882
+ "loss": 0.0472,
883
+ "num_input_tokens_seen": 153472,
884
+ "step": 490
885
+ },
886
+ {
887
+ "epoch": 5.5,
888
+ "grad_norm": 0.008177004754543304,
889
+ "learning_rate": 2.5096962493123012e-05,
890
+ "loss": 0.0002,
891
+ "num_input_tokens_seen": 154976,
892
+ "step": 495
893
+ },
894
+ {
895
+ "epoch": 5.5,
896
+ "eval_loss": 0.29022911190986633,
897
+ "eval_runtime": 0.9436,
898
+ "eval_samples_per_second": 42.392,
899
+ "eval_steps_per_second": 10.598,
900
+ "num_input_tokens_seen": 154976,
901
+ "step": 495
902
+ },
903
+ {
904
+ "epoch": 5.555555555555555,
905
+ "grad_norm": 0.12855511903762817,
906
+ "learning_rate": 2.461216461326642e-05,
907
+ "loss": 0.0007,
908
+ "num_input_tokens_seen": 156544,
909
+ "step": 500
910
+ },
911
+ {
912
+ "epoch": 5.611111111111111,
913
+ "grad_norm": 0.004401104990392923,
914
+ "learning_rate": 2.4127512582437485e-05,
915
+ "loss": 0.0186,
916
+ "num_input_tokens_seen": 158112,
917
+ "step": 505
918
+ },
919
+ {
920
+ "epoch": 5.666666666666667,
921
+ "grad_norm": 0.11994586884975433,
922
+ "learning_rate": 2.364318865844416e-05,
923
+ "loss": 0.0023,
924
+ "num_input_tokens_seen": 159680,
925
+ "step": 510
926
+ },
927
+ {
928
+ "epoch": 5.722222222222222,
929
+ "grad_norm": 0.0036123669706285,
930
+ "learning_rate": 2.3159374975706884e-05,
931
+ "loss": 0.0001,
932
+ "num_input_tokens_seen": 161312,
933
+ "step": 515
934
+ },
935
+ {
936
+ "epoch": 5.777777777777778,
937
+ "grad_norm": 0.005037173628807068,
938
+ "learning_rate": 2.2676253476765196e-05,
939
+ "loss": 0.0001,
940
+ "num_input_tokens_seen": 162880,
941
+ "step": 520
942
+ },
943
+ {
944
+ "epoch": 5.833333333333333,
945
+ "grad_norm": 0.004541746340692043,
946
+ "learning_rate": 2.2194005843856636e-05,
947
+ "loss": 0.0001,
948
+ "num_input_tokens_seen": 164448,
949
+ "step": 525
950
+ },
951
+ {
952
+ "epoch": 5.888888888888889,
953
+ "grad_norm": 0.003595268353819847,
954
+ "learning_rate": 2.1712813430593436e-05,
955
+ "loss": 0.0001,
956
+ "num_input_tokens_seen": 166016,
957
+ "step": 530
958
+ },
959
+ {
960
+ "epoch": 5.944444444444445,
961
+ "grad_norm": 0.05647879093885422,
962
+ "learning_rate": 2.1232857193762924e-05,
963
+ "loss": 0.1753,
964
+ "num_input_tokens_seen": 167552,
965
+ "step": 535
966
+ },
967
+ {
968
+ "epoch": 6.0,
969
+ "grad_norm": 0.024071840569376945,
970
+ "learning_rate": 2.0754317625276983e-05,
971
+ "loss": 0.0003,
972
+ "num_input_tokens_seen": 169056,
973
+ "step": 540
974
+ },
975
+ {
976
+ "epoch": 6.0,
977
+ "eval_loss": 0.18793392181396484,
978
+ "eval_runtime": 0.9426,
979
+ "eval_samples_per_second": 42.435,
980
+ "eval_steps_per_second": 10.609,
981
+ "num_input_tokens_seen": 169056,
982
+ "step": 540
983
+ },
984
+ {
985
+ "epoch": 6.055555555555555,
986
+ "grad_norm": 0.003708529518917203,
987
+ "learning_rate": 2.02773746842965e-05,
988
+ "loss": 0.0003,
989
+ "num_input_tokens_seen": 170592,
990
+ "step": 545
991
+ },
992
+ {
993
+ "epoch": 6.111111111111111,
994
+ "grad_norm": 0.029693368822336197,
995
+ "learning_rate": 1.980220772955602e-05,
996
+ "loss": 0.0837,
997
+ "num_input_tokens_seen": 172192,
998
+ "step": 550
999
+ },
1000
+ {
1001
+ "epoch": 6.166666666666667,
1002
+ "grad_norm": 1.9385651350021362,
1003
+ "learning_rate": 1.932899545191433e-05,
1004
+ "loss": 0.0069,
1005
+ "num_input_tokens_seen": 173792,
1006
+ "step": 555
1007
+ },
1008
+ {
1009
+ "epoch": 6.222222222222222,
1010
+ "grad_norm": 0.008497129194438457,
1011
+ "learning_rate": 1.8857915807156092e-05,
1012
+ "loss": 0.0003,
1013
+ "num_input_tokens_seen": 175360,
1014
+ "step": 560
1015
+ },
1016
+ {
1017
+ "epoch": 6.277777777777778,
1018
+ "grad_norm": 0.005275554955005646,
1019
+ "learning_rate": 1.838914594906995e-05,
1020
+ "loss": 0.0001,
1021
+ "num_input_tokens_seen": 176992,
1022
+ "step": 565
1023
+ },
1024
+ {
1025
+ "epoch": 6.333333333333333,
1026
+ "grad_norm": 0.007108105346560478,
1027
+ "learning_rate": 1.792286216282824e-05,
1028
+ "loss": 0.0007,
1029
+ "num_input_tokens_seen": 178592,
1030
+ "step": 570
1031
+ },
1032
+ {
1033
+ "epoch": 6.388888888888889,
1034
+ "grad_norm": 0.003662517061457038,
1035
+ "learning_rate": 1.7459239798693364e-05,
1036
+ "loss": 0.0003,
1037
+ "num_input_tokens_seen": 180128,
1038
+ "step": 575
1039
+ },
1040
+ {
1041
+ "epoch": 6.444444444444445,
1042
+ "grad_norm": 0.03587842732667923,
1043
+ "learning_rate": 1.699845320607571e-05,
1044
+ "loss": 0.0002,
1045
+ "num_input_tokens_seen": 181632,
1046
+ "step": 580
1047
+ },
1048
+ {
1049
+ "epoch": 6.5,
1050
+ "grad_norm": 0.5192928314208984,
1051
+ "learning_rate": 1.6540675667967974e-05,
1052
+ "loss": 0.0013,
1053
+ "num_input_tokens_seen": 183200,
1054
+ "step": 585
1055
+ },
1056
+ {
1057
+ "epoch": 6.5,
1058
+ "eval_loss": 0.23766329884529114,
1059
+ "eval_runtime": 0.9465,
1060
+ "eval_samples_per_second": 42.262,
1061
+ "eval_steps_per_second": 10.565,
1062
+ "num_input_tokens_seen": 183200,
1063
+ "step": 585
1064
+ },
1065
+ {
1066
+ "epoch": 6.555555555555555,
1067
+ "grad_norm": 0.02627461589872837,
1068
+ "learning_rate": 1.60860793357805e-05,
1069
+ "loss": 0.0004,
1070
+ "num_input_tokens_seen": 184800,
1071
+ "step": 590
1072
+ },
1073
+ {
1074
+ "epoch": 6.611111111111111,
1075
+ "grad_norm": 0.007544673513621092,
1076
+ "learning_rate": 1.56348351646022e-05,
1077
+ "loss": 0.0285,
1078
+ "num_input_tokens_seen": 186336,
1079
+ "step": 595
1080
+ },
1081
+ {
1082
+ "epoch": 6.666666666666667,
1083
+ "grad_norm": 0.0033945958130061626,
1084
+ "learning_rate": 1.5187112848911323e-05,
1085
+ "loss": 0.0001,
1086
+ "num_input_tokens_seen": 187904,
1087
+ "step": 600
1088
+ },
1089
+ {
1090
+ "epoch": 6.722222222222222,
1091
+ "grad_norm": 0.004758753813803196,
1092
+ "learning_rate": 1.47430807587603e-05,
1093
+ "loss": 0.0001,
1094
+ "num_input_tokens_seen": 189472,
1095
+ "step": 605
1096
+ },
1097
+ {
1098
+ "epoch": 6.777777777777778,
1099
+ "grad_norm": 0.33814820647239685,
1100
+ "learning_rate": 1.430290587645865e-05,
1101
+ "loss": 0.0013,
1102
+ "num_input_tokens_seen": 191072,
1103
+ "step": 610
1104
+ },
1105
+ {
1106
+ "epoch": 6.833333333333333,
1107
+ "grad_norm": 0.002982664154842496,
1108
+ "learning_rate": 1.3866753733777765e-05,
1109
+ "loss": 0.0001,
1110
+ "num_input_tokens_seen": 192608,
1111
+ "step": 615
1112
+ },
1113
+ {
1114
+ "epoch": 6.888888888888889,
1115
+ "grad_norm": 0.003444693749770522,
1116
+ "learning_rate": 1.343478834970121e-05,
1117
+ "loss": 0.0001,
1118
+ "num_input_tokens_seen": 194208,
1119
+ "step": 620
1120
+ },
1121
+ {
1122
+ "epoch": 6.944444444444445,
1123
+ "grad_norm": 0.0027797692455351353,
1124
+ "learning_rate": 1.3007172168743854e-05,
1125
+ "loss": 0.0001,
1126
+ "num_input_tokens_seen": 195776,
1127
+ "step": 625
1128
+ },
1129
+ {
1130
+ "epoch": 7.0,
1131
+ "grad_norm": 0.00582587905228138,
1132
+ "learning_rate": 1.2584065999863102e-05,
1133
+ "loss": 0.0001,
1134
+ "num_input_tokens_seen": 197344,
1135
+ "step": 630
1136
+ },
1137
+ {
1138
+ "epoch": 7.0,
1139
+ "eval_loss": 0.24828998744487762,
1140
+ "eval_runtime": 0.942,
1141
+ "eval_samples_per_second": 42.462,
1142
+ "eval_steps_per_second": 10.616,
1143
+ "num_input_tokens_seen": 197344,
1144
+ "step": 630
1145
+ },
1146
+ {
1147
+ "epoch": 7.055555555555555,
1148
+ "grad_norm": 0.012683599255979061,
1149
+ "learning_rate": 1.2165628955985314e-05,
1150
+ "loss": 0.0001,
1151
+ "num_input_tokens_seen": 198944,
1152
+ "step": 635
1153
+ },
1154
+ {
1155
+ "epoch": 7.111111111111111,
1156
+ "grad_norm": 0.0020441152155399323,
1157
+ "learning_rate": 1.175201839416988e-05,
1158
+ "loss": 0.0001,
1159
+ "num_input_tokens_seen": 200512,
1160
+ "step": 640
1161
+ },
1162
+ {
1163
+ "epoch": 7.166666666666667,
1164
+ "grad_norm": 0.0023130401968955994,
1165
+ "learning_rate": 1.1343389856433658e-05,
1166
+ "loss": 0.0001,
1167
+ "num_input_tokens_seen": 202016,
1168
+ "step": 645
1169
+ },
1170
+ {
1171
+ "epoch": 7.222222222222222,
1172
+ "grad_norm": 0.0022213098127394915,
1173
+ "learning_rate": 1.0939897011258001e-05,
1174
+ "loss": 0.0001,
1175
+ "num_input_tokens_seen": 203648,
1176
+ "step": 650
1177
+ },
1178
+ {
1179
+ "epoch": 7.277777777777778,
1180
+ "grad_norm": 0.002087386092171073,
1181
+ "learning_rate": 1.0541691595800337e-05,
1182
+ "loss": 0.0001,
1183
+ "num_input_tokens_seen": 205184,
1184
+ "step": 655
1185
+ },
1186
+ {
1187
+ "epoch": 7.333333333333333,
1188
+ "grad_norm": 0.0022648421581834555,
1189
+ "learning_rate": 1.0148923358832022e-05,
1190
+ "loss": 0.0001,
1191
+ "num_input_tokens_seen": 206720,
1192
+ "step": 660
1193
+ },
1194
+ {
1195
+ "epoch": 7.388888888888889,
1196
+ "grad_norm": 0.004804587922990322,
1197
+ "learning_rate": 9.761740004423927e-06,
1198
+ "loss": 0.0001,
1199
+ "num_input_tokens_seen": 208320,
1200
+ "step": 665
1201
+ },
1202
+ {
1203
+ "epoch": 7.444444444444445,
1204
+ "grad_norm": 0.0030888738110661507,
1205
+ "learning_rate": 9.380287136401e-06,
1206
+ "loss": 0.0001,
1207
+ "num_input_tokens_seen": 209856,
1208
+ "step": 670
1209
+ },
1210
+ {
1211
+ "epoch": 7.5,
1212
+ "grad_norm": 0.0017403181409463286,
1213
+ "learning_rate": 9.00470820358663e-06,
1214
+ "loss": 0.0002,
1215
+ "num_input_tokens_seen": 211392,
1216
+ "step": 675
1217
+ },
1218
+ {
1219
+ "epoch": 7.5,
1220
+ "eval_loss": 0.25386351346969604,
1221
+ "eval_runtime": 0.9402,
1222
+ "eval_samples_per_second": 42.544,
1223
+ "eval_steps_per_second": 10.636,
1224
+ "num_input_tokens_seen": 211392,
1225
+ "step": 675
1226
+ },
1227
+ {
1228
+ "epoch": 7.555555555555555,
1229
+ "grad_norm": 0.0020435000769793987,
1230
+ "learning_rate": 8.635144445857406e-06,
1231
+ "loss": 0.0001,
1232
+ "num_input_tokens_seen": 212960,
1233
+ "step": 680
1234
+ },
1235
+ {
1236
+ "epoch": 7.611111111111111,
1237
+ "grad_norm": 0.004450716078281403,
1238
+ "learning_rate": 8.271734841028553e-06,
1239
+ "loss": 0.0002,
1240
+ "num_input_tokens_seen": 214528,
1241
+ "step": 685
1242
+ },
1243
+ {
1244
+ "epoch": 7.666666666666667,
1245
+ "grad_norm": 0.002327506896108389,
1246
+ "learning_rate": 7.914616052590071e-06,
1247
+ "loss": 0.0001,
1248
+ "num_input_tokens_seen": 216000,
1249
+ "step": 690
1250
+ },
1251
+ {
1252
+ "epoch": 7.722222222222222,
1253
+ "grad_norm": 0.0016173458425328135,
1254
+ "learning_rate": 7.563922378313218e-06,
1255
+ "loss": 0.0097,
1256
+ "num_input_tokens_seen": 217632,
1257
+ "step": 695
1258
+ },
1259
+ {
1260
+ "epoch": 7.777777777777778,
1261
+ "grad_norm": 0.0022673553321510553,
1262
+ "learning_rate": 7.219785699746573e-06,
1263
+ "loss": 0.0001,
1264
+ "num_input_tokens_seen": 219232,
1265
+ "step": 700
1266
+ },
1267
+ {
1268
+ "epoch": 7.833333333333333,
1269
+ "grad_norm": 0.008347373455762863,
1270
+ "learning_rate": 6.882335432620779e-06,
1271
+ "loss": 0.0001,
1272
+ "num_input_tokens_seen": 220800,
1273
+ "step": 705
1274
+ },
1275
+ {
1276
+ "epoch": 7.888888888888889,
1277
+ "grad_norm": 0.002359850564971566,
1278
+ "learning_rate": 6.55169847818059e-06,
1279
+ "loss": 0.0001,
1280
+ "num_input_tokens_seen": 222368,
1281
+ "step": 710
1282
+ },
1283
+ {
1284
+ "epoch": 7.944444444444445,
1285
+ "grad_norm": 0.002948438050225377,
1286
+ "learning_rate": 6.22799917546252e-06,
1287
+ "loss": 0.0001,
1288
+ "num_input_tokens_seen": 223968,
1289
+ "step": 715
1290
+ },
1291
+ {
1292
+ "epoch": 8.0,
1293
+ "grad_norm": 0.0022273629438132048,
1294
+ "learning_rate": 5.9113592545359945e-06,
1295
+ "loss": 0.0001,
1296
+ "num_input_tokens_seen": 225536,
1297
+ "step": 720
1298
+ },
1299
+ {
1300
+ "epoch": 8.0,
1301
+ "eval_loss": 0.2521117627620697,
1302
+ "eval_runtime": 0.9388,
1303
+ "eval_samples_per_second": 42.606,
1304
+ "eval_steps_per_second": 10.651,
1305
+ "num_input_tokens_seen": 225536,
1306
+ "step": 720
1307
+ },
1308
+ {
1309
+ "epoch": 8.055555555555555,
1310
+ "grad_norm": 0.0022272937931120396,
1311
+ "learning_rate": 5.601897790725643e-06,
1312
+ "loss": 0.0001,
1313
+ "num_input_tokens_seen": 227168,
1314
+ "step": 725
1315
+ },
1316
+ {
1317
+ "epoch": 8.11111111111111,
1318
+ "grad_norm": 0.001649300567805767,
1319
+ "learning_rate": 5.299731159831953e-06,
1320
+ "loss": 0.0001,
1321
+ "num_input_tokens_seen": 228704,
1322
+ "step": 730
1323
+ },
1324
+ {
1325
+ "epoch": 8.166666666666666,
1326
+ "grad_norm": 0.0024456402752548456,
1327
+ "learning_rate": 5.004972994367102e-06,
1328
+ "loss": 0.0001,
1329
+ "num_input_tokens_seen": 230336,
1330
+ "step": 735
1331
+ },
1332
+ {
1333
+ "epoch": 8.222222222222221,
1334
+ "grad_norm": 0.022852079942822456,
1335
+ "learning_rate": 4.7177341408224e-06,
1336
+ "loss": 0.0001,
1337
+ "num_input_tokens_seen": 231936,
1338
+ "step": 740
1339
+ },
1340
+ {
1341
+ "epoch": 8.277777777777779,
1342
+ "grad_norm": 0.0026604924350976944,
1343
+ "learning_rate": 4.438122617983443e-06,
1344
+ "loss": 0.0001,
1345
+ "num_input_tokens_seen": 233472,
1346
+ "step": 745
1347
+ },
1348
+ {
1349
+ "epoch": 8.333333333333334,
1350
+ "grad_norm": 0.0019071005517616868,
1351
+ "learning_rate": 4.166243576308712e-06,
1352
+ "loss": 0.0001,
1353
+ "num_input_tokens_seen": 235040,
1354
+ "step": 750
1355
+ },
1356
+ {
1357
+ "epoch": 8.38888888888889,
1358
+ "grad_norm": 0.007083934266120195,
1359
+ "learning_rate": 3.9021992583867325e-06,
1360
+ "loss": 0.0001,
1361
+ "num_input_tokens_seen": 236608,
1362
+ "step": 755
1363
+ },
1364
+ {
1365
+ "epoch": 8.444444444444445,
1366
+ "grad_norm": 0.00216863676905632,
1367
+ "learning_rate": 3.6460889604868626e-06,
1368
+ "loss": 0.0008,
1369
+ "num_input_tokens_seen": 238144,
1370
+ "step": 760
1371
+ },
1372
+ {
1373
+ "epoch": 8.5,
1374
+ "grad_norm": 0.0029811670538038015,
1375
+ "learning_rate": 3.398008995217988e-06,
1376
+ "loss": 0.0001,
1377
+ "num_input_tokens_seen": 239680,
1378
+ "step": 765
1379
+ },
1380
+ {
1381
+ "epoch": 8.5,
1382
+ "eval_loss": 0.24617867171764374,
1383
+ "eval_runtime": 0.9394,
1384
+ "eval_samples_per_second": 42.581,
1385
+ "eval_steps_per_second": 10.645,
1386
+ "num_input_tokens_seen": 239680,
1387
+ "step": 765
1388
+ },
1389
+ {
1390
+ "epoch": 8.555555555555555,
1391
+ "grad_norm": 0.0019673961214721203,
1392
+ "learning_rate": 3.158052655309332e-06,
1393
+ "loss": 0.0001,
1394
+ "num_input_tokens_seen": 241280,
1395
+ "step": 770
1396
+ },
1397
+ {
1398
+ "epoch": 8.61111111111111,
1399
+ "grad_norm": 0.0018027002224698663,
1400
+ "learning_rate": 2.9263101785268254e-06,
1401
+ "loss": 0.0001,
1402
+ "num_input_tokens_seen": 242816,
1403
+ "step": 775
1404
+ },
1405
+ {
1406
+ "epoch": 8.666666666666666,
1407
+ "grad_norm": 0.0017442210810258985,
1408
+ "learning_rate": 2.7028687137384267e-06,
1409
+ "loss": 0.0001,
1410
+ "num_input_tokens_seen": 244352,
1411
+ "step": 780
1412
+ },
1413
+ {
1414
+ "epoch": 8.722222222222221,
1415
+ "grad_norm": 0.001671099103987217,
1416
+ "learning_rate": 2.487812288140945e-06,
1417
+ "loss": 0.0002,
1418
+ "num_input_tokens_seen": 245856,
1419
+ "step": 785
1420
+ },
1421
+ {
1422
+ "epoch": 8.777777777777779,
1423
+ "grad_norm": 0.012044212780892849,
1424
+ "learning_rate": 2.281221775660894e-06,
1425
+ "loss": 0.0001,
1426
+ "num_input_tokens_seen": 247456,
1427
+ "step": 790
1428
+ },
1429
+ {
1430
+ "epoch": 8.833333333333334,
1431
+ "grad_norm": 0.001784978318028152,
1432
+ "learning_rate": 2.0831748665410765e-06,
1433
+ "loss": 0.0001,
1434
+ "num_input_tokens_seen": 248992,
1435
+ "step": 795
1436
+ },
1437
+ {
1438
+ "epoch": 8.88888888888889,
1439
+ "grad_norm": 0.001969333505257964,
1440
+ "learning_rate": 1.893746038124497e-06,
1441
+ "loss": 0.0001,
1442
+ "num_input_tokens_seen": 250528,
1443
+ "step": 800
1444
+ },
1445
+ {
1446
+ "epoch": 8.944444444444445,
1447
+ "grad_norm": 0.002932202536612749,
1448
+ "learning_rate": 1.713006526846439e-06,
1449
+ "loss": 0.0001,
1450
+ "num_input_tokens_seen": 252128,
1451
+ "step": 805
1452
+ },
1453
+ {
1454
+ "epoch": 9.0,
1455
+ "grad_norm": 0.001777714816853404,
1456
+ "learning_rate": 1.541024301445404e-06,
1457
+ "loss": 0.0001,
1458
+ "num_input_tokens_seen": 253696,
1459
+ "step": 810
1460
+ },
1461
+ {
1462
+ "epoch": 9.0,
1463
+ "eval_loss": 0.2544807493686676,
1464
+ "eval_runtime": 0.9396,
1465
+ "eval_samples_per_second": 42.57,
1466
+ "eval_steps_per_second": 10.643,
1467
+ "num_input_tokens_seen": 253696,
1468
+ "step": 810
1469
+ },
1470
+ {
1471
+ "epoch": 9.055555555555555,
1472
+ "grad_norm": 0.0017283963970839977,
1473
+ "learning_rate": 1.3778640374027985e-06,
1474
+ "loss": 0.0001,
1475
+ "num_input_tokens_seen": 255296,
1476
+ "step": 815
1477
+ },
1478
+ {
1479
+ "epoch": 9.11111111111111,
1480
+ "grad_norm": 0.0016462679486721754,
1481
+ "learning_rate": 1.2235870926211619e-06,
1482
+ "loss": 0.0001,
1483
+ "num_input_tokens_seen": 256896,
1484
+ "step": 820
1485
+ },
1486
+ {
1487
+ "epoch": 9.166666666666666,
1488
+ "grad_norm": 0.0017024942208081484,
1489
+ "learning_rate": 1.0782514843499653e-06,
1490
+ "loss": 0.0001,
1491
+ "num_input_tokens_seen": 258432,
1492
+ "step": 825
1493
+ },
1494
+ {
1495
+ "epoch": 9.222222222222221,
1496
+ "grad_norm": 0.0017732703126966953,
1497
+ "learning_rate": 9.419118673676924e-07,
1498
+ "loss": 0.0001,
1499
+ "num_input_tokens_seen": 260000,
1500
+ "step": 830
1501
+ },
1502
+ {
1503
+ "epoch": 9.277777777777779,
1504
+ "grad_norm": 0.0017788108671084046,
1505
+ "learning_rate": 8.146195134284052e-07,
1506
+ "loss": 0.0001,
1507
+ "num_input_tokens_seen": 261568,
1508
+ "step": 835
1509
+ },
1510
+ {
1511
+ "epoch": 9.333333333333334,
1512
+ "grad_norm": 0.001521819387562573,
1513
+ "learning_rate": 6.964222919805391e-07,
1514
+ "loss": 0.0001,
1515
+ "num_input_tokens_seen": 263200,
1516
+ "step": 840
1517
+ },
1518
+ {
1519
+ "epoch": 9.38888888888889,
1520
+ "grad_norm": 0.07402946054935455,
1521
+ "learning_rate": 5.87364652165176e-07,
1522
+ "loss": 0.0002,
1523
+ "num_input_tokens_seen": 264736,
1524
+ "step": 845
1525
+ },
1526
+ {
1527
+ "epoch": 9.444444444444445,
1528
+ "grad_norm": 0.005844974424690008,
1529
+ "learning_rate": 4.874876061005173e-07,
1530
+ "loss": 0.0001,
1531
+ "num_input_tokens_seen": 266304,
1532
+ "step": 850
1533
+ },
1534
+ {
1535
+ "epoch": 9.5,
1536
+ "grad_norm": 0.0018154801800847054,
1537
+ "learning_rate": 3.9682871345891883e-07,
1538
+ "loss": 0.0001,
1539
+ "num_input_tokens_seen": 267840,
1540
+ "step": 855
1541
+ },
1542
+ {
1543
+ "epoch": 9.5,
1544
+ "eval_loss": 0.24860987067222595,
1545
+ "eval_runtime": 0.9418,
1546
+ "eval_samples_per_second": 42.472,
1547
+ "eval_steps_per_second": 10.618,
1548
+ "num_input_tokens_seen": 267840,
1549
+ "step": 855
1550
+ },
1551
+ {
1552
+ "epoch": 9.555555555555555,
1553
+ "grad_norm": 0.0017057630466297269,
1554
+ "learning_rate": 3.1542206734221924e-07,
1555
+ "loss": 0.0001,
1556
+ "num_input_tokens_seen": 269376,
1557
+ "step": 860
1558
+ },
1559
+ {
1560
+ "epoch": 9.61111111111111,
1561
+ "grad_norm": 0.004398024175316095,
1562
+ "learning_rate": 2.4329828146074095e-07,
1563
+ "loss": 0.0005,
1564
+ "num_input_tokens_seen": 270944,
1565
+ "step": 865
1566
+ },
1567
+ {
1568
+ "epoch": 9.666666666666666,
1569
+ "grad_norm": 0.022089840844273567,
1570
+ "learning_rate": 1.8048447862070718e-07,
1571
+ "loss": 0.0001,
1572
+ "num_input_tokens_seen": 272448,
1573
+ "step": 870
1574
+ },
1575
+ {
1576
+ "epoch": 9.722222222222221,
1577
+ "grad_norm": 0.0015576216392219067,
1578
+ "learning_rate": 1.2700428052447033e-07,
1579
+ "loss": 0.0001,
1580
+ "num_input_tokens_seen": 273984,
1581
+ "step": 875
1582
+ },
1583
+ {
1584
+ "epoch": 9.777777777777779,
1585
+ "grad_norm": 0.021998705342411995,
1586
+ "learning_rate": 8.28777988873486e-08,
1587
+ "loss": 0.0001,
1588
+ "num_input_tokens_seen": 275520,
1589
+ "step": 880
1590
+ },
1591
+ {
1592
+ "epoch": 9.833333333333334,
1593
+ "grad_norm": 0.0017123895231634378,
1594
+ "learning_rate": 4.8121627874450625e-08,
1595
+ "loss": 0.0001,
1596
+ "num_input_tokens_seen": 277152,
1597
+ "step": 885
1598
+ },
1599
+ {
1600
+ "epoch": 9.88888888888889,
1601
+ "grad_norm": 0.001752789132297039,
1602
+ "learning_rate": 2.2748837860270267e-08,
1603
+ "loss": 0.0001,
1604
+ "num_input_tokens_seen": 278688,
1605
+ "step": 890
1606
+ },
1607
+ {
1608
+ "epoch": 9.944444444444445,
1609
+ "grad_norm": 0.0014716936275362968,
1610
+ "learning_rate": 6.768970513457151e-09,
1611
+ "loss": 0.0001,
1612
+ "num_input_tokens_seen": 280256,
1613
+ "step": 895
1614
+ },
1615
+ {
1616
+ "epoch": 10.0,
1617
+ "grad_norm": 0.0017592560034245253,
1618
+ "learning_rate": 1.8803520859811406e-10,
1619
+ "loss": 0.0001,
1620
+ "num_input_tokens_seen": 281856,
1621
+ "step": 900
1622
+ },
1623
+ {
1624
+ "epoch": 10.0,
1625
+ "eval_loss": 0.24965079128742218,
1626
+ "eval_runtime": 0.9391,
1627
+ "eval_samples_per_second": 42.592,
1628
+ "eval_steps_per_second": 10.648,
1629
+ "num_input_tokens_seen": 281856,
1630
+ "step": 900
1631
+ },
1632
+ {
1633
+ "epoch": 10.0,
1634
+ "num_input_tokens_seen": 281856,
1635
+ "step": 900,
1636
+ "total_flos": 1.2691848290107392e+16,
1637
+ "train_loss": 0.195216263138508,
1638
+ "train_runtime": 204.8687,
1639
+ "train_samples_per_second": 17.572,
1640
+ "train_steps_per_second": 4.393
1641
+ }
1642
+ ],
1643
+ "logging_steps": 5,
1644
+ "max_steps": 900,
1645
+ "num_input_tokens_seen": 281856,
1646
+ "num_train_epochs": 10,
1647
+ "save_steps": 45,
1648
+ "stateful_callbacks": {
1649
+ "TrainerControl": {
1650
+ "args": {
1651
+ "should_epoch_stop": false,
1652
+ "should_evaluate": false,
1653
+ "should_log": false,
1654
+ "should_save": true,
1655
+ "should_training_stop": true
1656
+ },
1657
+ "attributes": {}
1658
+ }
1659
+ },
1660
+ "total_flos": 1.2691848290107392e+16,
1661
+ "train_batch_size": 4,
1662
+ "trial_name": null,
1663
+ "trial_params": null
1664
+ }
training_eval_loss.png ADDED
training_loss.png ADDED