rbelanec commited on
Commit
5dc4fcb
·
verified ·
1 Parent(s): a76193b

End of training

Browse files
Files changed (5) hide show
  1. README.md +3 -3
  2. all_results.json +6 -6
  3. eval_results.json +3 -3
  4. train_results.json +3 -3
  5. trainer_state.json +54 -54
README.md CHANGED
@@ -17,10 +17,10 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # test
19
 
20
- This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.3585
23
- - Num Input Tokens Seen: 46944
24
 
25
  ## Model description
26
 
 
17
 
18
  # test
19
 
20
+ This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on the wsc dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 0.3497
23
+ - Num Input Tokens Seen: 49376
24
 
25
  ## Model description
26
 
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "epoch": 1.0,
3
  "eval_loss": 0.3496828079223633,
4
- "eval_runtime": 1.4986,
5
- "eval_samples_per_second": 37.369,
6
- "eval_steps_per_second": 9.342,
7
  "num_input_tokens_seen": 49376,
8
  "total_flos": 2223378963628032.0,
9
  "train_loss": 1.100914571762085,
10
- "train_runtime": 86.254,
11
- "train_samples_per_second": 5.774,
12
- "train_steps_per_second": 1.449
13
  }
 
1
  {
2
  "epoch": 1.0,
3
  "eval_loss": 0.3496828079223633,
4
+ "eval_runtime": 1.4753,
5
+ "eval_samples_per_second": 37.957,
6
+ "eval_steps_per_second": 9.489,
7
  "num_input_tokens_seen": 49376,
8
  "total_flos": 2223378963628032.0,
9
  "train_loss": 1.100914571762085,
10
+ "train_runtime": 75.0364,
11
+ "train_samples_per_second": 6.637,
12
+ "train_steps_per_second": 1.666
13
  }
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
  "eval_loss": 0.3496828079223633,
4
- "eval_runtime": 1.4986,
5
- "eval_samples_per_second": 37.369,
6
- "eval_steps_per_second": 9.342,
7
  "num_input_tokens_seen": 49376
8
  }
 
1
  {
2
  "epoch": 1.0,
3
  "eval_loss": 0.3496828079223633,
4
+ "eval_runtime": 1.4753,
5
+ "eval_samples_per_second": 37.957,
6
+ "eval_steps_per_second": 9.489,
7
  "num_input_tokens_seen": 49376
8
  }
train_results.json CHANGED
@@ -3,7 +3,7 @@
3
  "num_input_tokens_seen": 49376,
4
  "total_flos": 2223378963628032.0,
5
  "train_loss": 1.100914571762085,
6
- "train_runtime": 86.254,
7
- "train_samples_per_second": 5.774,
8
- "train_steps_per_second": 1.449
9
  }
 
3
  "num_input_tokens_seen": 49376,
4
  "total_flos": 2223378963628032.0,
5
  "train_loss": 1.100914571762085,
6
+ "train_runtime": 75.0364,
7
+ "train_samples_per_second": 6.637,
8
+ "train_steps_per_second": 1.666
9
  }
trainer_state.json CHANGED
@@ -20,9 +20,9 @@
20
  {
21
  "epoch": 0.056,
22
  "eval_loss": 6.932018280029297,
23
- "eval_runtime": 1.4196,
24
- "eval_samples_per_second": 39.449,
25
- "eval_steps_per_second": 9.862,
26
  "num_input_tokens_seen": 2880,
27
  "step": 7
28
  },
@@ -37,9 +37,9 @@
37
  {
38
  "epoch": 0.112,
39
  "eval_loss": 1.6227623224258423,
40
- "eval_runtime": 1.3856,
41
- "eval_samples_per_second": 40.417,
42
- "eval_steps_per_second": 10.104,
43
  "num_input_tokens_seen": 5920,
44
  "step": 14
45
  },
@@ -62,9 +62,9 @@
62
  {
63
  "epoch": 0.168,
64
  "eval_loss": 0.4040253460407257,
65
- "eval_runtime": 1.3702,
66
- "eval_samples_per_second": 40.869,
67
- "eval_steps_per_second": 10.217,
68
  "num_input_tokens_seen": 8416,
69
  "step": 21
70
  },
@@ -79,9 +79,9 @@
79
  {
80
  "epoch": 0.224,
81
  "eval_loss": 3.618654489517212,
82
- "eval_runtime": 1.3779,
83
- "eval_samples_per_second": 40.642,
84
- "eval_steps_per_second": 10.161,
85
  "num_input_tokens_seen": 11264,
86
  "step": 28
87
  },
@@ -104,9 +104,9 @@
104
  {
105
  "epoch": 0.28,
106
  "eval_loss": 0.36673951148986816,
107
- "eval_runtime": 1.3899,
108
- "eval_samples_per_second": 40.29,
109
- "eval_steps_per_second": 10.072,
110
  "num_input_tokens_seen": 13824,
111
  "step": 35
112
  },
@@ -121,9 +121,9 @@
121
  {
122
  "epoch": 0.336,
123
  "eval_loss": 0.3661668598651886,
124
- "eval_runtime": 1.3878,
125
- "eval_samples_per_second": 40.353,
126
- "eval_steps_per_second": 10.088,
127
  "num_input_tokens_seen": 16672,
128
  "step": 42
129
  },
@@ -138,9 +138,9 @@
138
  {
139
  "epoch": 0.392,
140
  "eval_loss": 0.5265085101127625,
141
- "eval_runtime": 1.3816,
142
- "eval_samples_per_second": 40.532,
143
- "eval_steps_per_second": 10.133,
144
  "num_input_tokens_seen": 19296,
145
  "step": 49
146
  },
@@ -163,9 +163,9 @@
163
  {
164
  "epoch": 0.448,
165
  "eval_loss": 0.39251771569252014,
166
- "eval_runtime": 1.3876,
167
- "eval_samples_per_second": 40.359,
168
- "eval_steps_per_second": 10.09,
169
  "num_input_tokens_seen": 22432,
170
  "step": 56
171
  },
@@ -180,9 +180,9 @@
180
  {
181
  "epoch": 0.504,
182
  "eval_loss": 2.652535915374756,
183
- "eval_runtime": 1.3806,
184
- "eval_samples_per_second": 40.562,
185
- "eval_steps_per_second": 10.141,
186
  "num_input_tokens_seen": 25504,
187
  "step": 63
188
  },
@@ -205,9 +205,9 @@
205
  {
206
  "epoch": 0.56,
207
  "eval_loss": 0.37763792276382446,
208
- "eval_runtime": 1.3741,
209
- "eval_samples_per_second": 40.753,
210
- "eval_steps_per_second": 10.188,
211
  "num_input_tokens_seen": 28064,
212
  "step": 70
213
  },
@@ -222,9 +222,9 @@
222
  {
223
  "epoch": 0.616,
224
  "eval_loss": 0.36274200677871704,
225
- "eval_runtime": 1.4269,
226
- "eval_samples_per_second": 39.245,
227
- "eval_steps_per_second": 9.811,
228
  "num_input_tokens_seen": 30720,
229
  "step": 77
230
  },
@@ -239,9 +239,9 @@
239
  {
240
  "epoch": 0.672,
241
  "eval_loss": 0.35259029269218445,
242
- "eval_runtime": 1.4343,
243
- "eval_samples_per_second": 39.044,
244
- "eval_steps_per_second": 9.761,
245
  "num_input_tokens_seen": 33504,
246
  "step": 84
247
  },
@@ -264,9 +264,9 @@
264
  {
265
  "epoch": 0.728,
266
  "eval_loss": 0.3546493351459503,
267
- "eval_runtime": 1.4075,
268
- "eval_samples_per_second": 39.786,
269
- "eval_steps_per_second": 9.946,
270
  "num_input_tokens_seen": 36128,
271
  "step": 91
272
  },
@@ -281,9 +281,9 @@
281
  {
282
  "epoch": 0.784,
283
  "eval_loss": 0.3496828079223633,
284
- "eval_runtime": 1.4395,
285
- "eval_samples_per_second": 38.902,
286
- "eval_steps_per_second": 9.726,
287
  "num_input_tokens_seen": 38592,
288
  "step": 98
289
  },
@@ -306,9 +306,9 @@
306
  {
307
  "epoch": 0.84,
308
  "eval_loss": 0.35384148359298706,
309
- "eval_runtime": 1.4061,
310
- "eval_samples_per_second": 39.827,
311
- "eval_steps_per_second": 9.957,
312
  "num_input_tokens_seen": 41280,
313
  "step": 105
314
  },
@@ -323,9 +323,9 @@
323
  {
324
  "epoch": 0.896,
325
  "eval_loss": 0.3565780818462372,
326
- "eval_runtime": 1.4483,
327
- "eval_samples_per_second": 38.666,
328
- "eval_steps_per_second": 9.666,
329
  "num_input_tokens_seen": 44160,
330
  "step": 112
331
  },
@@ -340,9 +340,9 @@
340
  {
341
  "epoch": 0.952,
342
  "eval_loss": 0.35846200585365295,
343
- "eval_runtime": 1.4597,
344
- "eval_samples_per_second": 38.364,
345
- "eval_steps_per_second": 9.591,
346
  "num_input_tokens_seen": 46944,
347
  "step": 119
348
  },
@@ -368,9 +368,9 @@
368
  "step": 125,
369
  "total_flos": 2223378963628032.0,
370
  "train_loss": 1.100914571762085,
371
- "train_runtime": 86.254,
372
- "train_samples_per_second": 5.774,
373
- "train_steps_per_second": 1.449
374
  }
375
  ],
376
  "logging_steps": 5,
 
20
  {
21
  "epoch": 0.056,
22
  "eval_loss": 6.932018280029297,
23
+ "eval_runtime": 1.3801,
24
+ "eval_samples_per_second": 40.577,
25
+ "eval_steps_per_second": 10.144,
26
  "num_input_tokens_seen": 2880,
27
  "step": 7
28
  },
 
37
  {
38
  "epoch": 0.112,
39
  "eval_loss": 1.6227623224258423,
40
+ "eval_runtime": 1.3899,
41
+ "eval_samples_per_second": 40.291,
42
+ "eval_steps_per_second": 10.073,
43
  "num_input_tokens_seen": 5920,
44
  "step": 14
45
  },
 
62
  {
63
  "epoch": 0.168,
64
  "eval_loss": 0.4040253460407257,
65
+ "eval_runtime": 1.3955,
66
+ "eval_samples_per_second": 40.129,
67
+ "eval_steps_per_second": 10.032,
68
  "num_input_tokens_seen": 8416,
69
  "step": 21
70
  },
 
79
  {
80
  "epoch": 0.224,
81
  "eval_loss": 3.618654489517212,
82
+ "eval_runtime": 1.3972,
83
+ "eval_samples_per_second": 40.08,
84
+ "eval_steps_per_second": 10.02,
85
  "num_input_tokens_seen": 11264,
86
  "step": 28
87
  },
 
104
  {
105
  "epoch": 0.28,
106
  "eval_loss": 0.36673951148986816,
107
+ "eval_runtime": 1.3617,
108
+ "eval_samples_per_second": 41.124,
109
+ "eval_steps_per_second": 10.281,
110
  "num_input_tokens_seen": 13824,
111
  "step": 35
112
  },
 
121
  {
122
  "epoch": 0.336,
123
  "eval_loss": 0.3661668598651886,
124
+ "eval_runtime": 1.3859,
125
+ "eval_samples_per_second": 40.406,
126
+ "eval_steps_per_second": 10.102,
127
  "num_input_tokens_seen": 16672,
128
  "step": 42
129
  },
 
138
  {
139
  "epoch": 0.392,
140
  "eval_loss": 0.5265085101127625,
141
+ "eval_runtime": 1.3679,
142
+ "eval_samples_per_second": 40.94,
143
+ "eval_steps_per_second": 10.235,
144
  "num_input_tokens_seen": 19296,
145
  "step": 49
146
  },
 
163
  {
164
  "epoch": 0.448,
165
  "eval_loss": 0.39251771569252014,
166
+ "eval_runtime": 1.381,
167
+ "eval_samples_per_second": 40.55,
168
+ "eval_steps_per_second": 10.138,
169
  "num_input_tokens_seen": 22432,
170
  "step": 56
171
  },
 
180
  {
181
  "epoch": 0.504,
182
  "eval_loss": 2.652535915374756,
183
+ "eval_runtime": 1.3855,
184
+ "eval_samples_per_second": 40.419,
185
+ "eval_steps_per_second": 10.105,
186
  "num_input_tokens_seen": 25504,
187
  "step": 63
188
  },
 
205
  {
206
  "epoch": 0.56,
207
  "eval_loss": 0.37763792276382446,
208
+ "eval_runtime": 1.3818,
209
+ "eval_samples_per_second": 40.527,
210
+ "eval_steps_per_second": 10.132,
211
  "num_input_tokens_seen": 28064,
212
  "step": 70
213
  },
 
222
  {
223
  "epoch": 0.616,
224
  "eval_loss": 0.36274200677871704,
225
+ "eval_runtime": 1.3919,
226
+ "eval_samples_per_second": 40.232,
227
+ "eval_steps_per_second": 10.058,
228
  "num_input_tokens_seen": 30720,
229
  "step": 77
230
  },
 
239
  {
240
  "epoch": 0.672,
241
  "eval_loss": 0.35259029269218445,
242
+ "eval_runtime": 1.4073,
243
+ "eval_samples_per_second": 39.794,
244
+ "eval_steps_per_second": 9.948,
245
  "num_input_tokens_seen": 33504,
246
  "step": 84
247
  },
 
264
  {
265
  "epoch": 0.728,
266
  "eval_loss": 0.3546493351459503,
267
+ "eval_runtime": 1.4289,
268
+ "eval_samples_per_second": 39.191,
269
+ "eval_steps_per_second": 9.798,
270
  "num_input_tokens_seen": 36128,
271
  "step": 91
272
  },
 
281
  {
282
  "epoch": 0.784,
283
  "eval_loss": 0.3496828079223633,
284
+ "eval_runtime": 1.4365,
285
+ "eval_samples_per_second": 38.985,
286
+ "eval_steps_per_second": 9.746,
287
  "num_input_tokens_seen": 38592,
288
  "step": 98
289
  },
 
306
  {
307
  "epoch": 0.84,
308
  "eval_loss": 0.35384148359298706,
309
+ "eval_runtime": 1.4278,
310
+ "eval_samples_per_second": 39.222,
311
+ "eval_steps_per_second": 9.806,
312
  "num_input_tokens_seen": 41280,
313
  "step": 105
314
  },
 
323
  {
324
  "epoch": 0.896,
325
  "eval_loss": 0.3565780818462372,
326
+ "eval_runtime": 1.4257,
327
+ "eval_samples_per_second": 39.28,
328
+ "eval_steps_per_second": 9.82,
329
  "num_input_tokens_seen": 44160,
330
  "step": 112
331
  },
 
340
  {
341
  "epoch": 0.952,
342
  "eval_loss": 0.35846200585365295,
343
+ "eval_runtime": 1.4469,
344
+ "eval_samples_per_second": 38.704,
345
+ "eval_steps_per_second": 9.676,
346
  "num_input_tokens_seen": 46944,
347
  "step": 119
348
  },
 
368
  "step": 125,
369
  "total_flos": 2223378963628032.0,
370
  "train_loss": 1.100914571762085,
371
+ "train_runtime": 75.0364,
372
+ "train_samples_per_second": 6.637,
373
+ "train_steps_per_second": 1.666
374
  }
375
  ],
376
  "logging_steps": 5,