Kei5uke commited on
Commit
c618d76
·
verified ·
1 Parent(s): dfb6e6b

Training in progress, step 30, checkpoint

Browse files
checkpoint-30/adapter_config.json CHANGED
@@ -23,13 +23,13 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "k_proj",
27
- "o_proj",
28
  "v_proj",
29
- "gate_proj",
 
30
  "up_proj",
31
- "down_proj",
32
- "q_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "down_proj",
 
27
  "v_proj",
28
+ "o_proj",
29
+ "q_proj",
30
  "up_proj",
31
+ "gate_proj",
32
+ "k_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
checkpoint-30/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e42ff132ae9e27b25c14b1a0d2e98b623f5d8a78aac496813de619136d7fdb27
3
  size 131146352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e73ad3aca4ce497d080cb8a6c619cd7f5891102a2eedc87026fe96b9db237ea
3
  size 131146352
checkpoint-30/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a6f5950021d2a9b0bba0756d546bc9a2f40e210c0800cc0a3dfc50b8468c8e8
3
  size 262615650
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2ee9e574a683ed57cea4fc36a2dabd24fd7170daaab96491ae55ca3db567a46
3
  size 262615650
checkpoint-30/trainer_state.json CHANGED
@@ -10,292 +10,292 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
- "grad_norm": 0.39027148485183716,
14
  "learning_rate": 4.000000000000001e-06,
15
- "loss": 1.1401,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
- "grad_norm": 0.1772935390472412,
21
  "learning_rate": 8.000000000000001e-06,
22
- "loss": 0.9133,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
- "grad_norm": 0.192855104804039,
28
  "learning_rate": 1.2e-05,
29
- "loss": 0.8036,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
- "eval_loss": 0.993510365486145,
35
- "eval_runtime": 46.5106,
36
- "eval_samples_per_second": 2.15,
37
- "eval_steps_per_second": 2.15,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
- "grad_norm": 0.23705194890499115,
43
  "learning_rate": 1.6000000000000003e-05,
44
- "loss": 0.9389,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
- "grad_norm": 0.22702816128730774,
50
  "learning_rate": 2e-05,
51
- "loss": 0.7836,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
- "grad_norm": 0.22573529183864594,
57
  "learning_rate": 1.9200000000000003e-05,
58
- "loss": 0.8874,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
- "eval_loss": 0.9887735843658447,
64
- "eval_runtime": 47.2769,
65
- "eval_samples_per_second": 2.115,
66
- "eval_steps_per_second": 2.115,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
- "grad_norm": 0.231398344039917,
72
  "learning_rate": 1.8400000000000003e-05,
73
- "loss": 0.8755,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
- "grad_norm": 0.28229379653930664,
79
  "learning_rate": 1.76e-05,
80
- "loss": 0.971,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
- "grad_norm": 0.21991951763629913,
86
  "learning_rate": 1.6800000000000002e-05,
87
- "loss": 0.8546,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
- "eval_loss": 0.9774181842803955,
93
- "eval_runtime": 47.3982,
94
- "eval_samples_per_second": 2.11,
95
- "eval_steps_per_second": 2.11,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
- "grad_norm": 0.26028916239738464,
101
  "learning_rate": 1.6000000000000003e-05,
102
- "loss": 1.0344,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
- "grad_norm": 0.23867206275463104,
108
  "learning_rate": 1.5200000000000002e-05,
109
- "loss": 0.8132,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
- "grad_norm": 0.30641764402389526,
115
  "learning_rate": 1.4400000000000001e-05,
116
- "loss": 1.1168,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
- "eval_loss": 0.963714599609375,
122
- "eval_runtime": 47.188,
123
- "eval_samples_per_second": 2.119,
124
- "eval_steps_per_second": 2.119,
125
  "step": 12
126
  },
127
  {
128
  "epoch": 0.008060765772748411,
129
- "grad_norm": 0.25963348150253296,
130
  "learning_rate": 1.3600000000000002e-05,
131
- "loss": 0.9735,
132
  "step": 13
133
  },
134
  {
135
  "epoch": 0.008680824678344443,
136
- "grad_norm": 0.2063651829957962,
137
  "learning_rate": 1.2800000000000001e-05,
138
- "loss": 0.7957,
139
  "step": 14
140
  },
141
  {
142
  "epoch": 0.009300883583940475,
143
- "grad_norm": 0.2287115454673767,
144
  "learning_rate": 1.2e-05,
145
- "loss": 0.8343,
146
  "step": 15
147
  },
148
  {
149
  "epoch": 0.009300883583940475,
150
- "eval_loss": 0.9510444402694702,
151
- "eval_runtime": 47.3972,
152
- "eval_samples_per_second": 2.11,
153
- "eval_steps_per_second": 2.11,
154
  "step": 15
155
  },
156
  {
157
  "epoch": 0.009920942489536505,
158
- "grad_norm": 0.2136881947517395,
159
  "learning_rate": 1.1200000000000001e-05,
160
- "loss": 0.7861,
161
  "step": 16
162
  },
163
  {
164
  "epoch": 0.010541001395132537,
165
- "grad_norm": 0.16134004294872284,
166
  "learning_rate": 1.04e-05,
167
- "loss": 0.8375,
168
  "step": 17
169
  },
170
  {
171
  "epoch": 0.01116106030072857,
172
- "grad_norm": 0.20020240545272827,
173
  "learning_rate": 9.600000000000001e-06,
174
- "loss": 0.8086,
175
  "step": 18
176
  },
177
  {
178
  "epoch": 0.01116106030072857,
179
- "eval_loss": 0.9415214657783508,
180
- "eval_runtime": 47.2291,
181
- "eval_samples_per_second": 2.117,
182
- "eval_steps_per_second": 2.117,
183
  "step": 18
184
  },
185
  {
186
  "epoch": 0.011781119206324601,
187
- "grad_norm": 0.2330344021320343,
188
  "learning_rate": 8.8e-06,
189
- "loss": 0.7858,
190
  "step": 19
191
  },
192
  {
193
  "epoch": 0.012401178111920633,
194
- "grad_norm": 0.175759956240654,
195
  "learning_rate": 8.000000000000001e-06,
196
- "loss": 0.6359,
197
  "step": 20
198
  },
199
  {
200
  "epoch": 0.013021237017516665,
201
- "grad_norm": 0.21558189392089844,
202
  "learning_rate": 7.2000000000000005e-06,
203
- "loss": 0.8966,
204
  "step": 21
205
  },
206
  {
207
  "epoch": 0.013021237017516665,
208
- "eval_loss": 0.9336352348327637,
209
- "eval_runtime": 47.0196,
210
- "eval_samples_per_second": 2.127,
211
- "eval_steps_per_second": 2.127,
212
  "step": 21
213
  },
214
  {
215
  "epoch": 0.013641295923112695,
216
- "grad_norm": 0.2076757252216339,
217
  "learning_rate": 6.4000000000000006e-06,
218
- "loss": 0.8719,
219
  "step": 22
220
  },
221
  {
222
  "epoch": 0.014261354828708727,
223
- "grad_norm": 0.3216860890388489,
224
  "learning_rate": 5.600000000000001e-06,
225
- "loss": 1.0026,
226
  "step": 23
227
  },
228
  {
229
  "epoch": 0.014881413734304759,
230
- "grad_norm": 0.2905324697494507,
231
  "learning_rate": 4.800000000000001e-06,
232
- "loss": 0.7967,
233
  "step": 24
234
  },
235
  {
236
  "epoch": 0.014881413734304759,
237
- "eval_loss": 0.9275597333908081,
238
- "eval_runtime": 47.023,
239
- "eval_samples_per_second": 2.127,
240
- "eval_steps_per_second": 2.127,
241
  "step": 24
242
  },
243
  {
244
  "epoch": 0.01550147263990079,
245
- "grad_norm": 0.2005520761013031,
246
  "learning_rate": 4.000000000000001e-06,
247
- "loss": 0.8374,
248
  "step": 25
249
  },
250
  {
251
  "epoch": 0.016121531545496823,
252
- "grad_norm": 0.1872091293334961,
253
  "learning_rate": 3.2000000000000003e-06,
254
- "loss": 0.7055,
255
  "step": 26
256
  },
257
  {
258
  "epoch": 0.016741590451092855,
259
- "grad_norm": 0.1885146200656891,
260
  "learning_rate": 2.4000000000000003e-06,
261
- "loss": 0.7527,
262
  "step": 27
263
  },
264
  {
265
  "epoch": 0.016741590451092855,
266
- "eval_loss": 0.9237930774688721,
267
- "eval_runtime": 47.0424,
268
- "eval_samples_per_second": 2.126,
269
- "eval_steps_per_second": 2.126,
270
  "step": 27
271
  },
272
  {
273
  "epoch": 0.017361649356688887,
274
- "grad_norm": 0.20483697950839996,
275
  "learning_rate": 1.6000000000000001e-06,
276
- "loss": 0.6491,
277
  "step": 28
278
  },
279
  {
280
  "epoch": 0.01798170826228492,
281
- "grad_norm": 0.23465946316719055,
282
  "learning_rate": 8.000000000000001e-07,
283
- "loss": 0.9221,
284
  "step": 29
285
  },
286
  {
287
  "epoch": 0.01860176716788095,
288
- "grad_norm": 0.2271672636270523,
289
  "learning_rate": 0.0,
290
- "loss": 0.6819,
291
  "step": 30
292
  },
293
  {
294
  "epoch": 0.01860176716788095,
295
- "eval_loss": 0.9222295880317688,
296
- "eval_runtime": 47.4088,
297
- "eval_samples_per_second": 2.109,
298
- "eval_steps_per_second": 2.109,
299
  "step": 30
300
  }
301
  ],
@@ -316,7 +316,7 @@
316
  "attributes": {}
317
  }
318
  },
319
- "total_flos": 6.35700342245376e+16,
320
  "train_batch_size": 2,
321
  "trial_name": null,
322
  "trial_params": null
 
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
+ "grad_norm": 0.5875879526138306,
14
  "learning_rate": 4.000000000000001e-06,
15
+ "loss": 1.179,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
+ "grad_norm": 0.2559925317764282,
21
  "learning_rate": 8.000000000000001e-06,
22
+ "loss": 0.9296,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
+ "grad_norm": 0.27975356578826904,
28
  "learning_rate": 1.2e-05,
29
+ "loss": 0.8212,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
+ "eval_loss": 1.0227885246276855,
35
+ "eval_runtime": 47.9985,
36
+ "eval_samples_per_second": 2.083,
37
+ "eval_steps_per_second": 2.083,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
+ "grad_norm": 0.4077853858470917,
43
  "learning_rate": 1.6000000000000003e-05,
44
+ "loss": 0.9553,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
+ "grad_norm": 0.3481099307537079,
50
  "learning_rate": 2e-05,
51
+ "loss": 0.8071,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
+ "grad_norm": 0.33538541197776794,
57
  "learning_rate": 1.9200000000000003e-05,
58
+ "loss": 0.9046,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
+ "eval_loss": 1.0122885704040527,
64
+ "eval_runtime": 48.1202,
65
+ "eval_samples_per_second": 2.078,
66
+ "eval_steps_per_second": 2.078,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
+ "grad_norm": 0.2848486602306366,
72
  "learning_rate": 1.8400000000000003e-05,
73
+ "loss": 0.8963,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
+ "grad_norm": 0.29836800694465637,
79
  "learning_rate": 1.76e-05,
80
+ "loss": 0.9869,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
+ "grad_norm": 0.22438423335552216,
86
  "learning_rate": 1.6800000000000002e-05,
87
+ "loss": 0.8704,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
+ "eval_loss": 0.996995210647583,
93
+ "eval_runtime": 48.1511,
94
+ "eval_samples_per_second": 2.077,
95
+ "eval_steps_per_second": 2.077,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
+ "grad_norm": 0.26462429761886597,
101
  "learning_rate": 1.6000000000000003e-05,
102
+ "loss": 1.0479,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
+ "grad_norm": 0.2650766968727112,
108
  "learning_rate": 1.5200000000000002e-05,
109
+ "loss": 0.8352,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
+ "grad_norm": 0.3595620095729828,
115
  "learning_rate": 1.4400000000000001e-05,
116
+ "loss": 1.1297,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
+ "eval_loss": 0.9821679592132568,
122
+ "eval_runtime": 48.205,
123
+ "eval_samples_per_second": 2.074,
124
+ "eval_steps_per_second": 2.074,
125
  "step": 12
126
  },
127
  {
128
  "epoch": 0.008060765772748411,
129
+ "grad_norm": 0.2854643166065216,
130
  "learning_rate": 1.3600000000000002e-05,
131
+ "loss": 0.9925,
132
  "step": 13
133
  },
134
  {
135
  "epoch": 0.008680824678344443,
136
+ "grad_norm": 0.21441830694675446,
137
  "learning_rate": 1.2800000000000001e-05,
138
+ "loss": 0.8125,
139
  "step": 14
140
  },
141
  {
142
  "epoch": 0.009300883583940475,
143
+ "grad_norm": 0.21920695900917053,
144
  "learning_rate": 1.2e-05,
145
+ "loss": 0.8518,
146
  "step": 15
147
  },
148
  {
149
  "epoch": 0.009300883583940475,
150
+ "eval_loss": 0.9693244695663452,
151
+ "eval_runtime": 47.8177,
152
+ "eval_samples_per_second": 2.091,
153
+ "eval_steps_per_second": 2.091,
154
  "step": 15
155
  },
156
  {
157
  "epoch": 0.009920942489536505,
158
+ "grad_norm": 0.20094920694828033,
159
  "learning_rate": 1.1200000000000001e-05,
160
+ "loss": 0.8003,
161
  "step": 16
162
  },
163
  {
164
  "epoch": 0.010541001395132537,
165
+ "grad_norm": 0.21595342457294464,
166
  "learning_rate": 1.04e-05,
167
+ "loss": 0.8492,
168
  "step": 17
169
  },
170
  {
171
  "epoch": 0.01116106030072857,
172
+ "grad_norm": 0.20084670186042786,
173
  "learning_rate": 9.600000000000001e-06,
174
+ "loss": 0.8238,
175
  "step": 18
176
  },
177
  {
178
  "epoch": 0.01116106030072857,
179
+ "eval_loss": 0.9594716429710388,
180
+ "eval_runtime": 47.8775,
181
+ "eval_samples_per_second": 2.089,
182
+ "eval_steps_per_second": 2.089,
183
  "step": 18
184
  },
185
  {
186
  "epoch": 0.011781119206324601,
187
+ "grad_norm": 0.22359193861484528,
188
  "learning_rate": 8.8e-06,
189
+ "loss": 0.8032,
190
  "step": 19
191
  },
192
  {
193
  "epoch": 0.012401178111920633,
194
+ "grad_norm": 0.16142494976520538,
195
  "learning_rate": 8.000000000000001e-06,
196
+ "loss": 0.6508,
197
  "step": 20
198
  },
199
  {
200
  "epoch": 0.013021237017516665,
201
+ "grad_norm": 0.2063056379556656,
202
  "learning_rate": 7.2000000000000005e-06,
203
+ "loss": 0.9097,
204
  "step": 21
205
  },
206
  {
207
  "epoch": 0.013021237017516665,
208
+ "eval_loss": 0.9518673419952393,
209
+ "eval_runtime": 48.16,
210
+ "eval_samples_per_second": 2.076,
211
+ "eval_steps_per_second": 2.076,
212
  "step": 21
213
  },
214
  {
215
  "epoch": 0.013641295923112695,
216
+ "grad_norm": 0.1905093491077423,
217
  "learning_rate": 6.4000000000000006e-06,
218
+ "loss": 0.8873,
219
  "step": 22
220
  },
221
  {
222
  "epoch": 0.014261354828708727,
223
+ "grad_norm": 0.28930607438087463,
224
  "learning_rate": 5.600000000000001e-06,
225
+ "loss": 1.0172,
226
  "step": 23
227
  },
228
  {
229
  "epoch": 0.014881413734304759,
230
+ "grad_norm": 0.2751776874065399,
231
  "learning_rate": 4.800000000000001e-06,
232
+ "loss": 0.8155,
233
  "step": 24
234
  },
235
  {
236
  "epoch": 0.014881413734304759,
237
+ "eval_loss": 0.946439266204834,
238
+ "eval_runtime": 47.9311,
239
+ "eval_samples_per_second": 2.086,
240
+ "eval_steps_per_second": 2.086,
241
  "step": 24
242
  },
243
  {
244
  "epoch": 0.01550147263990079,
245
+ "grad_norm": 0.20159322023391724,
246
  "learning_rate": 4.000000000000001e-06,
247
+ "loss": 0.8539,
248
  "step": 25
249
  },
250
  {
251
  "epoch": 0.016121531545496823,
252
+ "grad_norm": 0.4606916010379791,
253
  "learning_rate": 3.2000000000000003e-06,
254
+ "loss": 0.7225,
255
  "step": 26
256
  },
257
  {
258
  "epoch": 0.016741590451092855,
259
+ "grad_norm": 0.18102441728115082,
260
  "learning_rate": 2.4000000000000003e-06,
261
+ "loss": 0.7702,
262
  "step": 27
263
  },
264
  {
265
  "epoch": 0.016741590451092855,
266
+ "eval_loss": 0.9430135488510132,
267
+ "eval_runtime": 47.8428,
268
+ "eval_samples_per_second": 2.09,
269
+ "eval_steps_per_second": 2.09,
270
  "step": 27
271
  },
272
  {
273
  "epoch": 0.017361649356688887,
274
+ "grad_norm": 0.2126253843307495,
275
  "learning_rate": 1.6000000000000001e-06,
276
+ "loss": 0.6669,
277
  "step": 28
278
  },
279
  {
280
  "epoch": 0.01798170826228492,
281
+ "grad_norm": 0.24684152007102966,
282
  "learning_rate": 8.000000000000001e-07,
283
+ "loss": 0.9388,
284
  "step": 29
285
  },
286
  {
287
  "epoch": 0.01860176716788095,
288
+ "grad_norm": 0.2677140235900879,
289
  "learning_rate": 0.0,
290
+ "loss": 0.7054,
291
  "step": 30
292
  },
293
  {
294
  "epoch": 0.01860176716788095,
295
+ "eval_loss": 0.9416863918304443,
296
+ "eval_runtime": 47.8668,
297
+ "eval_samples_per_second": 2.089,
298
+ "eval_steps_per_second": 2.089,
299
  "step": 30
300
  }
301
  ],
 
316
  "attributes": {}
317
  }
318
  },
319
+ "total_flos": 6.450802528167936e+16,
320
  "train_batch_size": 2,
321
  "trial_name": null,
322
  "trial_params": null
checkpoint-30/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:31008ae8f5fa3525c3b5855165b7a47a0cdb47e438b843022563680f47a27521
3
  size 5688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7726c9626d0a377a638b6674780a9e8d32aed5a1fc017243f1823b069fc15ac
3
  size 5688