Kei5uke commited on
Commit
9b06f37
·
verified ·
1 Parent(s): aeac2ad

Training in progress, step 30, checkpoint

Browse files
checkpoint-30/adapter_config.json CHANGED
@@ -23,13 +23,13 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "k_proj",
27
- "o_proj",
28
  "v_proj",
29
- "gate_proj",
 
30
  "up_proj",
31
- "down_proj",
32
- "q_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "down_proj",
 
27
  "v_proj",
28
+ "o_proj",
29
+ "q_proj",
30
  "up_proj",
31
+ "gate_proj",
32
+ "k_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
checkpoint-30/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a1a9e0b590ffdc0fac6d4895f608ffd2d0753ab395a71921639c2f9fd61ca545
3
  size 80013120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8b97773e04827e934022820bfa7fbd5bad93af3e2ed70e773e85740afe2cc6c
3
  size 80013120
checkpoint-30/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e7d2f51158926f839deaf2a2cb5ca95934615b5a99cad711ecdb2aab1e8f45f
3
  size 160284754
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:476729e13139d92a235311e1b8c488d8270f6b819b7960034a803a04f45d5278
3
  size 160284754
checkpoint-30/trainer_state.json CHANGED
@@ -10,292 +10,292 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
- "grad_norm": 0.18599237501621246,
14
  "learning_rate": 4.000000000000001e-06,
15
- "loss": 0.9319,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
- "grad_norm": 0.09086523950099945,
21
  "learning_rate": 8.000000000000001e-06,
22
- "loss": 0.6463,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
- "grad_norm": 0.09325356036424637,
28
  "learning_rate": 1.2e-05,
29
- "loss": 0.575,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
- "eval_loss": 0.7272388935089111,
35
- "eval_runtime": 33.9288,
36
- "eval_samples_per_second": 2.947,
37
- "eval_steps_per_second": 2.947,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
- "grad_norm": 0.09707680344581604,
43
  "learning_rate": 1.6000000000000003e-05,
44
- "loss": 0.7038,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
- "grad_norm": 0.10810782760381699,
50
  "learning_rate": 2e-05,
51
- "loss": 0.5471,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
- "grad_norm": 0.13030411303043365,
57
  "learning_rate": 1.9200000000000003e-05,
58
- "loss": 0.6735,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
- "eval_loss": 0.7256982922554016,
64
- "eval_runtime": 33.9528,
65
- "eval_samples_per_second": 2.945,
66
- "eval_steps_per_second": 2.945,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
- "grad_norm": 0.12336976826190948,
72
  "learning_rate": 1.8400000000000003e-05,
73
- "loss": 0.654,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
- "grad_norm": 0.16046489775180817,
79
  "learning_rate": 1.76e-05,
80
- "loss": 0.6301,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
- "grad_norm": 0.11589628458023071,
86
  "learning_rate": 1.6800000000000002e-05,
87
- "loss": 0.5326,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
- "eval_loss": 0.723076343536377,
93
- "eval_runtime": 34.5303,
94
- "eval_samples_per_second": 2.896,
95
- "eval_steps_per_second": 2.896,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
- "grad_norm": 0.12244553864002228,
101
  "learning_rate": 1.6000000000000003e-05,
102
- "loss": 0.8055,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
- "grad_norm": 0.10270956158638,
108
  "learning_rate": 1.5200000000000002e-05,
109
- "loss": 0.5492,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
- "grad_norm": 0.14439155161380768,
115
  "learning_rate": 1.4400000000000001e-05,
116
- "loss": 0.8579,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
- "eval_loss": 0.7205619215965271,
122
- "eval_runtime": 33.7297,
123
- "eval_samples_per_second": 2.965,
124
- "eval_steps_per_second": 2.965,
125
  "step": 12
126
  },
127
  {
128
  "epoch": 0.008060765772748411,
129
- "grad_norm": 0.1271936148405075,
130
  "learning_rate": 1.3600000000000002e-05,
131
- "loss": 0.6923,
132
  "step": 13
133
  },
134
  {
135
  "epoch": 0.008680824678344443,
136
- "grad_norm": 0.092640720307827,
137
  "learning_rate": 1.2800000000000001e-05,
138
- "loss": 0.5983,
139
  "step": 14
140
  },
141
  {
142
  "epoch": 0.009300883583940475,
143
- "grad_norm": 0.1161455512046814,
144
  "learning_rate": 1.2e-05,
145
- "loss": 0.6386,
146
  "step": 15
147
  },
148
  {
149
  "epoch": 0.009300883583940475,
150
- "eval_loss": 0.7178443074226379,
151
- "eval_runtime": 33.5878,
152
- "eval_samples_per_second": 2.977,
153
- "eval_steps_per_second": 2.977,
154
  "step": 15
155
  },
156
  {
157
  "epoch": 0.009920942489536505,
158
- "grad_norm": 0.09220079332590103,
159
  "learning_rate": 1.1200000000000001e-05,
160
- "loss": 0.4665,
161
  "step": 16
162
  },
163
  {
164
  "epoch": 0.010541001395132537,
165
- "grad_norm": 0.08713295310735703,
166
  "learning_rate": 1.04e-05,
167
- "loss": 0.6114,
168
  "step": 17
169
  },
170
  {
171
  "epoch": 0.01116106030072857,
172
- "grad_norm": 0.11913106590509415,
173
  "learning_rate": 9.600000000000001e-06,
174
- "loss": 0.5703,
175
  "step": 18
176
  },
177
  {
178
  "epoch": 0.01116106030072857,
179
- "eval_loss": 0.7152560949325562,
180
- "eval_runtime": 33.4806,
181
- "eval_samples_per_second": 2.987,
182
- "eval_steps_per_second": 2.987,
183
  "step": 18
184
  },
185
  {
186
  "epoch": 0.011781119206324601,
187
- "grad_norm": 0.11837327480316162,
188
  "learning_rate": 8.8e-06,
189
- "loss": 0.6049,
190
  "step": 19
191
  },
192
  {
193
  "epoch": 0.012401178111920633,
194
- "grad_norm": 0.08001928776502609,
195
  "learning_rate": 8.000000000000001e-06,
196
- "loss": 0.4691,
197
  "step": 20
198
  },
199
  {
200
  "epoch": 0.013021237017516665,
201
- "grad_norm": 0.11291633546352386,
202
  "learning_rate": 7.2000000000000005e-06,
203
- "loss": 0.6752,
204
  "step": 21
205
  },
206
  {
207
  "epoch": 0.013021237017516665,
208
- "eval_loss": 0.7128345370292664,
209
- "eval_runtime": 34.2622,
210
- "eval_samples_per_second": 2.919,
211
- "eval_steps_per_second": 2.919,
212
  "step": 21
213
  },
214
  {
215
  "epoch": 0.013641295923112695,
216
- "grad_norm": 0.1115325391292572,
217
  "learning_rate": 6.4000000000000006e-06,
218
- "loss": 0.5874,
219
  "step": 22
220
  },
221
  {
222
  "epoch": 0.014261354828708727,
223
- "grad_norm": 0.1305752545595169,
224
  "learning_rate": 5.600000000000001e-06,
225
- "loss": 0.7733,
226
  "step": 23
227
  },
228
  {
229
  "epoch": 0.014881413734304759,
230
- "grad_norm": 0.10779058188199997,
231
  "learning_rate": 4.800000000000001e-06,
232
- "loss": 0.6217,
233
  "step": 24
234
  },
235
  {
236
  "epoch": 0.014881413734304759,
237
- "eval_loss": 0.7108609080314636,
238
- "eval_runtime": 34.6034,
239
- "eval_samples_per_second": 2.89,
240
- "eval_steps_per_second": 2.89,
241
  "step": 24
242
  },
243
  {
244
  "epoch": 0.01550147263990079,
245
- "grad_norm": 0.11457020789384842,
246
  "learning_rate": 4.000000000000001e-06,
247
- "loss": 0.5838,
248
  "step": 25
249
  },
250
  {
251
  "epoch": 0.016121531545496823,
252
- "grad_norm": 0.1093108132481575,
253
  "learning_rate": 3.2000000000000003e-06,
254
- "loss": 0.504,
255
  "step": 26
256
  },
257
  {
258
  "epoch": 0.016741590451092855,
259
- "grad_norm": 0.11828811466693878,
260
  "learning_rate": 2.4000000000000003e-06,
261
- "loss": 0.5179,
262
  "step": 27
263
  },
264
  {
265
  "epoch": 0.016741590451092855,
266
- "eval_loss": 0.7094972133636475,
267
- "eval_runtime": 34.2275,
268
- "eval_samples_per_second": 2.922,
269
- "eval_steps_per_second": 2.922,
270
  "step": 27
271
  },
272
  {
273
  "epoch": 0.017361649356688887,
274
- "grad_norm": 0.09534204006195068,
275
  "learning_rate": 1.6000000000000001e-06,
276
- "loss": 0.4988,
277
  "step": 28
278
  },
279
  {
280
  "epoch": 0.01798170826228492,
281
- "grad_norm": 0.12317203730344772,
282
  "learning_rate": 8.000000000000001e-07,
283
- "loss": 0.6945,
284
  "step": 29
285
  },
286
  {
287
  "epoch": 0.01860176716788095,
288
- "grad_norm": 0.13696961104869843,
289
  "learning_rate": 0.0,
290
- "loss": 0.5378,
291
  "step": 30
292
  },
293
  {
294
  "epoch": 0.01860176716788095,
295
- "eval_loss": 0.7089178562164307,
296
- "eval_runtime": 33.3831,
297
- "eval_samples_per_second": 2.996,
298
- "eval_steps_per_second": 2.996,
299
  "step": 30
300
  }
301
  ],
@@ -316,7 +316,7 @@
316
  "attributes": {}
317
  }
318
  },
319
- "total_flos": 4.029866007950131e+16,
320
  "train_batch_size": 2,
321
  "trial_name": null,
322
  "trial_params": null
 
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
+ "grad_norm": 0.16878783702850342,
14
  "learning_rate": 4.000000000000001e-06,
15
+ "loss": 0.9729,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
+ "grad_norm": 0.0865061804652214,
21
  "learning_rate": 8.000000000000001e-06,
22
+ "loss": 0.6661,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
+ "grad_norm": 0.0920981764793396,
28
  "learning_rate": 1.2e-05,
29
+ "loss": 0.5954,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
+ "eval_loss": 0.762992262840271,
35
+ "eval_runtime": 33.3972,
36
+ "eval_samples_per_second": 2.994,
37
+ "eval_steps_per_second": 2.994,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
+ "grad_norm": 0.10451442003250122,
43
  "learning_rate": 1.6000000000000003e-05,
44
+ "loss": 0.7227,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
+ "grad_norm": 0.11351602524518967,
50
  "learning_rate": 2e-05,
51
+ "loss": 0.5719,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
+ "grad_norm": 0.12484906613826752,
57
  "learning_rate": 1.9200000000000003e-05,
58
+ "loss": 0.6972,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
+ "eval_loss": 0.7615777850151062,
64
+ "eval_runtime": 33.3804,
65
+ "eval_samples_per_second": 2.996,
66
+ "eval_steps_per_second": 2.996,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
+ "grad_norm": 0.11116621643304825,
72
  "learning_rate": 1.8400000000000003e-05,
73
+ "loss": 0.6789,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
+ "grad_norm": 0.16013608872890472,
79
  "learning_rate": 1.76e-05,
80
+ "loss": 0.6612,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
+ "grad_norm": 0.1211152970790863,
86
  "learning_rate": 1.6800000000000002e-05,
87
+ "loss": 0.5557,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
+ "eval_loss": 0.7586225867271423,
93
+ "eval_runtime": 33.7015,
94
+ "eval_samples_per_second": 2.967,
95
+ "eval_steps_per_second": 2.967,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
+ "grad_norm": 0.1299518495798111,
101
  "learning_rate": 1.6000000000000003e-05,
102
+ "loss": 0.8339,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
+ "grad_norm": 0.11476285755634308,
108
  "learning_rate": 1.5200000000000002e-05,
109
+ "loss": 0.5786,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
+ "grad_norm": 0.18253028392791748,
115
  "learning_rate": 1.4400000000000001e-05,
116
+ "loss": 0.8889,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
+ "eval_loss": 0.7548640966415405,
122
+ "eval_runtime": 34.763,
123
+ "eval_samples_per_second": 2.877,
124
+ "eval_steps_per_second": 2.877,
125
  "step": 12
126
  },
127
  {
128
  "epoch": 0.008060765772748411,
129
+ "grad_norm": 0.14364618062973022,
130
  "learning_rate": 1.3600000000000002e-05,
131
+ "loss": 0.7251,
132
  "step": 13
133
  },
134
  {
135
  "epoch": 0.008680824678344443,
136
+ "grad_norm": 0.11001864820718765,
137
  "learning_rate": 1.2800000000000001e-05,
138
+ "loss": 0.6209,
139
  "step": 14
140
  },
141
  {
142
  "epoch": 0.009300883583940475,
143
+ "grad_norm": 0.1239691898226738,
144
  "learning_rate": 1.2e-05,
145
+ "loss": 0.6641,
146
  "step": 15
147
  },
148
  {
149
  "epoch": 0.009300883583940475,
150
+ "eval_loss": 0.7505818009376526,
151
+ "eval_runtime": 34.0273,
152
+ "eval_samples_per_second": 2.939,
153
+ "eval_steps_per_second": 2.939,
154
  "step": 15
155
  },
156
  {
157
  "epoch": 0.009920942489536505,
158
+ "grad_norm": 0.11025531589984894,
159
  "learning_rate": 1.1200000000000001e-05,
160
+ "loss": 0.4851,
161
  "step": 16
162
  },
163
  {
164
  "epoch": 0.010541001395132537,
165
+ "grad_norm": 0.10391498357057571,
166
  "learning_rate": 1.04e-05,
167
+ "loss": 0.6274,
168
  "step": 17
169
  },
170
  {
171
  "epoch": 0.01116106030072857,
172
+ "grad_norm": 0.13879302144050598,
173
  "learning_rate": 9.600000000000001e-06,
174
+ "loss": 0.5931,
175
  "step": 18
176
  },
177
  {
178
  "epoch": 0.01116106030072857,
179
+ "eval_loss": 0.7464621067047119,
180
+ "eval_runtime": 34.3301,
181
+ "eval_samples_per_second": 2.913,
182
+ "eval_steps_per_second": 2.913,
183
  "step": 18
184
  },
185
  {
186
  "epoch": 0.011781119206324601,
187
+ "grad_norm": 0.14016444981098175,
188
  "learning_rate": 8.8e-06,
189
+ "loss": 0.6289,
190
  "step": 19
191
  },
192
  {
193
  "epoch": 0.012401178111920633,
194
+ "grad_norm": 0.11601278930902481,
195
  "learning_rate": 8.000000000000001e-06,
196
+ "loss": 0.4875,
197
  "step": 20
198
  },
199
  {
200
  "epoch": 0.013021237017516665,
201
+ "grad_norm": 0.13207638263702393,
202
  "learning_rate": 7.2000000000000005e-06,
203
+ "loss": 0.6942,
204
  "step": 21
205
  },
206
  {
207
  "epoch": 0.013021237017516665,
208
+ "eval_loss": 0.7427210807800293,
209
+ "eval_runtime": 34.6202,
210
+ "eval_samples_per_second": 2.888,
211
+ "eval_steps_per_second": 2.888,
212
  "step": 21
213
  },
214
  {
215
  "epoch": 0.013641295923112695,
216
+ "grad_norm": 0.12952734529972076,
217
  "learning_rate": 6.4000000000000006e-06,
218
+ "loss": 0.6095,
219
  "step": 22
220
  },
221
  {
222
  "epoch": 0.014261354828708727,
223
+ "grad_norm": 0.15325698256492615,
224
  "learning_rate": 5.600000000000001e-06,
225
+ "loss": 0.798,
226
  "step": 23
227
  },
228
  {
229
  "epoch": 0.014881413734304759,
230
+ "grad_norm": 0.12156686931848526,
231
  "learning_rate": 4.800000000000001e-06,
232
+ "loss": 0.6426,
233
  "step": 24
234
  },
235
  {
236
  "epoch": 0.014881413734304759,
237
+ "eval_loss": 0.7396314740180969,
238
+ "eval_runtime": 34.8471,
239
+ "eval_samples_per_second": 2.87,
240
+ "eval_steps_per_second": 2.87,
241
  "step": 24
242
  },
243
  {
244
  "epoch": 0.01550147263990079,
245
+ "grad_norm": 0.14465734362602234,
246
  "learning_rate": 4.000000000000001e-06,
247
+ "loss": 0.6037,
248
  "step": 25
249
  },
250
  {
251
  "epoch": 0.016121531545496823,
252
+ "grad_norm": 0.12176154553890228,
253
  "learning_rate": 3.2000000000000003e-06,
254
+ "loss": 0.5222,
255
  "step": 26
256
  },
257
  {
258
  "epoch": 0.016741590451092855,
259
+ "grad_norm": 0.13855800032615662,
260
  "learning_rate": 2.4000000000000003e-06,
261
+ "loss": 0.5386,
262
  "step": 27
263
  },
264
  {
265
  "epoch": 0.016741590451092855,
266
+ "eval_loss": 0.7376267910003662,
267
+ "eval_runtime": 34.7753,
268
+ "eval_samples_per_second": 2.876,
269
+ "eval_steps_per_second": 2.876,
270
  "step": 27
271
  },
272
  {
273
  "epoch": 0.017361649356688887,
274
+ "grad_norm": 0.1102200523018837,
275
  "learning_rate": 1.6000000000000001e-06,
276
+ "loss": 0.5182,
277
  "step": 28
278
  },
279
  {
280
  "epoch": 0.01798170826228492,
281
+ "grad_norm": 0.1387104094028473,
282
  "learning_rate": 8.000000000000001e-07,
283
+ "loss": 0.716,
284
  "step": 29
285
  },
286
  {
287
  "epoch": 0.01860176716788095,
288
+ "grad_norm": 0.16438902914524078,
289
  "learning_rate": 0.0,
290
+ "loss": 0.5635,
291
  "step": 30
292
  },
293
  {
294
  "epoch": 0.01860176716788095,
295
+ "eval_loss": 0.7367461323738098,
296
+ "eval_runtime": 34.8255,
297
+ "eval_samples_per_second": 2.871,
298
+ "eval_steps_per_second": 2.871,
299
  "step": 30
300
  }
301
  ],
 
316
  "attributes": {}
317
  }
318
  },
319
+ "total_flos": 4.068039818310451e+16,
320
  "train_batch_size": 2,
321
  "trial_name": null,
322
  "trial_params": null
checkpoint-30/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd165a72088a81e5441087b0ae01dd72f7f14c5dad6a2e407eebc961586b0fc2
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fba20505ee5f941c5085404aed8036910f9919d2bc7b2ba5ba4186849388b0c
3
  size 5752