irishprancer commited on
Commit
89335e9
·
verified ·
1 Parent(s): ae2af9d

Training in progress, step 300, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -22,8 +22,8 @@
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
25
- "q_proj",
26
- "k_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
25
+ "k_proj",
26
+ "q_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4afd5de1ba40a667d318a69f8ca1b9640abbb1894e9a31f968505bcc344b6854
3
  size 527048968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dffb58ead3a8261ad808ac56947c3cd3660dde80949934a70ee4007ecf573495
3
  size 527048968
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44b589f7ce6dc8df86706635a4a048b369cefac35fb9420912505e001b6668a2
3
  size 1054135994
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6d510ad6a7aadf1acfd0a7799fb32b8ccdd1d51e237076309c4a4f363e0b0c8
3
  size 1054135994
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:afdcd158786234085082ce38b1824c51dd8c72881220443fc2d1c6f4e031a983
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c84fe26b84a1b82bdd32297a5506f7b0ec5d1b51b65a2f281da22aea31edccd2
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cf3f988e8fed2daa2e801eb1f19b681872781cf57f0fb7b896e859a12cfe2bb
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f4c00d522bdde510099aafe1617b13d114dce17a17b44e05876f016f4e4d7af
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,125 +1,238 @@
1
  {
2
- "best_metric": 0.7960079908370972,
3
- "best_model_checkpoint": "./output/checkpoint-150",
4
- "epoch": 6.521739130434782,
5
  "eval_steps": 150,
6
- "global_step": 150,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.43478260869565216,
13
- "grad_norm": 1.5021412372589111,
14
  "learning_rate": 3e-06,
15
- "loss": 0.9061,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.8695652173913043,
20
- "grad_norm": 1.687071442604065,
21
  "learning_rate": 6e-06,
22
- "loss": 0.9026,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 1.3043478260869565,
27
- "grad_norm": 1.7296651601791382,
28
  "learning_rate": 9e-06,
29
- "loss": 0.9003,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 1.7391304347826086,
34
- "grad_norm": 1.4536513090133667,
35
  "learning_rate": 1.2e-05,
36
- "loss": 0.9093,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 2.1739130434782608,
41
- "grad_norm": 1.351776361465454,
42
  "learning_rate": 1.5e-05,
43
- "loss": 0.8358,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 2.608695652173913,
48
- "grad_norm": 2.031055212020874,
49
  "learning_rate": 1.8e-05,
50
- "loss": 0.8894,
51
  "step": 60
52
  },
53
  {
54
  "epoch": 3.0434782608695654,
55
- "grad_norm": 1.4920436143875122,
56
  "learning_rate": 2.1e-05,
57
- "loss": 0.8915,
58
  "step": 70
59
  },
60
  {
61
  "epoch": 3.4782608695652173,
62
- "grad_norm": 1.729347586631775,
63
  "learning_rate": 2.4e-05,
64
- "loss": 0.8237,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 3.9130434782608696,
69
- "grad_norm": 1.4314346313476562,
70
  "learning_rate": 2.7000000000000002e-05,
71
  "loss": 0.8529,
72
  "step": 90
73
  },
74
  {
75
  "epoch": 4.3478260869565215,
76
- "grad_norm": 1.3657989501953125,
77
  "learning_rate": 3e-05,
78
  "loss": 0.8649,
79
  "step": 100
80
  },
81
  {
82
  "epoch": 4.782608695652174,
83
- "grad_norm": 2.1989877223968506,
84
  "learning_rate": 2.999999702723963e-05,
85
  "loss": 0.8227,
86
  "step": 110
87
  },
88
  {
89
  "epoch": 5.217391304347826,
90
- "grad_norm": 1.0724585056304932,
91
  "learning_rate": 2.9999988108959687e-05,
92
- "loss": 0.7655,
93
  "step": 120
94
  },
95
  {
96
  "epoch": 5.6521739130434785,
97
- "grad_norm": 1.5675039291381836,
98
  "learning_rate": 2.9999973245163716e-05,
99
  "loss": 0.7413,
100
  "step": 130
101
  },
102
  {
103
  "epoch": 6.086956521739131,
104
- "grad_norm": 1.9068934917449951,
105
  "learning_rate": 2.99999524358576e-05,
106
- "loss": 0.7653,
107
  "step": 140
108
  },
109
  {
110
  "epoch": 6.521739130434782,
111
- "grad_norm": 1.121341586112976,
112
  "learning_rate": 2.9999925681049593e-05,
113
  "loss": 0.7855,
114
  "step": 150
115
  },
116
  {
117
  "epoch": 6.521739130434782,
118
- "eval_loss": 0.7960079908370972,
119
- "eval_runtime": 0.4677,
120
- "eval_samples_per_second": 21.382,
121
- "eval_steps_per_second": 21.382,
122
  "step": 150
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
123
  }
124
  ],
125
  "logging_steps": 10,
@@ -139,7 +252,7 @@
139
  "attributes": {}
140
  }
141
  },
142
- "total_flos": 3894839614291968.0,
143
  "train_batch_size": 4,
144
  "trial_name": null,
145
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.7312111854553223,
3
+ "best_model_checkpoint": "./output/checkpoint-300",
4
+ "epoch": 13.043478260869565,
5
  "eval_steps": 150,
6
+ "global_step": 300,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.43478260869565216,
13
+ "grad_norm": 1.5021647214889526,
14
  "learning_rate": 3e-06,
15
+ "loss": 0.9064,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.8695652173913043,
20
+ "grad_norm": 1.6793277263641357,
21
  "learning_rate": 6e-06,
22
+ "loss": 0.9024,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 1.3043478260869565,
27
+ "grad_norm": 1.7296056747436523,
28
  "learning_rate": 9e-06,
29
+ "loss": 0.9,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 1.7391304347826086,
34
+ "grad_norm": 1.4458602666854858,
35
  "learning_rate": 1.2e-05,
36
+ "loss": 0.9086,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 2.1739130434782608,
41
+ "grad_norm": 1.3516699075698853,
42
  "learning_rate": 1.5e-05,
43
+ "loss": 0.8359,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 2.608695652173913,
48
+ "grad_norm": 2.03114914894104,
49
  "learning_rate": 1.8e-05,
50
+ "loss": 0.8892,
51
  "step": 60
52
  },
53
  {
54
  "epoch": 3.0434782608695654,
55
+ "grad_norm": 1.4920729398727417,
56
  "learning_rate": 2.1e-05,
57
+ "loss": 0.8913,
58
  "step": 70
59
  },
60
  {
61
  "epoch": 3.4782608695652173,
62
+ "grad_norm": 1.7293280363082886,
63
  "learning_rate": 2.4e-05,
64
+ "loss": 0.8234,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 3.9130434782608696,
69
+ "grad_norm": 1.4240705966949463,
70
  "learning_rate": 2.7000000000000002e-05,
71
  "loss": 0.8529,
72
  "step": 90
73
  },
74
  {
75
  "epoch": 4.3478260869565215,
76
+ "grad_norm": 1.3656291961669922,
77
  "learning_rate": 3e-05,
78
  "loss": 0.8649,
79
  "step": 100
80
  },
81
  {
82
  "epoch": 4.782608695652174,
83
+ "grad_norm": 2.198136806488037,
84
  "learning_rate": 2.999999702723963e-05,
85
  "loss": 0.8227,
86
  "step": 110
87
  },
88
  {
89
  "epoch": 5.217391304347826,
90
+ "grad_norm": 1.0726988315582275,
91
  "learning_rate": 2.9999988108959687e-05,
92
+ "loss": 0.765,
93
  "step": 120
94
  },
95
  {
96
  "epoch": 5.6521739130434785,
97
+ "grad_norm": 1.5672078132629395,
98
  "learning_rate": 2.9999973245163716e-05,
99
  "loss": 0.7413,
100
  "step": 130
101
  },
102
  {
103
  "epoch": 6.086956521739131,
104
+ "grad_norm": 1.9071016311645508,
105
  "learning_rate": 2.99999524358576e-05,
106
+ "loss": 0.7655,
107
  "step": 140
108
  },
109
  {
110
  "epoch": 6.521739130434782,
111
+ "grad_norm": 1.1219961643218994,
112
  "learning_rate": 2.9999925681049593e-05,
113
  "loss": 0.7855,
114
  "step": 150
115
  },
116
  {
117
  "epoch": 6.521739130434782,
118
+ "eval_loss": 0.7960985898971558,
119
+ "eval_runtime": 0.5137,
120
+ "eval_samples_per_second": 19.465,
121
+ "eval_steps_per_second": 19.465,
122
  "step": 150
123
+ },
124
+ {
125
+ "epoch": 6.956521739130435,
126
+ "grad_norm": 1.5404783487319946,
127
+ "learning_rate": 2.9999892980750297e-05,
128
+ "loss": 0.6588,
129
+ "step": 160
130
+ },
131
+ {
132
+ "epoch": 7.391304347826087,
133
+ "grad_norm": 1.3464210033416748,
134
+ "learning_rate": 2.9999854334972675e-05,
135
+ "loss": 0.739,
136
+ "step": 170
137
+ },
138
+ {
139
+ "epoch": 7.826086956521739,
140
+ "grad_norm": 1.7264870405197144,
141
+ "learning_rate": 2.999980974373204e-05,
142
+ "loss": 0.7292,
143
+ "step": 180
144
+ },
145
+ {
146
+ "epoch": 8.26086956521739,
147
+ "grad_norm": 1.5393368005752563,
148
+ "learning_rate": 2.9999759207046075e-05,
149
+ "loss": 0.6244,
150
+ "step": 190
151
+ },
152
+ {
153
+ "epoch": 8.695652173913043,
154
+ "grad_norm": 1.7342770099639893,
155
+ "learning_rate": 2.9999702724934804e-05,
156
+ "loss": 0.6764,
157
+ "step": 200
158
+ },
159
+ {
160
+ "epoch": 9.130434782608695,
161
+ "grad_norm": 1.0416548252105713,
162
+ "learning_rate": 2.999964029742062e-05,
163
+ "loss": 0.6521,
164
+ "step": 210
165
+ },
166
+ {
167
+ "epoch": 9.565217391304348,
168
+ "grad_norm": 1.2193585634231567,
169
+ "learning_rate": 2.9999571924528263e-05,
170
+ "loss": 0.5592,
171
+ "step": 220
172
+ },
173
+ {
174
+ "epoch": 10.0,
175
+ "grad_norm": 1.5286610126495361,
176
+ "learning_rate": 2.9999497606284837e-05,
177
+ "loss": 0.756,
178
+ "step": 230
179
+ },
180
+ {
181
+ "epoch": 10.434782608695652,
182
+ "grad_norm": 1.427451252937317,
183
+ "learning_rate": 2.9999417342719796e-05,
184
+ "loss": 0.7118,
185
+ "step": 240
186
+ },
187
+ {
188
+ "epoch": 10.869565217391305,
189
+ "grad_norm": 0.9794074296951294,
190
+ "learning_rate": 2.9999331133864956e-05,
191
+ "loss": 0.5897,
192
+ "step": 250
193
+ },
194
+ {
195
+ "epoch": 11.304347826086957,
196
+ "grad_norm": 1.194082260131836,
197
+ "learning_rate": 2.9999238979754485e-05,
198
+ "loss": 0.6548,
199
+ "step": 260
200
+ },
201
+ {
202
+ "epoch": 11.73913043478261,
203
+ "grad_norm": 1.0489437580108643,
204
+ "learning_rate": 2.999914088042492e-05,
205
+ "loss": 0.6474,
206
+ "step": 270
207
+ },
208
+ {
209
+ "epoch": 12.173913043478262,
210
+ "grad_norm": 1.3105570077896118,
211
+ "learning_rate": 2.9999036835915132e-05,
212
+ "loss": 0.5936,
213
+ "step": 280
214
+ },
215
+ {
216
+ "epoch": 12.608695652173914,
217
+ "grad_norm": 1.0830979347229004,
218
+ "learning_rate": 2.9998926846266365e-05,
219
+ "loss": 0.6328,
220
+ "step": 290
221
+ },
222
+ {
223
+ "epoch": 13.043478260869565,
224
+ "grad_norm": 1.3943493366241455,
225
+ "learning_rate": 2.9998810911522213e-05,
226
+ "loss": 0.5806,
227
+ "step": 300
228
+ },
229
+ {
230
+ "epoch": 13.043478260869565,
231
+ "eval_loss": 0.7312111854553223,
232
+ "eval_runtime": 0.6239,
233
+ "eval_samples_per_second": 16.028,
234
+ "eval_steps_per_second": 16.028,
235
+ "step": 300
236
  }
237
  ],
238
  "logging_steps": 10,
 
252
  "attributes": {}
253
  }
254
  },
255
+ "total_flos": 7724643094462464.0,
256
  "train_batch_size": 4,
257
  "trial_name": null,
258
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4e26a07b48ab9f4eeb67e7d951dfa1f71b1fc6e6c955372cec1a3c50fa8b9a2
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c84fd1341bca21f546fb34f0984c2737ed54e8dbbe2ce01666781675b951db1
3
  size 5368