hajimemat commited on
Commit
e9b356b
·
verified ·
1 Parent(s): 9ede0bd

Training in progress, step 300, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a0ddf70e3b717c4942eb29e5cb233383ecd1afc20c58cc985636fa0e06d5b136
3
  size 194563400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:145ec1241d02e5ad7c0c5963a613fa26a9f50d41b13b54cf4e31c3a3fd967930
3
  size 194563400
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8b1d2da135690fff7c0e120e786351cf9750b8ea31f819ca37b532af0ea60ef3
3
- size 99240837
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fe898fa16b6709fa0c7a7e7fc599faecd85a247cbb9d57e586320b5d424d3ee
3
+ size 99241221
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:11ed94202eb6e3eaeb8f032cfd9fd46e9b4657a59638d69479f047c360d252a9
3
- size 14709
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecb3d01905ba3db4c3065c85697c168d9e32937350baf828e8a0685920ef9c35
3
+ size 14645
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:95f3a67d47428d0f6084a0c2e68b54a6b89dcabf900532b585b519c3b42aa7fc
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca4f7c0edc6c0f32a409cb7bdaa5a2587da793fda2041cd5c6f57585074491a5
3
  size 1465
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.4764031561709096,
5
  "eval_steps": 500,
6
- "global_step": 200,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -147,6 +147,76 @@
147
  "learning_rate": 5.6174528158664096e-05,
148
  "loss": 1.0359,
149
  "step": 200
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
150
  }
151
  ],
152
  "logging_steps": 10,
@@ -166,7 +236,7 @@
166
  "attributes": {}
167
  }
168
  },
169
- "total_flos": 1.691246653341696e+17,
170
  "train_batch_size": 2,
171
  "trial_name": null,
172
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.7146047342563644,
5
  "eval_steps": 500,
6
+ "global_step": 300,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
147
  "learning_rate": 5.6174528158664096e-05,
148
  "loss": 1.0359,
149
  "step": 200
150
+ },
151
+ {
152
+ "epoch": 0.5002233139794551,
153
+ "grad_norm": 0.11049015820026398,
154
+ "learning_rate": 5.232054001413941e-05,
155
+ "loss": 1.03,
156
+ "step": 210
157
+ },
158
+ {
159
+ "epoch": 0.5240434717880006,
160
+ "grad_norm": 0.11212771385908127,
161
+ "learning_rate": 4.845266449876645e-05,
162
+ "loss": 1.0016,
163
+ "step": 220
164
+ },
165
+ {
166
+ "epoch": 0.5478636295965461,
167
+ "grad_norm": 0.12015224993228912,
168
+ "learning_rate": 4.4594049078802925e-05,
169
+ "loss": 1.0311,
170
+ "step": 230
171
+ },
172
+ {
173
+ "epoch": 0.5716837874050915,
174
+ "grad_norm": 0.11489170789718628,
175
+ "learning_rate": 4.076778580306999e-05,
176
+ "loss": 1.0325,
177
+ "step": 240
178
+ },
179
+ {
180
+ "epoch": 0.595503945213637,
181
+ "grad_norm": 0.11050307750701904,
182
+ "learning_rate": 3.69967731075996e-05,
183
+ "loss": 0.9939,
184
+ "step": 250
185
+ },
186
+ {
187
+ "epoch": 0.6193241030221825,
188
+ "grad_norm": 0.11331106722354889,
189
+ "learning_rate": 3.330357877896577e-05,
190
+ "loss": 1.0314,
191
+ "step": 260
192
+ },
193
+ {
194
+ "epoch": 0.643144260830728,
195
+ "grad_norm": 0.1162668839097023,
196
+ "learning_rate": 2.9710304896401802e-05,
197
+ "loss": 0.9822,
198
+ "step": 270
199
+ },
200
+ {
201
+ "epoch": 0.6669644186392735,
202
+ "grad_norm": 0.11404977738857269,
203
+ "learning_rate": 2.6238455560962884e-05,
204
+ "loss": 0.9947,
205
+ "step": 280
206
+ },
207
+ {
208
+ "epoch": 0.690784576447819,
209
+ "grad_norm": 0.11031963676214218,
210
+ "learning_rate": 2.2908808203314635e-05,
211
+ "loss": 0.9889,
212
+ "step": 290
213
+ },
214
+ {
215
+ "epoch": 0.7146047342563644,
216
+ "grad_norm": 0.11861108988523483,
217
+ "learning_rate": 1.9741289240311755e-05,
218
+ "loss": 1.0303,
219
+ "step": 300
220
  }
221
  ],
222
  "logging_steps": 10,
 
236
  "attributes": {}
237
  }
238
  },
239
+ "total_flos": 2.536869980012544e+17,
240
  "train_batch_size": 2,
241
  "trial_name": null,
242
  "trial_params": null