rootxhacker commited on
Commit
66325c3
·
verified ·
1 Parent(s): b21cd97

Training in progress, step 1500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f02ec6b82a831f2902a910cf8fafe953154448b3b45acb102d57d8391c23b96
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45253cd6a22264fcd7ca6604e87717a83e19ed0d2d4435aba88387d4537d6529
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cbe39982c4bebe4e0b3f0f76afe1c485909ad223aac0a865ef49d364896c3cff
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b17ffeb44d19aff3ac9c819aa738ee5ab85c93083a940599d75b2a7ffda4f79
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c326c13c75092245145edf95faedf4bbc2e6d95bbc78269d58ccc6219fd83b99
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ad00466a0acdb40237dae4c0b4283060a706960c994c5977eb55d8a845c8d60
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:31bb3f2abd407f62d555fa53e113286139c6abb36d00dab486b238b592c7d71a
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97a66f532d27e837c455454684371b16caffae096e4b153335efaed54e929716
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2229a15bdf42c8dbae70a65f18e33cde5c16363745edce9ae10c3d4c94d5ed8e
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8167a7a628c5d6d5061531d993a99d8970329d2f198daf9e2bd5698793632aed
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9d88837be00a1ec9be1702d7436c136fd03c05aff969a737326a957a88901e17
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8648ee281a6907a159c6d2a012bfd17e90caa435b3ae05a6accbf2bacc5a6ed9
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 1000,
3
- "best_metric": 1.6695575714111328,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-1000",
5
- "epoch": 0.07691716021844473,
6
  "eval_steps": 250,
7
- "global_step": 1000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -180,6 +180,92 @@
180
  "eval_samples_per_second": 59.406,
181
  "eval_steps_per_second": 14.851,
182
  "step": 1000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
183
  }
184
  ],
185
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 1500,
3
+ "best_metric": 1.6279098987579346,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-1500",
5
+ "epoch": 0.1153757403276671,
6
  "eval_steps": 250,
7
+ "global_step": 1500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
180
  "eval_samples_per_second": 59.406,
181
  "eval_steps_per_second": 14.851,
182
  "step": 1000
183
+ },
184
+ {
185
+ "epoch": 0.08076301822936698,
186
+ "grad_norm": 2.8161075115203857,
187
+ "learning_rate": 0.00019716905176220034,
188
+ "loss": 1.6006,
189
+ "step": 1050
190
+ },
191
+ {
192
+ "epoch": 0.0846088762402892,
193
+ "grad_norm": 1.2744417190551758,
194
+ "learning_rate": 0.00019690933174038388,
195
+ "loss": 1.6306,
196
+ "step": 1100
197
+ },
198
+ {
199
+ "epoch": 0.08845473425121145,
200
+ "grad_norm": 1.0525975227355957,
201
+ "learning_rate": 0.0001966496117185674,
202
+ "loss": 1.6414,
203
+ "step": 1150
204
+ },
205
+ {
206
+ "epoch": 0.09230059226213368,
207
+ "grad_norm": 1.7405261993408203,
208
+ "learning_rate": 0.0001963898916967509,
209
+ "loss": 1.5997,
210
+ "step": 1200
211
+ },
212
+ {
213
+ "epoch": 0.09614645027305592,
214
+ "grad_norm": 2.664368152618408,
215
+ "learning_rate": 0.00019613017167493445,
216
+ "loss": 1.6865,
217
+ "step": 1250
218
+ },
219
+ {
220
+ "epoch": 0.09614645027305592,
221
+ "eval_loss": 1.653170108795166,
222
+ "eval_runtime": 16.8496,
223
+ "eval_samples_per_second": 59.348,
224
+ "eval_steps_per_second": 14.837,
225
+ "step": 1250
226
+ },
227
+ {
228
+ "epoch": 0.09999230828397816,
229
+ "grad_norm": 2.0688107013702393,
230
+ "learning_rate": 0.00019587045165311796,
231
+ "loss": 1.6575,
232
+ "step": 1300
233
+ },
234
+ {
235
+ "epoch": 0.10383816629490039,
236
+ "grad_norm": 0.8756892085075378,
237
+ "learning_rate": 0.00019561073163130145,
238
+ "loss": 1.6426,
239
+ "step": 1350
240
+ },
241
+ {
242
+ "epoch": 0.10768402430582263,
243
+ "grad_norm": 1.398770809173584,
244
+ "learning_rate": 0.000195351011609485,
245
+ "loss": 1.5807,
246
+ "step": 1400
247
+ },
248
+ {
249
+ "epoch": 0.11152988231674486,
250
+ "grad_norm": 2.2140140533447266,
251
+ "learning_rate": 0.0001950912915876685,
252
+ "loss": 1.5556,
253
+ "step": 1450
254
+ },
255
+ {
256
+ "epoch": 0.1153757403276671,
257
+ "grad_norm": 1.9329601526260376,
258
+ "learning_rate": 0.000194831571565852,
259
+ "loss": 1.6597,
260
+ "step": 1500
261
+ },
262
+ {
263
+ "epoch": 0.1153757403276671,
264
+ "eval_loss": 1.6279098987579346,
265
+ "eval_runtime": 16.9897,
266
+ "eval_samples_per_second": 58.859,
267
+ "eval_steps_per_second": 14.715,
268
+ "step": 1500
269
  }
270
  ],
271
  "logging_steps": 50,