rootxhacker commited on
Commit
c1a0e3b
·
verified ·
1 Parent(s): 92a3503

Training in progress, step 1500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74ea70ab5c294f3be223b52902ee924adb6f2a217f316a909c64447e3b17f8f8
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12b232b0278243d25903d8789b4dba3a6c48116f80cf44f29b224b3a9213825f
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa476b430a4a63248ac2bc81d08a889c28813dea25aaa9356f70340bb7c10542
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f87fd12b6e3592e8e2a065da27b2e13a73a88e79824ecd70f4ced521e1d64d89
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bd88c5d89e3266252b8212d40d27f742ba6d2099c1dc476b1334d403bca3f20
3
- size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42de46b2e4c5179833d46fe28b80f469c3d40255a80ec3a104ef7ad1fd3d8210
3
+ size 14180
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:23b493d7ad156dc53747ae8d013133956196b2f038f30fe21dd5039c094ada9d
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34bac16fbc99181cf3c30f05bcb7e648ad2c92eef3f334c8cbe294300c1d92e7
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7d109e68a9027ba37776e3e3b4c867f8592e84e6e4f33bcbb6b2e5ff20129448
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c205be8a1beb83fa60ead2e2dc0acbbf0de06ae533775758f9d9f2396e1a4003
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 1000,
3
- "best_metric": 4.709664344787598,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-1000",
5
- "epoch": 0.07691716021844473,
6
  "eval_steps": 250,
7
- "global_step": 1000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -180,6 +180,92 @@
180
  "eval_samples_per_second": 53.177,
181
  "eval_steps_per_second": 13.294,
182
  "step": 1000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
183
  }
184
  ],
185
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 1500,
3
+ "best_metric": 4.696172714233398,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-1500",
5
+ "epoch": 0.1153757403276671,
6
  "eval_steps": 250,
7
+ "global_step": 1500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
180
  "eval_samples_per_second": 53.177,
181
  "eval_steps_per_second": 13.294,
182
  "step": 1000
183
+ },
184
+ {
185
+ "epoch": 0.08076301822936698,
186
+ "grad_norm": 5.442048072814941,
187
+ "learning_rate": 0.00013514135509723907,
188
+ "loss": 4.6134,
189
+ "step": 1050
190
+ },
191
+ {
192
+ "epoch": 0.0846088762402892,
193
+ "grad_norm": 4.779583930969238,
194
+ "learning_rate": 0.00013488401161721995,
195
+ "loss": 4.7226,
196
+ "step": 1100
197
+ },
198
+ {
199
+ "epoch": 0.08845473425121145,
200
+ "grad_norm": 3.221238851547241,
201
+ "learning_rate": 0.0001346266681372008,
202
+ "loss": 4.6837,
203
+ "step": 1150
204
+ },
205
+ {
206
+ "epoch": 0.09230059226213368,
207
+ "grad_norm": 5.55983304977417,
208
+ "learning_rate": 0.0001343693246571817,
209
+ "loss": 4.672,
210
+ "step": 1200
211
+ },
212
+ {
213
+ "epoch": 0.09614645027305592,
214
+ "grad_norm": 6.964417934417725,
215
+ "learning_rate": 0.0001341119811771626,
216
+ "loss": 4.9043,
217
+ "step": 1250
218
+ },
219
+ {
220
+ "epoch": 0.09614645027305592,
221
+ "eval_loss": 4.7052001953125,
222
+ "eval_runtime": 18.9307,
223
+ "eval_samples_per_second": 52.824,
224
+ "eval_steps_per_second": 13.206,
225
+ "step": 1250
226
+ },
227
+ {
228
+ "epoch": 0.09999230828397816,
229
+ "grad_norm": 7.476005554199219,
230
+ "learning_rate": 0.00013385463769714347,
231
+ "loss": 4.7776,
232
+ "step": 1300
233
+ },
234
+ {
235
+ "epoch": 0.10383816629490039,
236
+ "grad_norm": 3.4916040897369385,
237
+ "learning_rate": 0.00013359729421712435,
238
+ "loss": 4.7738,
239
+ "step": 1350
240
+ },
241
+ {
242
+ "epoch": 0.10768402430582263,
243
+ "grad_norm": 4.028671741485596,
244
+ "learning_rate": 0.00013333995073710526,
245
+ "loss": 4.6459,
246
+ "step": 1400
247
+ },
248
+ {
249
+ "epoch": 0.11152988231674486,
250
+ "grad_norm": 4.597095489501953,
251
+ "learning_rate": 0.0001330826072570861,
252
+ "loss": 4.6778,
253
+ "step": 1450
254
+ },
255
+ {
256
+ "epoch": 0.1153757403276671,
257
+ "grad_norm": 5.779391288757324,
258
+ "learning_rate": 0.000132825263777067,
259
+ "loss": 4.7938,
260
+ "step": 1500
261
+ },
262
+ {
263
+ "epoch": 0.1153757403276671,
264
+ "eval_loss": 4.696172714233398,
265
+ "eval_runtime": 18.8705,
266
+ "eval_samples_per_second": 52.993,
267
+ "eval_steps_per_second": 13.248,
268
+ "step": 1500
269
  }
270
  ],
271
  "logging_steps": 50,