Baselhany commited on
Commit
d16a29b
·
verified ·
1 Parent(s): 648a112

Training in progress, step 3600, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77c42fb89ca5e7eeacd8dee347614384694d8e88a850be6f81c4be67614baaea
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69193aaf7ea71a75c6a53dabfa0a65ddf34dc8a499a127bc128b2aa53579f9fa
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d33bb8913e40252c03691681597e0dda6da0e5d3ecd015fac2bf0734e9edc3fd
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b8a538434890bc9b555c3d707ff536182a1ce9e72fd2075b1e4b25a5bfb7b54
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a0b9074bf4ad76d06be1f1774df5d1b38e357952bea3c631cc895318d367b54d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d30d045d4139d3f55dfce92596184610f9efa25b6cf0587ae9fb8624b44b114d
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:124d751b7518ff7dfcb4911295c0150ebaf171d8ca363a18657a8580b6e7df96
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8dfdc3004aaa3171852de1590839deeeabef4dbb22c233c90f7014a88b5dd61
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0560bb7d987b0d9f7026aea630b8dd8d632a86b445987fc463b416ee149d3af2
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93bedcf8402af55b095bf6d5c557a90f995c7c5096c156990bbf9a11e0c8faf9
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 3200,
3
  "best_metric": 0.22575891541408782,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-3200",
5
- "epoch": 2.2793515054338145,
6
  "eval_steps": 400,
7
- "global_step": 3200,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -304,6 +304,43 @@
304
  "eval_steps_per_second": 0.407,
305
  "eval_wer": 0.22575891541408782,
306
  "step": 3200
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
307
  }
308
  ],
309
  "logging_steps": 100,
@@ -323,7 +360,7 @@
323
  "attributes": {}
324
  }
325
  },
326
- "total_flos": 4.16027087732736e+18,
327
  "train_batch_size": 8,
328
  "trial_name": null,
329
  "trial_params": null
 
2
  "best_global_step": 3200,
3
  "best_metric": 0.22575891541408782,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-3200",
5
+ "epoch": 2.564404061998931,
6
  "eval_steps": 400,
7
+ "global_step": 3600,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
304
  "eval_steps_per_second": 0.407,
305
  "eval_wer": 0.22575891541408782,
306
  "step": 3200
307
+ },
308
+ {
309
+ "epoch": 2.3506146445750935,
310
+ "grad_norm": 39.37460708618164,
311
+ "learning_rate": 2.491237530331626e-05,
312
+ "loss": 4.4,
313
+ "step": 3300
314
+ },
315
+ {
316
+ "epoch": 2.4218777837163725,
317
+ "grad_norm": 81.92980194091797,
318
+ "learning_rate": 2.2216230789970345e-05,
319
+ "loss": 4.5424,
320
+ "step": 3400
321
+ },
322
+ {
323
+ "epoch": 2.493140922857652,
324
+ "grad_norm": 69.34239959716797,
325
+ "learning_rate": 1.9520086276624428e-05,
326
+ "loss": 4.3866,
327
+ "step": 3500
328
+ },
329
+ {
330
+ "epoch": 2.564404061998931,
331
+ "grad_norm": 42.60212326049805,
332
+ "learning_rate": 1.682394176327851e-05,
333
+ "loss": 4.1488,
334
+ "step": 3600
335
+ },
336
+ {
337
+ "epoch": 2.564404061998931,
338
+ "eval_loss": 0.11083566397428513,
339
+ "eval_runtime": 157.7821,
340
+ "eval_samples_per_second": 3.169,
341
+ "eval_steps_per_second": 0.399,
342
+ "eval_wer": 0.23887415266725612,
343
+ "step": 3600
344
  }
345
  ],
346
  "logging_steps": 100,
 
360
  "attributes": {}
361
  }
362
  },
363
+ "total_flos": 4.68059945435136e+18,
364
  "train_batch_size": 8,
365
  "trial_name": null,
366
  "trial_params": null