Baselhany commited on
Commit
37b1382
·
verified ·
1 Parent(s): d9f68aa

Training in progress, step 32000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cafca5fce7ab45753baba0338f50645e6a1b9b7421c299ea82fec48d3bf3d36
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee40e71245bf0f66c619c4566f44ff371a139d5158466fdc9d52c3e8ec2e3176
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b35c5d0465d9b3ece1745281e3d3c84209edd4483c5aea554c50fe7150a4481
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23d7f9f3b44eb0e3f4ac71847a56e56f764cdaa70a0e972cf685f8a504364deb
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d58b6a9a7b8cc6729fbb3b5459ccf762552d9a3d564aec2634f31bb934eb9e50
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:889fba2dda612c6c0d4296e738b798d52e27bd546723613a9f1022c81f0a1b37
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dbf4261ff9d86adf30ca05b86d56b875327e2250853d232d8b94713198d83e00
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01a2425ae56cb630c35a174629553276a10429b35fc53ac726db0dfb5fc49bea
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f300f5e6b28e35d46bb757938c0a422473b4c31b18fb2753c64908a1536391d7
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e27c322f011be2d3e98ac457c9a136957b37708f9a786651d2be8677d0d76424
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 18400,
3
  "best_metric": 0.19658119658119658,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-18400",
5
- "epoch": 22.519508284339924,
6
  "eval_steps": 400,
7
- "global_step": 31600,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2931,6 +2931,43 @@
2931
  "eval_steps_per_second": 0.4,
2932
  "eval_wer": 0.1987916298261126,
2933
  "step": 31600
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2934
  }
2935
  ],
2936
  "logging_steps": 100,
@@ -2950,7 +2987,7 @@
2950
  "attributes": {}
2951
  }
2952
  },
2953
- "total_flos": 4.108368101769216e+19,
2954
  "train_batch_size": 8,
2955
  "trial_name": null,
2956
  "trial_params": null
 
2
  "best_global_step": 18400,
3
  "best_metric": 0.19658119658119658,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-18400",
5
+ "epoch": 22.80456084090504,
6
  "eval_steps": 400,
7
+ "global_step": 32000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2931
  "eval_steps_per_second": 0.4,
2932
  "eval_wer": 0.1987916298261126,
2933
  "step": 31600
2934
+ },
2935
+ {
2936
+ "epoch": 22.590771423481204,
2937
+ "grad_norm": 16.371854782104492,
2938
+ "learning_rate": 9.816341287057123e-06,
2939
+ "loss": 1.0875,
2940
+ "step": 31700
2941
+ },
2942
+ {
2943
+ "epoch": 22.662034562622484,
2944
+ "grad_norm": 12.373185157775879,
2945
+ "learning_rate": 9.527114967462039e-06,
2946
+ "loss": 1.0525,
2947
+ "step": 31800
2948
+ },
2949
+ {
2950
+ "epoch": 22.733297701763764,
2951
+ "grad_norm": 16.901241302490234,
2952
+ "learning_rate": 9.237888647866955e-06,
2953
+ "loss": 0.9976,
2954
+ "step": 31900
2955
+ },
2956
+ {
2957
+ "epoch": 22.80456084090504,
2958
+ "grad_norm": 11.734404563903809,
2959
+ "learning_rate": 8.948662328271873e-06,
2960
+ "loss": 1.0248,
2961
+ "step": 32000
2962
+ },
2963
+ {
2964
+ "epoch": 22.80456084090504,
2965
+ "eval_loss": 0.09294673800468445,
2966
+ "eval_runtime": 159.1061,
2967
+ "eval_samples_per_second": 3.143,
2968
+ "eval_steps_per_second": 0.396,
2969
+ "eval_wer": 0.201886236368995,
2970
+ "step": 32000
2971
  }
2972
  ],
2973
  "logging_steps": 100,
 
2987
  "attributes": {}
2988
  }
2989
  },
2990
+ "total_flos": 4.160400959471616e+19,
2991
  "train_batch_size": 8,
2992
  "trial_name": null,
2993
  "trial_params": null