Baselhany commited on
Commit
c4ca024
·
verified ·
1 Parent(s): f082bb7

Training in progress, step 23600, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ae610937e70b8821d3a95f673534ececaf32e1a5c7a91cc91f420b8d0b3504a6
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d9a3bf99664cae27c073055518be5e618859219ce0c4cc126973844088b23e7
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9d9553129d74646d0142913b2ff5bc7d5b652b86694f6f9afdd91837a9e6adbe
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d112c3743b72b3d287af6914aa1b449485771c9973ab53bd8f8fa446e77f384
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd2a16c621a6088b552c6371de52b5b43ee839d9d4e2a55fdb4031859156826b
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:926d748c36044c22d8aef249e66668c13c49aaf45151cb27dffb465ad9d47cf2
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3519c66bef03d90fe5e133b9e08c9eb6330151ddd8811e14cc9972ce967e5af2
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2229a15bdf42c8dbae70a65f18e33cde5c16363745edce9ae10c3d4c94d5ed8e
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f584f04892a93fc2b5b61b9739764ab14cba2a6af1d9d003974cfd05c836d94
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6ab58750c0d8932deb4621f0a9a0b81209e9c8b7842a4cbe67e31f97248e146
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 18400,
3
  "best_metric": 0.19658119658119658,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-18400",
5
- "epoch": 16.531623017993944,
6
  "eval_steps": 400,
7
- "global_step": 23200,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2154,6 +2154,43 @@
2154
  "eval_steps_per_second": 0.426,
2155
  "eval_wer": 0.2002652519893899,
2156
  "step": 23200
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2157
  }
2158
  ],
2159
  "logging_steps": 100,
@@ -2173,7 +2210,7 @@
2173
  "attributes": {}
2174
  }
2175
  },
2176
- "total_flos": 3.016267524734976e+19,
2177
  "train_batch_size": 8,
2178
  "trial_name": null,
2179
  "trial_params": null
 
2
  "best_global_step": 18400,
3
  "best_metric": 0.19658119658119658,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-18400",
5
+ "epoch": 16.81667557455906,
6
  "eval_steps": 400,
7
+ "global_step": 23600,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2154
  "eval_steps_per_second": 0.426,
2155
  "eval_wer": 0.2002652519893899,
2156
  "step": 23200
2157
+ },
2158
+ {
2159
+ "epoch": 16.60288615713522,
2160
+ "grad_norm": 12.050436019897461,
2161
+ "learning_rate": 2.4281615348379086e-06,
2162
+ "loss": 1.1627,
2163
+ "step": 23300
2164
+ },
2165
+ {
2166
+ "epoch": 16.6741492962765,
2167
+ "grad_norm": 13.23355484008789,
2168
+ "learning_rate": 1.999914350563145e-06,
2169
+ "loss": 1.1094,
2170
+ "step": 23400
2171
+ },
2172
+ {
2173
+ "epoch": 16.74541243541778,
2174
+ "grad_norm": 11.773246765136719,
2175
+ "learning_rate": 1.5716671662883815e-06,
2176
+ "loss": 1.1767,
2177
+ "step": 23500
2178
+ },
2179
+ {
2180
+ "epoch": 16.81667557455906,
2181
+ "grad_norm": 11.992323875427246,
2182
+ "learning_rate": 1.1434199820136182e-06,
2183
+ "loss": 1.1193,
2184
+ "step": 23600
2185
+ },
2186
+ {
2187
+ "epoch": 16.81667557455906,
2188
+ "eval_loss": 0.09421209245920181,
2189
+ "eval_runtime": 146.8439,
2190
+ "eval_samples_per_second": 3.405,
2191
+ "eval_steps_per_second": 0.429,
2192
+ "eval_wer": 0.19908635425876806,
2193
+ "step": 23600
2194
  }
2195
  ],
2196
  "logging_steps": 100,
 
2210
  "attributes": {}
2211
  }
2212
  },
2213
+ "total_flos": 3.068300382437376e+19,
2214
  "train_batch_size": 8,
2215
  "trial_name": null,
2216
  "trial_params": null