Baselhany commited on
Commit
9575a37
·
verified ·
1 Parent(s): d6a332b

Training in progress, step 9600, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2634ee5c91e47c9d9438951c5a31b6505f921ccdc3c0a21f3663947156853947
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ca4effc12eba3aac6be54b6ebefbe335a68a355bfa1a4dd5f6a757e3d20e67a
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75bb3146d72d9dd817a812afea831681c0b29bc8550b271772873f4b41dbd457
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fa103db88bb063249cf00198f06c391c0cbe109211197c08de8bcb1f223a081
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e464bae5ad4106f5a3627e7d672fcdac62ea5f98f2885a50ca5598f6f6382a9
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fe4aa3c6d64b2359fdca2b677edc0f16c40c47a59d8fb0f684e2e556f14ffab
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a55405e1660e5257b9afd90e5fb43c8cb59734b3562afe4dea9642684f8e6541
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2679b4f3b91289b00dcc38839d58e85292314e6e8c351585202758bc552e8648
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba8b9029ae05745a1673f92384ec22df6e7e540c2703e5221e20ea21bb31a083
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab9e93c2a1c338d0c2d563b755e9dc5f46b92db971a762ae1d3d9695b77e067d
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 4000,
3
  "best_metric": 0.18950781019746538,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-4000",
5
- "epoch": 5.453467694131595,
6
  "eval_steps": 400,
7
- "global_step": 9200,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -859,6 +859,43 @@
859
  "eval_steps_per_second": 0.432,
860
  "eval_wer": 0.1943707633362806,
861
  "step": 9200
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
862
  }
863
  ],
864
  "logging_steps": 100,
@@ -878,7 +915,7 @@
878
  "attributes": {}
879
  }
880
  },
881
- "total_flos": 1.19661344980992e+19,
882
  "train_batch_size": 8,
883
  "trial_name": null,
884
  "trial_params": null
 
2
  "best_global_step": 4000,
3
  "best_metric": 0.18950781019746538,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-4000",
5
+ "epoch": 5.6905749851807945,
6
  "eval_steps": 400,
7
+ "global_step": 9600,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
859
  "eval_steps_per_second": 0.432,
860
  "eval_wer": 0.1943707633362806,
861
  "step": 9200
862
+ },
863
+ {
864
+ "epoch": 5.512744516893894,
865
+ "grad_norm": 15.388890266418457,
866
+ "learning_rate": 3.235610957217606e-05,
867
+ "loss": 1.3068,
868
+ "step": 9300
869
+ },
870
+ {
871
+ "epoch": 5.572021339656194,
872
+ "grad_norm": 12.056999206542969,
873
+ "learning_rate": 3.1586642043705757e-05,
874
+ "loss": 1.3009,
875
+ "step": 9400
876
+ },
877
+ {
878
+ "epoch": 5.631298162418494,
879
+ "grad_norm": 15.635717391967773,
880
+ "learning_rate": 3.081717451523546e-05,
881
+ "loss": 1.2486,
882
+ "step": 9500
883
+ },
884
+ {
885
+ "epoch": 5.6905749851807945,
886
+ "grad_norm": 15.254015922546387,
887
+ "learning_rate": 3.004770698676516e-05,
888
+ "loss": 1.3753,
889
+ "step": 9600
890
+ },
891
+ {
892
+ "epoch": 5.6905749851807945,
893
+ "eval_loss": 0.09110399335622787,
894
+ "eval_runtime": 146.1078,
895
+ "eval_samples_per_second": 3.422,
896
+ "eval_steps_per_second": 0.431,
897
+ "eval_wer": 0.1940760389036251,
898
+ "step": 9600
899
  }
900
  ],
901
  "logging_steps": 100,
 
915
  "attributes": {}
916
  }
917
  },
918
+ "total_flos": 1.24864630751232e+19,
919
  "train_batch_size": 8,
920
  "trial_name": null,
921
  "trial_params": null