Baselhany commited on
Commit
ddb8cc3
·
verified ·
1 Parent(s): 27397e4

Training in progress, step 37554, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82e580b76e08cdfb8448ca6d1803c64c049f1cef861419e8036adf1677312206
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8b9281a6b09625fe8ef329d325b3617b261175ce263da64422ce3eb4bf32fc7
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:51fc0f82c59c2b71d10f095d4baf2da47474fd2fe1bbcbc826cc6c39a81ee2ec
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ddbf47924ce0b202878aefa08df60d1dfdac5bc57f06d3e7653e962c8348d4b
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f328fe058de8a228ddebf585a5f24f49207d162996c245c7975352b2646f7ee
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b1ee0a7cbe660ccdfb692cb8233c5f11e5835a36f7d6d7c2ae404eb3aaeb573
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:29c8852b8972fe5d4076d17cae40f3bcca1d0af379559b08bc3be9a35e8e8830
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:615a2a304c106eccc92843b64971e0c98b40f210bc51b0ae9becd0dae6a72ab5
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c5779eae6320dd873f29bcf99b0bddfe558494d35dfd606a3b158d47caa27415
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e229fb3370f6503adbe15ce94a76f296a892df2f922fc7e71b59c654950e7987
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 28000,
3
  "best_metric": 0.18110816386678455,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-28000",
5
- "epoch": 21.675454012888107,
6
  "eval_steps": 1000,
7
- "global_step": 37000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2931,6 +2931,41 @@
2931
  "eval_steps_per_second": 0.427,
2932
  "eval_wer": 0.1861184792219275,
2933
  "step": 37000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2934
  }
2935
  ],
2936
  "logging_steps": 100,
@@ -2945,12 +2980,12 @@
2945
  "should_evaluate": false,
2946
  "should_log": false,
2947
  "should_save": true,
2948
- "should_training_stop": false
2949
  },
2950
  "attributes": {}
2951
  }
2952
  },
2953
- "total_flos": 4.812697871843328e+19,
2954
  "train_batch_size": 8,
2955
  "trial_name": null,
2956
  "trial_params": null
 
2
  "best_global_step": 28000,
3
  "best_metric": 0.18110816386678455,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-28000",
5
+ "epoch": 22.0,
6
  "eval_steps": 1000,
7
+ "global_step": 37554,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2931
  "eval_steps_per_second": 0.427,
2932
  "eval_wer": 0.1861184792219275,
2933
  "step": 37000
2934
+ },
2935
+ {
2936
+ "epoch": 21.73403632103105,
2937
+ "grad_norm": 6.6599273681640625,
2938
+ "learning_rate": 1.2738165920008636e-06,
2939
+ "loss": 0.6701,
2940
+ "step": 37100
2941
+ },
2942
+ {
2943
+ "epoch": 21.79261862917399,
2944
+ "grad_norm": 4.662840366363525,
2945
+ "learning_rate": 1.0039401953905112e-06,
2946
+ "loss": 0.7006,
2947
+ "step": 37200
2948
+ },
2949
+ {
2950
+ "epoch": 21.851200937316932,
2951
+ "grad_norm": 6.501893043518066,
2952
+ "learning_rate": 7.340637987801587e-07,
2953
+ "loss": 0.66,
2954
+ "step": 37300
2955
+ },
2956
+ {
2957
+ "epoch": 21.90978324545987,
2958
+ "grad_norm": 5.708708763122559,
2959
+ "learning_rate": 4.6418740216980626e-07,
2960
+ "loss": 0.6819,
2961
+ "step": 37400
2962
+ },
2963
+ {
2964
+ "epoch": 21.968365553602812,
2965
+ "grad_norm": 4.885653018951416,
2966
+ "learning_rate": 1.943110055594538e-07,
2967
+ "loss": 0.7,
2968
+ "step": 37500
2969
  }
2970
  ],
2971
  "logging_steps": 100,
 
2980
  "should_evaluate": false,
2981
  "should_log": false,
2982
  "should_save": true,
2983
+ "should_training_stop": true
2984
  },
2985
  "attributes": {}
2986
  }
2987
  },
2988
+ "total_flos": 4.88474711949312e+19,
2989
  "train_batch_size": 8,
2990
  "trial_name": null,
2991
  "trial_params": null