Baselhany commited on
Commit
3b14954
·
verified ·
1 Parent(s): fdde1f3

Training in progress, step 33200, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f5c9a0489a2cf843b764d138398df3a7a6a5d1d0e8773f219c72b4c09ccfae61
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26cd6571ff6013ee4040c7e9bab5317447cfb4cc34e7b1da8ba6fb8c2ffae362
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:db6620df8d73410ffb0bf9409a6ade7370440551ac5d5b4428a111e658eecd44
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9c377312e70d82cb6f732e1d40d46c81b7b9474fba935ebf032c3f1f2757d90
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28cbe766e4153cd167aa13892e6d8f2b2fb01f0cf38c806d1ede3210aea189b8
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7233ed33137c98d9b12160c311872e55d6db55aa61e9d7f77192b7e6a13fa95
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ef2649ceb258c6256acc0ae4c4605bb5b2c0540ea4c5d0deec7e198b382fb7a
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b25248087d69965ddef920c8054bb52ba085ee231272e0795b2da361d2f441be
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:51726f789396c230f9beb438f4af63a6120e53498ea8e472d26397da95a5f066
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db00d50dd421d79ba139d2536bd27ca7a13377d9e22c077973ca6936cb06d8eb
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 32400,
3
  "best_metric": 0.19510757441791923,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-32400",
5
- "epoch": 23.374131480491716,
6
  "eval_steps": 400,
7
- "global_step": 32800,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -3042,6 +3042,43 @@
3042
  "eval_steps_per_second": 0.399,
3043
  "eval_wer": 0.19952844090775126,
3044
  "step": 32800
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3045
  }
3046
  ],
3047
  "logging_steps": 100,
@@ -3061,7 +3098,7 @@
3061
  "attributes": {}
3062
  }
3063
  },
3064
- "total_flos": 4.264348787933184e+19,
3065
  "train_batch_size": 8,
3066
  "trial_name": null,
3067
  "trial_params": null
 
2
  "best_global_step": 32400,
3
  "best_metric": 0.19510757441791923,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-32400",
5
+ "epoch": 23.659184037056832,
6
  "eval_steps": 400,
7
+ "global_step": 33200,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
3042
  "eval_steps_per_second": 0.399,
3043
  "eval_wer": 0.19952844090775126,
3044
  "step": 32800
3045
+ },
3046
+ {
3047
+ "epoch": 23.445394619632996,
3048
+ "grad_norm": 17.588510513305664,
3049
+ "learning_rate": 6.3456254519161245e-06,
3050
+ "loss": 1.019,
3051
+ "step": 32900
3052
+ },
3053
+ {
3054
+ "epoch": 23.516657758774272,
3055
+ "grad_norm": 13.247255325317383,
3056
+ "learning_rate": 6.0563991323210416e-06,
3057
+ "loss": 1.0711,
3058
+ "step": 33000
3059
+ },
3060
+ {
3061
+ "epoch": 23.587920897915552,
3062
+ "grad_norm": 10.2926025390625,
3063
+ "learning_rate": 5.7671728127259586e-06,
3064
+ "loss": 1.0257,
3065
+ "step": 33100
3066
+ },
3067
+ {
3068
+ "epoch": 23.659184037056832,
3069
+ "grad_norm": 13.390459060668945,
3070
+ "learning_rate": 5.4779464931308756e-06,
3071
+ "loss": 0.9972,
3072
+ "step": 33200
3073
+ },
3074
+ {
3075
+ "epoch": 23.659184037056832,
3076
+ "eval_loss": 0.09225763380527496,
3077
+ "eval_runtime": 158.7123,
3078
+ "eval_samples_per_second": 3.15,
3079
+ "eval_steps_per_second": 0.397,
3080
+ "eval_wer": 0.19952844090775126,
3081
+ "step": 33200
3082
  }
3083
  ],
3084
  "logging_steps": 100,
 
3098
  "attributes": {}
3099
  }
3100
  },
3101
+ "total_flos": 4.316381645635584e+19,
3102
  "train_batch_size": 8,
3103
  "trial_name": null,
3104
  "trial_params": null