Baselhany commited on
Commit
63a0b06
·
verified ·
1 Parent(s): c84833f

Training in progress, step 34000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f22325aa1f5ed494b0d5c0d16a8a913f20d38ec71bd0dd5e939c08738dde49d
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20bdfe64bb7787dbde06ed0249c104c2a75f74f77a4da7b6d17b0be3c907bffe
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8e0d19a7a0c0b61e3d26116ae68f7eaa58e61ff51d6f44d3d0bf74bdf2300ee
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55dd9570954d73b04e0eb476a55f9ed87ca0571abe03bdd05b897829def60e9b
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:071d1078d3ac4159702884d4d75a81a594e66bcc8d271e9eff2fa9013e87bfc4
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19d659484e215c1796fcedcf712444963aec532ba7ef9b9f53c66f0a4322a2d0
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7716c9ce498bee52d2573a1cacdfdb44092db12b548eaa85005da9fb2a335024
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e82de6898de222ba34f053188f784b1daff292a8caca719707d751a738a4d585
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfbb59a646988d0220b43b90f89f163169bfcd3208a594ad2cc8a4e9b2b6b2ea
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8627c62832e908e5ee44d6dc9ec10410d33b0d8b29cb328e2941165b3e5c98eb
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 32400,
3
  "best_metric": 0.19510757441791923,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-32400",
5
- "epoch": 23.944236593621948,
6
  "eval_steps": 400,
7
- "global_step": 33600,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -3116,6 +3116,43 @@
3116
  "eval_steps_per_second": 0.397,
3117
  "eval_wer": 0.19643383436486886,
3118
  "step": 33600
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3119
  }
3120
  ],
3121
  "logging_steps": 100,
@@ -3135,7 +3172,7 @@
3135
  "attributes": {}
3136
  }
3137
  },
3138
- "total_flos": 4.368414503337984e+19,
3139
  "train_batch_size": 8,
3140
  "trial_name": null,
3141
  "trial_params": null
 
2
  "best_global_step": 32400,
3
  "best_metric": 0.19510757441791923,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-32400",
5
+ "epoch": 24.228754676643508,
6
  "eval_steps": 400,
7
+ "global_step": 34000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
3116
  "eval_steps_per_second": 0.397,
3117
  "eval_wer": 0.19643383436486886,
3118
  "step": 33600
3119
+ },
3120
+ {
3121
+ "epoch": 24.014965259219668,
3122
+ "grad_norm": 12.55706787109375,
3123
+ "learning_rate": 4.03470715835141e-06,
3124
+ "loss": 0.9904,
3125
+ "step": 33700
3126
+ },
3127
+ {
3128
+ "epoch": 24.086228398360948,
3129
+ "grad_norm": 15.65674114227295,
3130
+ "learning_rate": 3.745480838756327e-06,
3131
+ "loss": 1.0292,
3132
+ "step": 33800
3133
+ },
3134
+ {
3135
+ "epoch": 24.157491537502228,
3136
+ "grad_norm": 12.829319953918457,
3137
+ "learning_rate": 3.4562545191612434e-06,
3138
+ "loss": 0.9981,
3139
+ "step": 33900
3140
+ },
3141
+ {
3142
+ "epoch": 24.228754676643508,
3143
+ "grad_norm": 10.557638168334961,
3144
+ "learning_rate": 3.1670281995661605e-06,
3145
+ "loss": 0.9927,
3146
+ "step": 34000
3147
+ },
3148
+ {
3149
+ "epoch": 24.228754676643508,
3150
+ "eval_loss": 0.09272577613592148,
3151
+ "eval_runtime": 160.295,
3152
+ "eval_samples_per_second": 3.119,
3153
+ "eval_steps_per_second": 0.393,
3154
+ "eval_wer": 0.19790745652814618,
3155
+ "step": 34000
3156
  }
3157
  ],
3158
  "logging_steps": 100,
 
3172
  "attributes": {}
3173
  }
3174
  },
3175
+ "total_flos": 4.420329474097152e+19,
3176
  "train_batch_size": 8,
3177
  "trial_name": null,
3178
  "trial_params": null