adiren7 commited on
Commit
c9e3eb9
·
verified ·
1 Parent(s): f6db793

Training in progress, step 3000

Browse files
fine-tune-whisper-non-streaming.ipynb CHANGED
@@ -1178,8 +1178,8 @@
1178
  "\n",
1179
  " <div>\n",
1180
  " \n",
1181
- " <progress value='2302' max='5000' style='width:300px; height:20px; vertical-align: middle;'></progress>\n",
1182
- " [2302/5000 56:12 < 1:05:56, 0.68 it/s, Epoch 2301/5000]\n",
1183
  " </div>\n",
1184
  " <table border=\"1\" class=\"dataframe\">\n",
1185
  " <thead>\n",
@@ -1203,6 +1203,12 @@
1203
  " <td>2.064901</td>\n",
1204
  " <td>94.433998</td>\n",
1205
  " </tr>\n",
 
 
 
 
 
 
1206
  " </tbody>\n",
1207
  "</table><p>"
1208
  ],
@@ -1226,6 +1232,10 @@
1226
  "Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41.\n",
1227
  "Non-default generation parameters: {'max_length': 448, 'suppress_tokens': [], 'begin_suppress_tokens': [220, 50257]}\n",
1228
  "/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:460: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.\n",
 
 
 
 
1229
  " warnings.warn(\n"
1230
  ]
1231
  }
 
1178
  "\n",
1179
  " <div>\n",
1180
  " \n",
1181
+ " <progress value='3224' max='5000' style='width:300px; height:20px; vertical-align: middle;'></progress>\n",
1182
+ " [3224/5000 1:19:09 < 43:38, 0.68 it/s, Epoch 3223/5000]\n",
1183
  " </div>\n",
1184
  " <table border=\"1\" class=\"dataframe\">\n",
1185
  " <thead>\n",
 
1203
  " <td>2.064901</td>\n",
1204
  " <td>94.433998</td>\n",
1205
  " </tr>\n",
1206
+ " <tr>\n",
1207
+ " <td>3000</td>\n",
1208
+ " <td>0.000000</td>\n",
1209
+ " <td>2.257328</td>\n",
1210
+ " <td>95.468221</td>\n",
1211
+ " </tr>\n",
1212
  " </tbody>\n",
1213
  "</table><p>"
1214
  ],
 
1232
  "Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41.\n",
1233
  "Non-default generation parameters: {'max_length': 448, 'suppress_tokens': [], 'begin_suppress_tokens': [220, 50257]}\n",
1234
  "/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:460: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.\n",
1235
+ " warnings.warn(\n",
1236
+ "Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41.\n",
1237
+ "Non-default generation parameters: {'max_length': 448, 'suppress_tokens': [], 'begin_suppress_tokens': [220, 50257]}\n",
1238
+ "/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:460: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.\n",
1239
  " warnings.warn(\n"
1240
  ]
1241
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6351cfef62d46c9a60b67d430cc00296c31489b653f72f99115248453837721a
3
  size 966995080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f44b16a0d4692512391b72d61c3cf40ae8bfe5c27cbd8a2be38b9df5ef80719c
3
  size 966995080
runs/Jul16_23-56-00_e25538e1b1ec/events.out.tfevents.1721233484.e25538e1b1ec CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1bba34266a7c7c61e3692f165f29f1c49ed86a34516f15d6221a2726697e5369
3
- size 22828
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d150045e3a490bf882835a665f66b1747c9e63025bcd05caa75f5c1855435af
3
+ size 31904