adiren7 commited on
Commit
f6db793
·
verified ·
1 Parent(s): 1471824

Training in progress, step 2000

Browse files
fine-tune-whisper-non-streaming.ipynb CHANGED
@@ -1178,8 +1178,8 @@
1178
  "\n",
1179
  " <div>\n",
1180
  " \n",
1181
- " <progress value='1398' max='5000' style='width:300px; height:20px; vertical-align: middle;'></progress>\n",
1182
- " [1398/5000 33:48 < 1:27:12, 0.69 it/s, Epoch 1397/5000]\n",
1183
  " </div>\n",
1184
  " <table border=\"1\" class=\"dataframe\">\n",
1185
  " <thead>\n",
@@ -1197,6 +1197,12 @@
1197
  " <td>1.685839</td>\n",
1198
  " <td>87.307258</td>\n",
1199
  " </tr>\n",
 
 
 
 
 
 
1200
  " </tbody>\n",
1201
  "</table><p>"
1202
  ],
@@ -1216,6 +1222,10 @@
1216
  "Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41.\n",
1217
  "Non-default generation parameters: {'max_length': 448, 'suppress_tokens': [], 'begin_suppress_tokens': [220, 50257]}\n",
1218
  "/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:460: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.\n",
 
 
 
 
1219
  " warnings.warn(\n"
1220
  ]
1221
  }
 
1178
  "\n",
1179
  " <div>\n",
1180
  " \n",
1181
+ " <progress value='2302' max='5000' style='width:300px; height:20px; vertical-align: middle;'></progress>\n",
1182
+ " [2302/5000 56:12 < 1:05:56, 0.68 it/s, Epoch 2301/5000]\n",
1183
  " </div>\n",
1184
  " <table border=\"1\" class=\"dataframe\">\n",
1185
  " <thead>\n",
 
1197
  " <td>1.685839</td>\n",
1198
  " <td>87.307258</td>\n",
1199
  " </tr>\n",
1200
+ " <tr>\n",
1201
+ " <td>2000</td>\n",
1202
+ " <td>0.000100</td>\n",
1203
+ " <td>2.064901</td>\n",
1204
+ " <td>94.433998</td>\n",
1205
+ " </tr>\n",
1206
  " </tbody>\n",
1207
  "</table><p>"
1208
  ],
 
1222
  "Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41.\n",
1223
  "Non-default generation parameters: {'max_length': 448, 'suppress_tokens': [], 'begin_suppress_tokens': [220, 50257]}\n",
1224
  "/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:460: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.\n",
1225
+ " warnings.warn(\n",
1226
+ "Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41.\n",
1227
+ "Non-default generation parameters: {'max_length': 448, 'suppress_tokens': [], 'begin_suppress_tokens': [220, 50257]}\n",
1228
+ "/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:460: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.\n",
1229
  " warnings.warn(\n"
1230
  ]
1231
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61d7ebbbc4f9fb7a5627d20f9476b64ea138b4de962f89b67e85e4b826c56319
3
  size 966995080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6351cfef62d46c9a60b67d430cc00296c31489b653f72f99115248453837721a
3
  size 966995080
runs/Jul16_23-56-00_e25538e1b1ec/events.out.tfevents.1721233484.e25538e1b1ec CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e75659c736d7176958dae9acd85dabc20778c001e08181c9c27b179ec653f810
3
- size 14070
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bba34266a7c7c61e3692f165f29f1c49ed86a34516f15d6221a2726697e5369
3
+ size 22828