Update main.py
Browse files
main.py
CHANGED
|
@@ -16,6 +16,13 @@ import logging
|
|
| 16 |
import ffmpeg
|
| 17 |
import tempfile
|
| 18 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
# Silence all transformers and huggingface logging
|
| 20 |
logging.getLogger("transformers").setLevel(logging.ERROR)
|
| 21 |
logging.getLogger("urllib3").setLevel(logging.ERROR)
|
|
@@ -108,8 +115,10 @@ def correct_grammar(text, grammar_pipeline):
|
|
| 108 |
return '. '.join([r['generated_text'] for r in results])
|
| 109 |
|
| 110 |
# ========== Initialize Models ==========
|
| 111 |
-
processor, whisper_model = load_whisper_model("small")
|
| 112 |
-
grammar_pipeline = load_grammar_model()
|
|
|
|
|
|
|
| 113 |
|
| 114 |
# ========== Warm-Up Models ==========
|
| 115 |
def warm_up_models():
|
|
|
|
| 16 |
import ffmpeg
|
| 17 |
import tempfile
|
| 18 |
|
| 19 |
+
# Force HF cache to /tmp
|
| 20 |
+
os.environ["HF_HOME"] = "/tmp/huggingface"
|
| 21 |
+
os.environ["TRANSFORMERS_CACHE"] = "/tmp/huggingface/transformers"
|
| 22 |
+
os.environ["HF_HUB_CACHE"] = "/tmp/huggingface/hub"
|
| 23 |
+
|
| 24 |
+
os.makedirs(os.environ["HF_HOME"], exist_ok=True)
|
| 25 |
+
|
| 26 |
# Silence all transformers and huggingface logging
|
| 27 |
logging.getLogger("transformers").setLevel(logging.ERROR)
|
| 28 |
logging.getLogger("urllib3").setLevel(logging.ERROR)
|
|
|
|
| 115 |
return '. '.join([r['generated_text'] for r in results])
|
| 116 |
|
| 117 |
# ========== Initialize Models ==========
|
| 118 |
+
# processor, whisper_model = load_whisper_model("small")
|
| 119 |
+
# grammar_pipeline = load_grammar_model()
|
| 120 |
+
processor = WhisperProcessor.from_pretrained(model_name)
|
| 121 |
+
model = WhisperForConditionalGeneration.from_pretrained(model_name)
|
| 122 |
|
| 123 |
# ========== Warm-Up Models ==========
|
| 124 |
def warm_up_models():
|