Spaces:
Runtime error
Runtime error
liuyang
commited on
Commit
·
ba3077f
1
Parent(s):
1d18680
fix audio param
Browse files
app.py
CHANGED
|
@@ -514,10 +514,6 @@ class WhisperTranscriber:
|
|
| 514 |
|
| 515 |
start_time = time.time()
|
| 516 |
|
| 517 |
-
# Load audio (float32, 16k) once
|
| 518 |
-
audio = whisperx.load_audio(audio_path)
|
| 519 |
-
print(audio_path)
|
| 520 |
-
|
| 521 |
# Resolve engine (allow override from transcribe_options)
|
| 522 |
if transcribe_options and isinstance(transcribe_options, dict) and transcribe_options.get("engine"):
|
| 523 |
engine = str(transcribe_options.get("engine")).strip().lower()
|
|
@@ -527,6 +523,10 @@ class WhisperTranscriber:
|
|
| 527 |
detected_language = language if language else "unknown"
|
| 528 |
|
| 529 |
if engine == "whisperx":
|
|
|
|
|
|
|
|
|
|
|
|
|
| 530 |
# Lazy-load WhisperX model on first use
|
| 531 |
if model_name not in _whipser_x_transcribe_models:
|
| 532 |
print(f"Loading WhisperX transcribe model '{model_name}' on GPU...")
|
|
@@ -647,7 +647,7 @@ class WhisperTranscriber:
|
|
| 647 |
_faster_whisper_batched_pipelines[model_name] = BatchedInferencePipeline(model=fw_model)
|
| 648 |
batched_model = _faster_whisper_batched_pipelines[model_name]
|
| 649 |
segments_iter, info = batched_model.transcribe(
|
| 650 |
-
|
| 651 |
language=language,
|
| 652 |
task=task,
|
| 653 |
**fw_kwargs,
|
|
@@ -655,7 +655,7 @@ class WhisperTranscriber:
|
|
| 655 |
else:
|
| 656 |
fw_kwargs.pop("batch_size", None)
|
| 657 |
segments_iter, info = fw_model.transcribe(
|
| 658 |
-
|
| 659 |
language=language,
|
| 660 |
task=task,
|
| 661 |
**fw_kwargs,
|
|
|
|
| 514 |
|
| 515 |
start_time = time.time()
|
| 516 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 517 |
# Resolve engine (allow override from transcribe_options)
|
| 518 |
if transcribe_options and isinstance(transcribe_options, dict) and transcribe_options.get("engine"):
|
| 519 |
engine = str(transcribe_options.get("engine")).strip().lower()
|
|
|
|
| 523 |
detected_language = language if language else "unknown"
|
| 524 |
|
| 525 |
if engine == "whisperx":
|
| 526 |
+
# Load audio (float32, 16k) once
|
| 527 |
+
audio = whisperx.load_audio(audio_path)
|
| 528 |
+
print(audio_path)
|
| 529 |
+
|
| 530 |
# Lazy-load WhisperX model on first use
|
| 531 |
if model_name not in _whipser_x_transcribe_models:
|
| 532 |
print(f"Loading WhisperX transcribe model '{model_name}' on GPU...")
|
|
|
|
| 647 |
_faster_whisper_batched_pipelines[model_name] = BatchedInferencePipeline(model=fw_model)
|
| 648 |
batched_model = _faster_whisper_batched_pipelines[model_name]
|
| 649 |
segments_iter, info = batched_model.transcribe(
|
| 650 |
+
audio_path,
|
| 651 |
language=language,
|
| 652 |
task=task,
|
| 653 |
**fw_kwargs,
|
|
|
|
| 655 |
else:
|
| 656 |
fw_kwargs.pop("batch_size", None)
|
| 657 |
segments_iter, info = fw_model.transcribe(
|
| 658 |
+
audio_path,
|
| 659 |
language=language,
|
| 660 |
task=task,
|
| 661 |
**fw_kwargs,
|