liuyang commited on
Commit
ba3077f
·
1 Parent(s): 1d18680

fix audio param

Browse files
Files changed (1) hide show
  1. app.py +6 -6
app.py CHANGED
@@ -514,10 +514,6 @@ class WhisperTranscriber:
514
 
515
  start_time = time.time()
516
 
517
- # Load audio (float32, 16k) once
518
- audio = whisperx.load_audio(audio_path)
519
- print(audio_path)
520
-
521
  # Resolve engine (allow override from transcribe_options)
522
  if transcribe_options and isinstance(transcribe_options, dict) and transcribe_options.get("engine"):
523
  engine = str(transcribe_options.get("engine")).strip().lower()
@@ -527,6 +523,10 @@ class WhisperTranscriber:
527
  detected_language = language if language else "unknown"
528
 
529
  if engine == "whisperx":
 
 
 
 
530
  # Lazy-load WhisperX model on first use
531
  if model_name not in _whipser_x_transcribe_models:
532
  print(f"Loading WhisperX transcribe model '{model_name}' on GPU...")
@@ -647,7 +647,7 @@ class WhisperTranscriber:
647
  _faster_whisper_batched_pipelines[model_name] = BatchedInferencePipeline(model=fw_model)
648
  batched_model = _faster_whisper_batched_pipelines[model_name]
649
  segments_iter, info = batched_model.transcribe(
650
- audio,
651
  language=language,
652
  task=task,
653
  **fw_kwargs,
@@ -655,7 +655,7 @@ class WhisperTranscriber:
655
  else:
656
  fw_kwargs.pop("batch_size", None)
657
  segments_iter, info = fw_model.transcribe(
658
- audio,
659
  language=language,
660
  task=task,
661
  **fw_kwargs,
 
514
 
515
  start_time = time.time()
516
 
 
 
 
 
517
  # Resolve engine (allow override from transcribe_options)
518
  if transcribe_options and isinstance(transcribe_options, dict) and transcribe_options.get("engine"):
519
  engine = str(transcribe_options.get("engine")).strip().lower()
 
523
  detected_language = language if language else "unknown"
524
 
525
  if engine == "whisperx":
526
+ # Load audio (float32, 16k) once
527
+ audio = whisperx.load_audio(audio_path)
528
+ print(audio_path)
529
+
530
  # Lazy-load WhisperX model on first use
531
  if model_name not in _whipser_x_transcribe_models:
532
  print(f"Loading WhisperX transcribe model '{model_name}' on GPU...")
 
647
  _faster_whisper_batched_pipelines[model_name] = BatchedInferencePipeline(model=fw_model)
648
  batched_model = _faster_whisper_batched_pipelines[model_name]
649
  segments_iter, info = batched_model.transcribe(
650
+ audio_path,
651
  language=language,
652
  task=task,
653
  **fw_kwargs,
 
655
  else:
656
  fw_kwargs.pop("batch_size", None)
657
  segments_iter, info = fw_model.transcribe(
658
+ audio_path,
659
  language=language,
660
  task=task,
661
  **fw_kwargs,