| |
|
| | import torch |
| | import torchaudio |
| | import transformers |
| | from config import ModelConfig |
| | from model import MultiModalModel |
| |
|
| | def run_inference(audio_path: str, model_path: str = None): |
| | |
| | config = ModelConfig() |
| | |
| | |
| | model = MultiModalModel(config) |
| | |
| | if model_path: |
| | state_dict = torch.load(f"{model_path}/pytorch_model.bin", map_location="cpu") |
| | model.load_state_dict(state_dict, strict=False) |
| | |
| | model.eval() |
| | |
| | |
| | processor = transformers.AutoProcessor.from_pretrained(config.audio_model_id) |
| | audio, sr = torchaudio.load(audio_path) |
| | if sr != 16000: |
| | audio = torchaudio.functional.resample(audio, sr, 16000) |
| | if audio.shape[0] > 1: |
| | audio = audio.mean(dim=0, keepdim=True) |
| | |
| | audio_inputs = processor(audio.squeeze().numpy(), sampling_rate=16000, return_tensors="pt") |
| | audio_values = audio_inputs.input_features |
| | |
| | |
| | tokenizer = transformers.AutoTokenizer.from_pretrained(config.text_model_id) |
| | text = "Transcribe the following audio:" |
| | text_inputs = tokenizer(text, return_tensors="pt") |
| | |
| | |
| | with torch.no_grad(): |
| | generated_ids = model.generate( |
| | input_ids=text_inputs.input_ids, |
| | audio_values=audio_values, |
| | max_new_tokens=200 |
| | ) |
| | |
| | transcription = tokenizer.decode(generated_ids[0], skip_special_tokens=True) |
| | print("Transcription:", transcription) |
| | return transcription |
| |
|
| | if __name__ == "__main__": |
| | import sys |
| | if len(sys.argv) > 1: |
| | run_inference(sys.argv[1]) |
| | else: |
| | print("Usage: python -m audio_lm.inference path/to/audio.wav") |
| |
|