Spaces:
Sleeping
Sleeping
| import gradio as gr | |
| from transformers import pipeline | |
| import torchaudio | |
| import time | |
| # Load Whisper ASR model | |
| transcriber = pipeline(model="openai/whisper-base") | |
| # Load summarization model | |
| summarization_model = pipeline("summarization") | |
| def translate_audio(audio): | |
| # Step 1: Transcribe audio to text | |
| transcription = transcriber(audio) | |
| print('transcription', transcription) | |
| # Step 2: Translate text to Hindi | |
| summary = summarization_model(transcription['text']) | |
| print('summary', summary) | |
| return transcription['text'], summary[0]['summary_text'] | |
| # Create Gradio interface | |
| with gr.Blocks() as iface: | |
| gr.Markdown("# Audio Translator, Summarizer") | |
| with gr.Row(): | |
| audio_input = gr.Audio(type="filepath", label="Upload Audio") | |
| transcription_output = gr.Textbox( | |
| label="Transcribed Text", | |
| info="Initial text") | |
| translation_output = gr.Textbox( | |
| label="Summary", | |
| info="Meeting minute") | |
| translate_button = gr.Button("Translate Audio") | |
| translate_button.click( | |
| translate_audio, | |
| inputs=[audio_input], | |
| outputs=[transcription_output, translation_output] | |
| ) | |
| # Launch the app | |
| iface.launch(share=True) # 'share=True' to get a public link |