Drbrain0620's picture
Update main.py
b9a4cf0 verified
raw
history blame contribute delete
922 Bytes
import gradio as gr
import base64
import torch
from transformers import pipeline
# Whisper ๋ชจ๋ธ์„ pipeline์œผ๋กœ ๋ถˆ๋Ÿฌ์˜ค๊ธฐ
whisper = pipeline("automatic-speech-recognition", model="openai/whisper-small")
# ์Œ์„ฑ์„ ํ…์ŠคํŠธ๋กœ ๋ณ€ํ™˜ํ•˜๋Š” ํ•จ์ˆ˜
def transcribe_audio(audio):
if audio is None:
return "์—๋Ÿฌ: ์˜ค๋””์˜ค ์—†์Œ", ""
result = whisper(audio)
return result["text"], base64.b64encode(result["text"].encode()).decode()
# Gradio ์ธํ„ฐํŽ˜์ด์Šค
demo = gr.Interface(
fn=transcribe_audio,
inputs=gr.Audio(label = '์˜ค๋””์˜ค', sources="microphone", type='filepath'),
outputs=[gr.Textbox(label='๊ฒฐ๊ณผ'), gr.Textbox(label='์•”ํ˜ธํ™”๋œ ๊ฒฐ๊ณผ')],
title='์ด์šฐ์ง„์˜ Speech to Text (โ€ป ๋…น์Œ ํ›„ ๋ฐ”๋กœ ์‹คํ–‰ ๋ˆ„๋ฅด๋ฉด ์—๋Ÿฌ๋‚จ)',
description='๊ธฐ์—ฌ๋„: AI 60% ๋‚˜ 40%',
submit_btn='์‹คํ–‰',
clear_btn='์ง€์šฐ๊ธฐ')
# ์•ฑ ์‹คํ–‰
demo.launch()