File size: 1,945 Bytes
c0b9248
 
 
 
 
 
 
 
6b29f15
c0b9248
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
610a57e
c0b9248
 
 
 
 
 
 
 
 
 
 
 
 
0451166
c0b9248
610a57e
c0b9248
 
610a57e
c0b9248
 
 
 
 
 
610a57e
c0b9248
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
import gradio as gr
from iman.sad_tfpy10 import *
from autosub import SpeechRecognizer
from autosub import GOOGLE_SPEECH_API_KEY
import soundfile as sf
import io

css = """ 
 textarea { direction: ltr; text-align: left; font-family: Calibri, sans-serif; font-size: 16px;} 
"""


recognizer = SpeechRecognizer(language="az", rate=16000,api_key=GOOGLE_SPEECH_API_KEY, proxies=None)


seg = Segmenter(ffmpeg_path="ffmpeg",model_path="keras_speech_music_noise_cnn.hdf5" , device="cpu",vad_type="vad")



def process_segment(args):
 segment, wav = args
 start, stop = segment
 # pp = converter((start, stop))
 pp = pcm_to_flac(wav[int(start*16000) : int(stop*16000)])
 tr_beamsearch_lm = recognizer(pp) 
 print("%s\t%s\t%s" %(start,stop , tr_beamsearch_lm ))   
 return start, stop, tr_beamsearch_lm

def pcm_to_flac(pcm_data, sample_rate=16000):
 buffer = io.BytesIO()
 sf.write(buffer, pcm_data, sample_rate, format='FLAC')
 flac_data = buffer.getvalue() 
 return flac_data

    
def transcribe_audio(audio_file):

    text=""
    isig,wav =  seg(audio_file)
    isig = filter_output(isig , max_silence=0.5 ,ignore_small_speech_segments=0.1 , max_speech_len=10 ,split_speech_bigger_than=10)   
    isig = [(a,b) for x,a,b,_,_ in isig]
   
    results=[]
    for segment in isig:
       
         results.append (process_segment((segment, wav)))
    for start, stop, tr_beamsearch_lm in results:
          
          try:
              
              text += ' ' + tr_beamsearch_lm + '\r\n'
         
          except:
               pass
    
    return text

# Define the Gradio interface
interface = gr.Interface(
    fn=transcribe_audio,
    inputs=gr.Audio(type="filepath"),  

    outputs=gr.Textbox(label="Transcription", elem_id="output-text",interactive=True),
    title="Azari Audio Transcription",
    description="Upload an audio file or record audio to get the transcription.",
    css=css
)

# Launch the Gradio app
interface.launch()