File size: 6,725 Bytes
0ba63c7
 
 
 
 
 
 
 
 
382faa1
 
 
c7676b1
0ba63c7
93057e6
 
 
cb16cfc
 
 
93057e6
0ba63c7
93057e6
0ba63c7
 
 
 
 
 
 
 
 
 
 
 
93057e6
c7676b1
93057e6
 
 
0ba63c7
 
 
 
 
 
 
 
93057e6
 
0ba63c7
93057e6
0ba63c7
93057e6
cb16cfc
0ba63c7
 
 
 
93057e6
 
 
0ba63c7
 
93057e6
0ba63c7
 
 
93057e6
0ba63c7
93057e6
 
 
cb16cfc
0ba63c7
cb16cfc
93057e6
0ba63c7
 
93057e6
0ba63c7
 
 
 
 
 
 
 
93057e6
0ba63c7
 
 
 
 
 
 
 
93057e6
0ba63c7
93057e6
 
382faa1
 
93057e6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cb16cfc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
93057e6
cb16cfc
 
93057e6
 
382faa1
 
93057e6
 
 
cb16cfc
93057e6
0ba63c7
cb16cfc
93057e6
 
 
cb16cfc
 
93057e6
0ba63c7
93057e6
 
cb16cfc
93057e6
 
 
 
 
 
 
cb16cfc
93057e6
 
 
 
cb16cfc
 
 
0ba63c7
 
 
382faa1
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
import gradio as gr
import numpy as np
import torch
import soundfile as sf
import librosa
from transformers import AutoFeatureExtractor, AutoModelForAudioFrameClassification
from recitations_segmenter import segment_recitations, clean_speech_intervals
import io
from PIL import Image
import tempfile
import os
import zipfile
from gradio_client import Client

# ๐Ÿ”น ASR client
from gradio_client import Client, handle_file

# ๐Ÿ”น Arabic Aligner
from arabic_aligner import ArabicAligner  # ุงู„ู…ู„ู ุงู„ู„ูŠ ููŠู‡ ุงู„ูƒูˆุฏ ุงู„ู„ูŠ ุจุนุชุชู‡ ู‚ุจู„ ูƒุฏู‡

# ======================
# Setup device and model
# ======================
device = 'cuda' if torch.cuda.is_available() else 'cpu'
dtype = torch.bfloat16 if torch.cuda.is_available() else torch.float32

print(f"Loading model on {device}...")
processor = AutoFeatureExtractor.from_pretrained("obadx/recitation-segmenter-v2")
model = AutoModelForAudioFrameClassification.from_pretrained(
    "obadx/recitation-segmenter-v2",
    torch_dtype=dtype,
    device_map=device
)
print("Model loaded successfully!")

# ๐Ÿ”น ASR Space
asr_client = Client("aboalaa1472/Quran_ASR", timeout=120)
# ======================
# Utils
# ======================
def read_audio(path, sampling_rate=16000):
    audio, sr = sf.read(path)
    if len(audio.shape) > 1:
        audio = audio.mean(axis=1)
    if sr != sampling_rate:
        audio = librosa.resample(audio, orig_sr=sr, target_sr=sampling_rate)
    return torch.tensor(audio).float()

def get_interval(x, intervals, idx, sr=16000):
    start = int(intervals[idx][0] * sr)
    end = int(intervals[idx][1] * sr)
    return x[start:end]

def plot_signal(x, intervals, sr=16000):
    import matplotlib.pyplot as plt
    fig, ax = plt.subplots(figsize=(20, 4))
    if isinstance(x, torch.Tensor):
        x = x.numpy()
    ax.plot(x, linewidth=0.5)
    for s, e in intervals:
        ax.axvline(x=s * sr, color='red', alpha=0.4)
        ax.axvline(x=e * sr, color='red', alpha=0.4)
    plt.tight_layout()
    buf = io.BytesIO()
    plt.savefig(buf, format="png")
    buf.seek(0)
    img = Image.open(buf)
    plt.close()
    return img

# ======================
# Main processing
# ======================
def process_audio_and_compare(audio_file, reference_text, min_silence_ms, min_speech_ms, pad_ms):
    if audio_file is None:
        return None, "โš ๏ธ ุงุฑูุน ู…ู„ู ุตูˆุชูŠ ุฃูˆู„ุงู‹", None

    try:
        wav = read_audio(audio_file)

        sampled_outputs = segment_recitations(
            [wav],
            model,
            processor,
            device=device,
            dtype=dtype,
            batch_size=4,
        )

        clean_out = clean_speech_intervals(
            sampled_outputs[0].speech_intervals,
            sampled_outputs[0].is_complete,
            min_silence_duration_ms=min_silence_ms,
            min_speech_duration_ms=min_speech_ms,
            pad_duration_ms=pad_ms,
            return_seconds=True,
        )

        intervals = clean_out.clean_speech_intervals
        plot_img = plot_signal(wav, intervals)

        temp_dir = tempfile.mkdtemp()
        segment_files = []
        full_asr_text = []

        result_text = f"โœ… ุนุฏุฏ ุงู„ู…ู‚ุงุทุน: {len(intervals)}\n\n"

        for i in range(len(intervals)):
            seg = get_interval(wav, intervals, i)
            if isinstance(seg, torch.Tensor):
                seg = seg.cpu().numpy()

            seg_path = os.path.join(temp_dir, f"segment_{i+1:03d}.wav")
            sf.write(seg_path, seg, 16000)
            segment_files.append(seg_path)

            # ๐Ÿ”น ASR CALL
            asr_text = asr_client.predict(
                uploaded_audio=handle_file(seg_path),
                mic_audio=handle_file(seg_path),
                api_name="/run"
            )
            full_asr_text.append(asr_text)
            result_text += f"๐ŸŽต ู…ู‚ุทุน {i+1} ({intervals[i][0]:.2f}s โ†’ {intervals[i][1]:.2f}s)\n๐Ÿ“œ {asr_text}\n\n"

        full_asr_text_str = " ".join(full_asr_text)
        result_text += f"\n๐Ÿงพ ุงู„ู†ุต ุงู„ูƒุงู…ู„:\n{full_asr_text_str}\n\n"

        # ๐Ÿ”น ArabicAligner comparison
        aligner = ArabicAligner()
        align_results = aligner.align_and_compare(full_asr_text_str, reference_text)

        stats = align_results['statistics']
        result_text += (
            f"๐Ÿ“Š ุฅุญุตุงุฆูŠุงุช ุงู„ู…ู‚ุงุฑู†ุฉ:\n"
            f"- ุฅุฌู…ุงู„ูŠ ูƒู„ู…ุงุช ุงู„ู…ุฑุฌุน: {stats['total_reference_words']}\n"
            f"- ุฅุฌู…ุงู„ูŠ ูƒู„ู…ุงุช ASR: {stats['total_user_words']}\n"
            f"- ุฅุฌู…ุงู„ูŠ ุงู„ุฃุฎุทุงุก: {stats['total_errors']}\n"
            f"  - ุฃุฎุทุงุก ุงู„ูƒู„ู…ุงุช: {stats['word_level_errors']}\n"
            f"  - ุฃุฎุทุงุก ุงู„ุญุฑูƒุงุช: {stats['diacritic_errors']}\n"
            f"- ุงู„ุฏู‚ุฉ: {stats['accuracy']:.2f}%\n\n"
            f"โœ๏ธ ุชูุงุตูŠู„ ุงู„ุฃุฎุทุงุก:\n"
        )

        for i, error in enumerate(align_results['errors'], 1):
            result_text += f"[{i}] Type: {error.error_type.value.upper()} | User: '{error.user_word}' | Expected: '{error.reference_word}' | Details: {error.details}\n"

        # ZIP
        zip_path = os.path.join(temp_dir, "segments.zip")
        with zipfile.ZipFile(zip_path, 'w') as zipf:
            for f in segment_files:
                zipf.write(f, os.path.basename(f))

        return plot_img, result_text, zip_path

    except Exception as e:
        return None, f"โŒ ุฎุทุฃ: {str(e)}", None

# Gradio UI
# ======================
with gr.Blocks(title="Quran Segmentation + ASR + Comparison") as demo:
    gr.Markdown("## ๐Ÿ•Œ ุชู‚ุทูŠุน ุงู„ุชู„ุงูˆุงุช + ุงู„ุชุนุฑู ุนู„ู‰ ุงู„ู†ุต ุงู„ู‚ุฑุขู†ูŠ + ุงู„ู…ู‚ุงุฑู†ุฉ ุจุงู„ู†ุต ุงู„ู…ุดูƒูˆู„")

    with gr.Row():
        with gr.Column():
            audio_input = gr.Audio(type="filepath", label="๐Ÿ“ค ุงุฑูุน ุงู„ุชู„ุงูˆุฉ")
            reference_text_input = gr.Textbox(label="๐Ÿ“– ุฃุฏุฎู„ ู†ุต ุงู„ู‚ุฑุขู† ุงู„ู…ุดูƒูˆู„ ู„ู„ู…ู‚ุงุฑู†ุฉ", lines=10)
            min_silence = gr.Slider(10, 500, 30, step=10, label="Min Silence (ms)")
            min_speech = gr.Slider(10, 500, 30, step=10, label="Min Speech (ms)")
            padding = gr.Slider(0, 200, 30, step=10, label="Padding (ms)")
            btn = gr.Button("๐Ÿš€ ุงุจุฏุฃ")

        with gr.Column():
            plot_out = gr.Image(label="๐Ÿ“ˆ ุงู„ุฅุดุงุฑุฉ")
            text_out = gr.Textbox(lines=30, label="๐Ÿ“œ ุงู„ู†ุชุงุฆุฌ")

    zip_out = gr.File(label="๐Ÿ“ฆ ุชุญู…ูŠู„ ุงู„ู…ู‚ุงุทุน")

    btn.click(
        fn=process_audio_and_compare,
        inputs=[audio_input, reference_text_input, min_silence, min_speech, padding],
        outputs=[plot_out, text_out, zip_out]
    )

if __name__ == "__main__":
    demo.launch()