Spaces:
Sleeping
Sleeping
File size: 9,106 Bytes
9ba2a1c a67942c 9ba2a1c bef26f3 9ba2a1c a67942c bef26f3 12deec3 bef26f3 51499e8 9ba2a1c a67942c 9ba2a1c bef26f3 a67942c 9ba2a1c 3eb8d72 9ba2a1c 3eb8d72 9ba2a1c 3eb8d72 9ba2a1c 3eb8d72 9ba2a1c bef26f3 9ba2a1c |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 |
import torch
import gradio as gr
import yt_dlp as youtube_dl
from transformers import pipeline
from transformers.pipelines.audio_utils import ffmpeg_read
import tempfile
import os
import numpy as np
from gensim.models import Word2Vec
from sklearn.cluster import KMeans
from sklearn.metrics.pairwise import cosine_similarity
from collections import defaultdict
import spacy
from transformers import pipeline
from sklearn.metrics import davies_bouldin_score
MODEL_NAME = "openai/whisper-large-v3"
BATCH_SIZE = 8
FILE_LIMIT_MB = 1000
YT_LENGTH_LIMIT_S = 3600 # limit to 1 hour YouTube files
device = 0 if torch.cuda.is_available() else "cpu"
pipe = pipeline(
task="automatic-speech-recognition",
model=MODEL_NAME,
chunk_length_s=30,
device=device,
)
summarizer = pipeline("summarization", model="facebook/bart-large-cnn")
# Download the 'en_core_web_sm' model
spacy.cli.download("en_core_web_sm")
# Load the model
nlp = spacy.load("en_core_web_sm")
def summarize(text, max_length=1000):
return summarizer(text, max_length=min(max_length, len(text)), min_length=1, do_sample=False)[0]["summary_text"]
def segment_sentences(text):
# Process the text using spaCy
doc = nlp(text)
# Extract sentences from the processed document
return [sent.text for sent in doc.sents]
def preprocess_sentences(sentences):
preprocessed_sentences = []
for sentence in sentences:
# Tokenize and lemmatize the sentence using spaCy
doc = nlp(sentence.lower())
tokens = [token.lemma_ for token in doc if not token.is_stop and token.is_alpha]
preprocessed_sentences.append(tokens)
return preprocessed_sentences
def embedding(preprocessed_sentences):
model = Word2Vec(preprocessed_sentences, vector_size=100, window=5, min_count=1, sg=1)
sentence_embeddings = []
for sentence in preprocessed_sentences:
word_embeddings = [model.wv[word] for word in sentence if word in model.wv]
if word_embeddings:
sentence_embeddings.append(np.mean(word_embeddings, axis=0))
else:
# Handle the case when none of the words in the sentence exist in the Word2Vec vocabulary
sentence_embeddings.append(np.zeros(model.vector_size)) # Use zero vector as placeholder
return sentence_embeddings
def optimal_n_clusters(sentence_embeddings):
cosine_sim_matrix = cosine_similarity(sentence_embeddings)
db_scores = []
k_values = range(2, len(sentence_embeddings))
for k in k_values:
kmeans = KMeans(n_clusters=k, n_init=10, random_state=42)
cluster_labels = kmeans.fit_predict(cosine_sim_matrix)
db_scores.append(davies_bouldin_score(cosine_sim_matrix, cluster_labels))
# Choose the optimal number of clusters based on Davies-Bouldin index
return (cosine_sim_matrix, np.argmin(db_scores) + 2) # Add 2 to account for skipping k=1
def cluster_assignments(cosine_sim_matrix, optimal_n_clusters):
# Cluster sentence embeddings using KMeans with the optimal number of clusters
kmeans = KMeans(n_clusters=optimal_n_clusters, n_init=10, random_state=42)
return kmeans.fit_predict(cosine_sim_matrix)
def clusters(sentences, cluster_assignments):
# Group sentences into clusters
clusters = defaultdict(list)
for i, sentence in enumerate(sentences):
clusters[cluster_assignments[i]].append(sentence)
result = defaultdict(list)
for i in range(len(clusters)):
cluster = ' '.join(clusters[i])
title = summarize(cluster, 10)
result[title].extend(clusters[i])
return result
def format_as_bullet_points(dictionary):
bullet_points = ""
for key, values in dictionary.items():
bullet_points += f"- {key}:\n"
for value in values:
bullet_points += f" - {value}\n"
return bullet_points.strip()
def final_result(input):
text = summarize(input)
sentences = segment_sentences(text)
preprocessed_sentences = preprocess_sentences(sentences)
sentence_embeddings = embedding(preprocessed_sentences)
cosine_sim_matrix, optimal_number_of_clusters = optimal_n_clusters(sentence_embeddings)
clusters_assignments = cluster_assignments(cosine_sim_matrix, optimal_number_of_clusters)
all_clusters = clusters(sentences, clusters_assignments)
return format_as_bullet_points(all_clusters)
def transcribe(inputs, task):
if inputs is None:
raise gr.Error("No audio file submitted! Please upload or record an audio file before submitting your request.")
text = pipe(inputs, batch_size=BATCH_SIZE, generate_kwargs={"task": task}, return_timestamps=True)["text"]
return final_result(text)
def _return_yt_html_embed(yt_url):
video_id = yt_url.split("?v=")[-1]
HTML_str = (
f'<center> <iframe width="500" height="320" src="https://www.youtube.com/embed/{video_id}"> </iframe>'
" </center>"
)
return HTML_str
def download_yt_audio(yt_url, filename):
info_loader = youtube_dl.YoutubeDL()
try:
info = info_loader.extract_info(yt_url, download=False)
except youtube_dl.utils.DownloadError as err:
raise gr.Error(str(err))
file_length = info["duration_string"]
file_h_m_s = file_length.split(":")
file_h_m_s = [int(sub_length) for sub_length in file_h_m_s]
if len(file_h_m_s) == 1:
file_h_m_s.insert(0, 0)
if len(file_h_m_s) == 2:
file_h_m_s.insert(0, 0)
file_length_s = file_h_m_s[0] * 3600 + file_h_m_s[1] * 60 + file_h_m_s[2]
if file_length_s > YT_LENGTH_LIMIT_S:
yt_length_limit_hms = time.strftime("%HH:%MM:%SS", time.gmtime(YT_LENGTH_LIMIT_S))
file_length_hms = time.strftime("%HH:%MM:%SS", time.gmtime(file_length_s))
raise gr.Error(f"Maximum YouTube length is {yt_length_limit_hms}, got {file_length_hms} YouTube video.")
ydl_opts = {"outtmpl": filename, "format": "worstvideo[ext=mp4]+bestaudio[ext=m4a]/best[ext=mp4]/best"}
with youtube_dl.YoutubeDL(ydl_opts) as ydl:
try:
ydl.download([yt_url])
except youtube_dl.utils.ExtractorError as err:
raise gr.Error(str(err))
def yt_transcribe(yt_url, task, max_filesize=75.0):
html_embed_str = _return_yt_html_embed(yt_url)
with tempfile.TemporaryDirectory() as tmpdirname:
filepath = os.path.join(tmpdirname, "video.mp4")
download_yt_audio(yt_url, filepath)
with open(filepath, "rb") as f:
inputs = f.read()
inputs = ffmpeg_read(inputs, pipe.feature_extractor.sampling_rate)
inputs = {"array": inputs, "sampling_rate": pipe.feature_extractor.sampling_rate}
text = pipe(inputs, batch_size=BATCH_SIZE, generate_kwargs={"task": task}, return_timestamps=True)["text"]
return html_embed_str, final_result(text)
demo = gr.Blocks()
mf_transcribe = gr.Interface(
fn=transcribe,
inputs=[
gr.inputs.Audio(source="microphone", type="filepath", optional=True),
gr.inputs.Radio(["transcribe", "translate"], label="Task", default="transcribe"),
],
outputs="text",
layout="horizontal",
theme="huggingface",
title="Whisper Large V3: Transcribe Audio",
description=(
"Transcribe long-form microphone or audio inputs with the click of a button! Demo uses the OpenAI Whisper"
f" checkpoint [{MODEL_NAME}](https://huggingface.co/{MODEL_NAME}) and 🤗 Transformers to transcribe audio files"
" of arbitrary length."
),
allow_flagging="never",
)
file_transcribe = gr.Interface(
fn=transcribe,
inputs=[
gr.inputs.Audio(source="upload", type="filepath", optional=True, label="Audio file"),
gr.inputs.Radio(["transcribe", "translate"], label="Task", default="transcribe"),
],
outputs="text",
layout="horizontal",
theme="huggingface",
title="Whisper Large V3: Transcribe Audio",
description=(
"Transcribe long-form microphone or audio inputs with the click of a button! Demo uses the OpenAI Whisper"
f" checkpoint [{MODEL_NAME}](https://huggingface.co/{MODEL_NAME}) and 🤗 Transformers to transcribe audio files"
" of arbitrary length."
),
allow_flagging="never",
)
yt_transcribe = gr.Interface(
fn=yt_transcribe,
inputs=[
gr.inputs.Textbox(lines=1, placeholder="Paste the URL to a YouTube video here", label="YouTube URL"),
gr.inputs.Radio(["transcribe", "translate"], label="Task", default="transcribe")
],
outputs=["html", "text"],
layout="horizontal",
theme="huggingface",
title="Whisper Large V3: Transcribe YouTube",
description=(
"Transcribe long-form YouTube videos with the click of a button! Demo uses the OpenAI Whisper checkpoint"
f" [{MODEL_NAME}](https://huggingface.co/{MODEL_NAME}) and 🤗 Transformers to transcribe video files of"
" arbitrary length."
),
allow_flagging="never",
)
with demo:
gr.TabbedInterface([mf_transcribe, file_transcribe, yt_transcribe], ["Microphone", "Audio file", "YouTube"])
demo.launch(enable_queue=True)
|