create app.py
Browse files
app.py
ADDED
|
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
Created on Mon Mar 28 01:04:50 2022
|
| 3 |
+
@author: adeep
|
| 4 |
+
"""
|
| 5 |
+
from fnmatch import translate
|
| 6 |
+
import cv2 as cv
|
| 7 |
+
import tempfile
|
| 8 |
+
import numpy as np
|
| 9 |
+
import pandas as pd
|
| 10 |
+
import streamlit as st
|
| 11 |
+
import joblib
|
| 12 |
+
import os
|
| 13 |
+
from moviepy.editor import VideoFileClip
|
| 14 |
+
import speech_recognition as sr
|
| 15 |
+
from pydub import AudioSegment
|
| 16 |
+
from pydub.silence import split_on_silence
|
| 17 |
+
import transformers
|
| 18 |
+
from transformers import pipeline
|
| 19 |
+
import nltk
|
| 20 |
+
nltk.download('punkt')
|
| 21 |
+
nltk.download('averaged_perceptron_tagger')
|
| 22 |
+
import nltk
|
| 23 |
+
nltk.download('punkt')
|
| 24 |
+
nltk.download('averaged_perceptron_tagger')
|
| 25 |
+
from nltk.tokenize import sent_tokenize
|
| 26 |
+
import re
|
| 27 |
+
from utils import get_translation, welcome, get_large_audio_transcription
|
| 28 |
+
|
| 29 |
+
from PIL import Image
|
| 30 |
+
|
| 31 |
+
#import stanfordnlp
|
| 32 |
+
|
| 33 |
+
def main():
|
| 34 |
+
|
| 35 |
+
|
| 36 |
+
st.title("Summarize Text")
|
| 37 |
+
video = st.file_uploader("Choose a file", type=['mp4'])
|
| 38 |
+
button = st.button("Summarize")
|
| 39 |
+
|
| 40 |
+
max_c = st.sidebar.slider('Select max words', 50, 500, step=10, value=150)
|
| 41 |
+
min_c = st.sidebar.slider('Select min words', 10, 450, step=10, value=50)
|
| 42 |
+
gen_summ = False
|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
|
| 46 |
+
with st.spinner("Running.."):
|
| 47 |
+
|
| 48 |
+
if button and video:
|
| 49 |
+
tfile = tempfile.NamedTemporaryFile(delete=False)
|
| 50 |
+
tfile.write(video.read())
|
| 51 |
+
#st.write(tfile.name)
|
| 52 |
+
v = VideoFileClip(tfile.name)
|
| 53 |
+
v.audio.write_audiofile("movie.wav")
|
| 54 |
+
#st.video(video, format="video/mp4", start_time=0)
|
| 55 |
+
#st.audio("movie.wav")
|
| 56 |
+
whole_text=get_large_audio_transcription("movie.wav")
|
| 57 |
+
#st.write(whole_text)
|
| 58 |
+
#summarizer = pipeline("summarization")
|
| 59 |
+
#summarizer = pipeline("summarization", model="t5-base", tokenizer="t5-base", framework="pt")
|
| 60 |
+
summarizer = pipeline("summarization", model="t5-large", tokenizer="t5-large", framework="pt")
|
| 61 |
+
summarized = summarizer(whole_text, min_length=min_c, max_length=max_c)
|
| 62 |
+
summ=summarized[0]['summary_text']
|
| 63 |
+
#st.write(summ)
|
| 64 |
+
gen_summ = True
|
| 65 |
+
#stf_nlp = stanfordnlp.Pipeline(processors='tokenize,mwt,pos')
|
| 66 |
+
#doc = stf_nlp(summ)
|
| 67 |
+
#l=[w.text.capitalize() if w.upos in ["PROPN","NNS"] else w.text for sent in doc.sentences for w in sent.words]
|
| 68 |
+
#text=" ".join(l)
|
| 69 |
+
#summ=truecasing_by_sentence_segmentation(summ)
|
| 70 |
+
sentences = sent_tokenize(summ, language='english')
|
| 71 |
+
# capitalize the sentences
|
| 72 |
+
sentences_capitalized = [s.capitalize() for s in sentences]
|
| 73 |
+
# join the capitalized sentences
|
| 74 |
+
summ = re.sub(" (?=[\.,'!?:;])", "", ' '.join(sentences_capitalized))
|
| 75 |
+
|
| 76 |
+
if 'summary' not in st.session_state:
|
| 77 |
+
st.session_state.summary=True
|
| 78 |
+
st.session_state.summarization = summ
|
| 79 |
+
|
| 80 |
+
|
| 81 |
+
|
| 82 |
+
translate = st.sidebar.radio('Do you want to translate the text to any different language?', ('No', 'Yes'))
|
| 83 |
+
|
| 84 |
+
if translate == 'Yes' and gen_summ == True:
|
| 85 |
+
lang_list = ['Hindi', 'Marathi', 'Malayalam', 'Kannada', 'Telugu', 'Tamil', 'Oriya', 'Bengali', 'Gujarati', 'Urdu']
|
| 86 |
+
|
| 87 |
+
s_type = st.sidebar.selectbox('Select the Language in which you want to Translate:',lang_list)
|
| 88 |
+
st.sidebar.write('You selected:', s_type)
|
| 89 |
+
|
| 90 |
+
if 'summary' in st.session_state:
|
| 91 |
+
summarized_text = st.session_state.summarization
|
| 92 |
+
st.write(summarized_text)
|
| 93 |
+
translation = get_translation(source='English', dest=s_type, text=summarized_text)
|
| 94 |
+
|
| 95 |
+
st.sidebar.write(translation)
|
| 96 |
+
elif translate == 'Yes' and gen_summ == False:
|
| 97 |
+
st.error("The summary has not been generated yet. Please generate the summary first and then translate")
|
| 98 |
+
|
| 99 |
+
else:
|
| 100 |
+
st.write('')
|
| 101 |
+
|
| 102 |
+
if __name__ == '__main__':
|
| 103 |
+
|
| 104 |
+
main()
|