KAPtechies commited on
Commit
f762839
·
1 Parent(s): b69664e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +58 -0
app.py CHANGED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ import easyocr
3
+ import transformer
4
+
5
+ reader=easyocr.Reader(['en'])
6
+ # this needs to run only once to load the model into memory
7
+ result=reader.readtext('/content/WhatsApp Image 2023-09-23 at 8.03.28 AM.jpeg',detail=0)
8
+ news=" ".join(result)
9
+ from transformers import AutoTokenizer
10
+ tokenizer=AutoTokenizer.from_pretrained("facebook/mbart-large-50-one-to-many-mmt",use_fast=False)
11
+ from transformers import MBartForConditionalGeneration
12
+ # download and save model
13
+ model=MBartForConditionalGeneration.from_pretrained("facebook/mbart-large-50-one-to-many-mmt")
14
+ input_text=[news]
15
+ # convert sentences to tensors
16
+ model_inputs=tokenizer(input_text,return_tensors="pt",padding=True,truncation=True)
17
+ # translate from English to Hindi
18
+ generated_tokens=model.generate(
19
+ **model_inputs,
20
+ forced_bos_token_id=tokenizer.lang_code_to_id["hi_IN"]
21
+ )
22
+ translation=tokenizer.batch_decode(generated_tokens,skip_special_tokens=True)
23
+ translation
24
+
25
+ from transformers import AutoTokenizer
26
+
27
+
28
+ tokenizer = AutoTokenizer.from_pretrained("facebook/mbart-large-50-one-to-many-mmt", use_fast=False)
29
+
30
+ from transformers import MBartForConditionalGeneration
31
+
32
+ # download and save model
33
+ model = MBartForConditionalGeneration.from_pretrained("facebook/mbart-large-50-one-to-many-mmt")
34
+
35
+ def translator(img):
36
+
37
+ reader = easyocr.Reader(['en'])
38
+ result = reader.readtext(img,detail = 0)
39
+ news= " ".join(result)
40
+ input_text = [news]
41
+
42
+ # convert sentences to tensors
43
+ model_inputs = tokenizer(input_text, return_tensors="pt", padding=True, truncation=True)
44
+
45
+
46
+ # translate from English to Hindi
47
+ generated_tokens = model.generate(
48
+ **model_inputs,
49
+ forced_bos_token_id=tokenizer.lang_code_to_id["hi_IN"]
50
+ )
51
+
52
+ translation = tokenizer.batch_decode(generated_tokens, skip_special_tokens=True)
53
+
54
+ return translation
55
+
56
+ demo = gr.Interface(fn=translator, inputs=gr.Image(), outputs="text")
57
+
58
+ demo.launch(inline=False)