Spaces:
Running
Running
Update src/Translate.py
Browse files- src/Translate.py +7 -4
src/Translate.py
CHANGED
|
@@ -42,11 +42,14 @@ class Translators:
|
|
| 42 |
tgt_lang = f"{self.tl}_{self.tl.upper()}"
|
| 43 |
# Load model and tokenizer
|
| 44 |
# from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
| 45 |
-
|
| 46 |
-
|
| 47 |
-
model = MBartForConditionalGeneration.from_pretrained(self.model_name)
|
| 48 |
-
tokenizer = MBartTokenizer.from_pretrained(self.model_name, src_lang=src_lang)
|
| 49 |
# pipe = pipeline("translation", model="facebook/mbart-large-cc25")
|
|
|
|
|
|
|
|
|
|
| 50 |
# Tokenize and translate
|
| 51 |
inputs = tokenizer(self.input_text, return_tensors="pt")
|
| 52 |
translated_tokens = model.generate(
|
|
|
|
| 42 |
tgt_lang = f"{self.tl}_{self.tl.upper()}"
|
| 43 |
# Load model and tokenizer
|
| 44 |
# from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
| 45 |
+
tokenizer = AutoTokenizer.from_pretrained(self.model_name)
|
| 46 |
+
model = AutoModelForSeq2SeqLM.from_pretrained(self.model_name)
|
| 47 |
+
# model = MBartForConditionalGeneration.from_pretrained(self.model_name)
|
| 48 |
+
# tokenizer = MBartTokenizer.from_pretrained(self.model_name, src_lang=src_lang)
|
| 49 |
# pipe = pipeline("translation", model="facebook/mbart-large-cc25")
|
| 50 |
+
# translator = pipeline('translation', model=model, tokenizer=tokenizer, src_lang=self.sl, tgt_lang=self.tl)
|
| 51 |
+
# translated_text = translator(text, max_length=512)
|
| 52 |
+
# return translated_text[0]['translation_text']
|
| 53 |
# Tokenize and translate
|
| 54 |
inputs = tokenizer(self.input_text, return_tensors="pt")
|
| 55 |
translated_tokens = model.generate(
|