| | from transformers import AutoModelForSeq2SeqLM |
| | import torch |
| | |
| | from transformers import AutoTokenizer |
| | tokenizer = AutoTokenizer.from_pretrained("Sunbird/sunbird-mul-en-mbart-merged") |
| |
|
| | |
| | model = AutoModelForSeq2SeqLM.from_pretrained("Sunbird/sunbird-mul-en-mbart-merged") |
| | model.load_state_dict(torch.load("Total Combined Data V2 Aug 16 2023/Models/mul_en_base.bin")) |
| | model = model.to("cuda") |
| |
|
| | sentence = "Kya busiru okuluubirira eby'obugagga, Eddwaliro terisobola kusuza balwadde bangi nnyo. Tekisoboka" |
| | |
| | tokenized_sentence = tokenizer(sentence, return_tensors="pt") |
| | tokenized_sentence = tokenized_sentence.to("cuda") |
| | |
| | translated_tokens = model.generate(**tokenized_sentence) |
| | |
| | translation = tokenizer.decode(translated_tokens[0]) |
| | print("*"*6) |
| | print("\n") |
| | print(translation) |
| | print("\n") |
| | print("*"*6) |
| |
|