| import gradio as gr | |
| from transformers import AutoModelForSeq2SeqLM | |
| from transformers import AutoTokenizer | |
| model = AutoModelForSeq2SeqLM.from_pretrained('hackathon-pln-es/t5-small-spanish-nahuatl') | |
| tokenizer = AutoTokenizer.from_pretrained('hackathon-pln-es/t5-small-spanish-nahuatl') | |
| def predict(input): | |
| input_ids = tokenizer('translate Spanish to Nahuatl: ' + input, return_tensors='pt').input_ids | |
| outputs = model.generate(input_ids) | |
| outputs = tokenizer.batch_decode(outputs, skip_special_tokens=True)[0] | |
| return outputs | |
| gr.Interface( | |
| fn=predict, | |
| inputs=gr.inputs.Textbox(lines=1, label="Input Text in Spanish"), | |
| outputs=[ | |
| gr.outputs.Textbox(label="Translated text in Nahuatl"), | |
| ], | |
| theme="peach", | |
| title='🌽 Spanish to Nahuatl Automatic Translation', | |
| description='This model is a T5 Transformer (t5-small) fine-tuned on 29,007 spanish and nahuatl sentences using 12,890 samples collected from the web and 16,117 samples from the Axolotl dataset. The dataset is normalized using "sep" normalization from py-elotl. For more details visit https://huggingface.co/hackathon-pln-es/t5-small-spanish-nahuatl', | |
| examples=[ | |
| 'hola', | |
| 'conejo', | |
| 'estrella', | |
| 'te quiero mucho', | |
| 'te amo', | |
| 'quiero comer', | |
| 'esto se llama agua', | |
| 'mi abuelo se llama Juan', | |
| 'te amo con todo mi corazón'], | |
| allow_flagging="manual", | |
| flagging_options=["right translation", "wrong translation", "error", "other"], | |
| flagging_dir="logs" | |
| ).launch(enable_queue=True, debug=True) | |