| import gradio as gr | |
| from transformers import AutoTokenizer, AutoModelForCausalLM | |
| def predict_code(input): | |
| tokenizer = AutoTokenizer.from_pretrained('JetBrains/Mellum-4b-base') | |
| model = AutoModelForCausalLM.from_pretrained('JetBrains/Mellum-4b-base') | |
| encoded_input = tokenizer(input, return_tensors='pt', return_token_type_ids=False) | |
| input_len = len(encoded_input["input_ids"][0]) | |
| out = model.generate( | |
| **encoded_input, | |
| max_new_tokens=100, | |
| ) | |
| prediction = tokenizer.decode(out[0][input_len:]) | |
| return prediction | |
| def run(input): | |
| return predict_code(input) | |
| app = gr.Interface( | |
| fn=run, | |
| inputs=["text"], | |
| outputs=["text"] | |
| ) | |
| app.launch() | |