Spaces:
Sleeping
Sleeping
| import gradio as gr | |
| from transformers import pipeline | |
| import os | |
| access_token = os.getenv("Token") | |
| pipe = pipeline("text-generation", model="meta-llama/Llama-3.2-1B-Instruct", token=access_token) | |
| def predict(query): | |
| messages = [] | |
| messages.append({"role": "system","content": "you are a helpful assistant"}) | |
| messages.append({"role": "user","content": query}) | |
| return pipe(messages,max_new_tokens=256)[0]["generated_text"][-1]['content'] | |
| gradio_app = gr.Interface( | |
| predict, | |
| inputs="text", outputs="text", | |
| title="Llama-3.2-1B-Instruct", | |
| ) | |
| if __name__ == "__main__": | |
| gradio_app.launch() |