Mauriciotuks commited on
Commit
bce2cd6
·
verified ·
1 Parent(s): 25b6334

Update app.py

Browse files

![IMG_6026.png](https://cdn-uploads.huggingface.co/production/uploads/688dcc38d03635bba0fbc3fb/wWOdbzNSoHvcUI0AuuZQe.png)

Files changed (1) hide show
  1. app.py +8 -16
app.py CHANGED
@@ -1,20 +1,12 @@
1
- import gradio as gr
2
  from transformers import AutoTokenizer, AutoModelForCausalLM
3
 
4
- model = "meta-llama/Meta-Llama-3-8B"
5
 
6
- tokenizer = AutoTokenizer.from_pretrained(model)
7
- model = AutoModelForCausalLM.from_pretrained(model, device_map="auto", torch_dtype="auto")
8
 
9
- def chat_io(message):
10
- inputs = tokenizer(message, return_tensors="pt").to(model.device)
11
- outputs = model.generate(**inputs, max_new_tokens=200)
12
- return tokenizer.decode(outputs[0], skip_special_tokens=True)
13
-
14
- gr.Interface(
15
- fn=chat_io,
16
- inputs=gr.Textbox(placeholder="Pose ta question à IO..."),
17
- outputs="text",
18
- title="IO AI",
19
- description="Assistant IA super puissant basé sur LLaMA 3."
20
- ).launch()
 
 
1
  from transformers import AutoTokenizer, AutoModelForCausalLM
2
 
3
+ model_id = "meta-llama/Meta-Llama-Guard-2-8B"
4
 
5
+ tokenizer = AutoTokenizer.from_pretrained(model_id, use_auth_token=True)
6
+ model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", torch_dtype="auto")
7
 
8
+ def ask_llama(prompt):
9
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
10
+ output = model.generate(**inputs, max_new_tokens=200)
11
+ response = tokenizer.decode(output[0], skip_special_tokens=True)
12
+ return response