Mauriciotuks commited on
Commit
526b76f
·
verified ·
1 Parent(s): d3085c9

Update app.py

Browse files

![IMG_6026.png](https://cdn-uploads.huggingface.co/production/uploads/688dcc38d03635bba0fbc3fb/zDXUPrDGXYTGjBnLCPPww.png)

Files changed (1) hide show
  1. app.py +15 -8
app.py CHANGED
@@ -1,13 +1,20 @@
1
  import gradio as gr
2
- from transformers import AutoModelForCausalLM, AutoTokenizer
3
 
4
- model_name = "meta-llama/Meta-Llama-3-8B" # Pour 70B, tu dois avoir une organisation certifiée
5
 
6
- tokenizer = AutoTokenizer.from_pretrained(model_name)
7
- model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
8
 
9
- def chat_io(prompt):inputs = tokenizer(prompt, return_tensors="pt")
10
- outputs = model.generate(**inputs, max_new_tokens=100)
11
- return tokenizer.decode(outputs[0])
 
12
 
13
- gr.Interface(fn=chat_io, inputs="text", outputs="text", title="IO AI").launch()
 
 
 
 
 
 
 
1
  import gradio as gr
2
+ from transformers import AutoTokenizer, AutoModelForCausalLM
3
 
4
+ model = "meta-llama/Meta-Llama-3-8B"
5
 
6
+ tokenizer = AutoTokenizer.from_pretrained(model)
7
+ model = AutoModelForCausalLM.from_pretrained(model, device_map="auto", torch_dtype="auto")
8
 
9
+ def chat_io(message):
10
+ inputs = tokenizer(message, return_tensors="pt").to(model.device)
11
+ outputs = model.generate(**inputs, max_new_tokens=200)
12
+ return tokenizer.decode(outputs[0], skip_special_tokens=True)
13
 
14
+ gr.Interface(
15
+ fn=chat_io,
16
+ inputs=gr.Textbox(placeholder="Pose ta question à IO..."),
17
+ outputs="text",
18
+ title="IO AI",
19
+ description="Assistant IA super puissant basé sur LLaMA 3."
20
+ ).launch()