alishabhale commited on
Commit
e29c41b
·
1 Parent(s): a2ff5d8

Add application file

Browse files
Files changed (1) hide show
  1. app.py +23 -4
app.py CHANGED
@@ -1,7 +1,26 @@
1
  import gradio as gr
 
 
2
 
3
- def greet(name):
4
- return "Hello " + name + "!!"
 
 
 
 
5
 
6
- demo = gr.Interface(fn=greet, inputs="text", outputs="text")
7
- demo.launch()
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  import gradio as gr
2
+ from transformers import AutoModelForCausalLM, AutoTokenizer
3
+ import torch
4
 
5
+ # Load Model & Tokenizer
6
+ model_name = "meta-llama/Llama-2-7b-chat-hf"
7
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
8
+ model = AutoModelForCausalLM.from_pretrained(
9
+ model_name, torch_dtype=torch.float16, device_map="auto"
10
+ )
11
 
12
+ # Function to generate responses
13
+ def chat_with_llama(input_text):
14
+ inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
15
+ output = model.generate(**inputs, max_new_tokens=200)
16
+ return tokenizer.decode(output[0], skip_special_tokens=True)
17
+
18
+ # Gradio UI
19
+ iface = gr.Interface(
20
+ fn=chat_with_llama,
21
+ inputs="text",
22
+ outputs="text",
23
+ title="Llama-2 Chatbot"
24
+ )
25
+
26
+ iface.launch()