sharp8 commited on
Commit
eec43e1
Β·
verified Β·
1 Parent(s): bd49995

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +16 -11
app.py CHANGED
@@ -1,17 +1,22 @@
1
  import gradio as gr
2
- from transformers import AutoModelForCausalLM, AutoTokenizer
3
 
4
- model_id = "INTERX/Qwen2.5-GenX-14B"
5
 
6
- tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
7
- model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", torch_dtype="auto", trust_remote_code=True)
8
 
9
- def chat(prompt):
10
- messages = [{"role": "user", "content": prompt}]
11
- tokenized_chat = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors='pt').to(model.device)
12
- generated_ids = model.generate(tokenized_chat, max_new_tokens=512)
13
- response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
14
- return response
 
 
15
 
16
- gr.Interface(fn=chat, inputs="text", outputs="text", title="Qwen2.5-GenX-14B Chat").launch()
 
 
17
 
 
 
1
  import gradio as gr
2
+ from transformers import AutoTokenizer, AutoModelForCausalLM
3
 
4
+ MODEL_ID = "INTERX/Qwen2.5-GenX-7B"
5
 
6
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
7
+ model = AutoModelForCausalLM.from_pretrained(MODEL_ID, device_map="auto", torch_dtype="auto", trust_remote_code=True)
8
 
9
+ def chat(user_input):
10
+ messages = [{"role": "user", "content": user_input}]
11
+ inputs = tokenizer.apply_chat_template(
12
+ messages,
13
+ tokenize=True,
14
+ add_generation_prompt=True,
15
+ return_tensors="pt"
16
+ ).to(model.device)
17
 
18
+ output = model.generate(inputs, max_new_tokens=512)
19
+ response = tokenizer.decode(output[0], skip_special_tokens=True)
20
+ return response
21
 
22
+ gr.Interface(fn=chat, inputs="text", outputs="text", title="Chat con Qwen2.5-GenX-7B").launch()