File size: 808 Bytes
bd49995
eec43e1
1aa0476
eec43e1
1aa0476
eec43e1
 
1aa0476
eec43e1
 
 
 
 
 
 
 
d286610
eec43e1
 
 
9c0f6f5
eec43e1
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
import gradio as gr
from transformers import AutoTokenizer, AutoModelForCausalLM

MODEL_ID = "INTERX/Qwen2.5-GenX-7B"

tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(MODEL_ID, device_map="auto", torch_dtype="auto", trust_remote_code=True)

def chat(user_input):
    messages = [{"role": "user", "content": user_input}]
    inputs = tokenizer.apply_chat_template(
        messages,
        tokenize=True,
        add_generation_prompt=True,
        return_tensors="pt"
    ).to(model.device)

    output = model.generate(inputs, max_new_tokens=512)
    response = tokenizer.decode(output[0], skip_special_tokens=True)
    return response

gr.Interface(fn=chat, inputs="text", outputs="text", title="Chat con Qwen2.5-GenX-7B").launch()