Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -7,10 +7,11 @@ chat_model_state = None
|
|
| 7 |
chat_tokenizer_state = None
|
| 8 |
|
| 9 |
def load_chat_model():
|
| 10 |
-
"""Función para cargar el modelo de chat"""
|
| 11 |
global chat_model_state, chat_tokenizer_state
|
| 12 |
try:
|
| 13 |
model_name = "Qwen/Qwen2.5-3B-Instruct"
|
|
|
|
| 14 |
chat_model_state = AutoModelForCausalLM.from_pretrained(
|
| 15 |
model_name,
|
| 16 |
torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
|
|
@@ -22,7 +23,7 @@ def load_chat_model():
|
|
| 22 |
print(f"Error al cargar el modelo de chat: {e}")
|
| 23 |
|
| 24 |
def generate_response(messages, model, tokenizer):
|
| 25 |
-
"""Genera una respuesta usando el modelo de chat"""
|
| 26 |
try:
|
| 27 |
if model is None or tokenizer is None:
|
| 28 |
raise ValueError("El modelo de chat o el tokenizer no están cargados.")
|
|
@@ -63,7 +64,7 @@ def generate_response(messages, model, tokenizer):
|
|
| 63 |
|
| 64 |
# Gradio Interface
|
| 65 |
with gr.Blocks() as app_chat:
|
| 66 |
-
gr.Markdown("### Chatbot Simple")
|
| 67 |
chatbot_interface = gr.Chatbot(label="Conversación")
|
| 68 |
text_input_chat = gr.Textbox(label="Escribe tu mensaje", lines=1)
|
| 69 |
send_btn_chat = gr.Button("Enviar")
|
|
@@ -119,5 +120,10 @@ with gr.Blocks() as app_chat:
|
|
| 119 |
# Cargar el modelo al iniciar
|
| 120 |
load_chat_model()
|
| 121 |
|
| 122 |
-
# Ejecutar la aplicación
|
| 123 |
-
app_chat.launch(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
chat_tokenizer_state = None
|
| 8 |
|
| 9 |
def load_chat_model():
|
| 10 |
+
"""Función para cargar el modelo de chat."""
|
| 11 |
global chat_model_state, chat_tokenizer_state
|
| 12 |
try:
|
| 13 |
model_name = "Qwen/Qwen2.5-3B-Instruct"
|
| 14 |
+
# Cargar el modelo en CPU o GPU según disponibilidad
|
| 15 |
chat_model_state = AutoModelForCausalLM.from_pretrained(
|
| 16 |
model_name,
|
| 17 |
torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
|
|
|
|
| 23 |
print(f"Error al cargar el modelo de chat: {e}")
|
| 24 |
|
| 25 |
def generate_response(messages, model, tokenizer):
|
| 26 |
+
"""Genera una respuesta usando el modelo de chat."""
|
| 27 |
try:
|
| 28 |
if model is None or tokenizer is None:
|
| 29 |
raise ValueError("El modelo de chat o el tokenizer no están cargados.")
|
|
|
|
| 64 |
|
| 65 |
# Gradio Interface
|
| 66 |
with gr.Blocks() as app_chat:
|
| 67 |
+
gr.Markdown("### Chatbot Simple (CPU/GPU Compatible)")
|
| 68 |
chatbot_interface = gr.Chatbot(label="Conversación")
|
| 69 |
text_input_chat = gr.Textbox(label="Escribe tu mensaje", lines=1)
|
| 70 |
send_btn_chat = gr.Button("Enviar")
|
|
|
|
| 120 |
# Cargar el modelo al iniciar
|
| 121 |
load_chat_model()
|
| 122 |
|
| 123 |
+
# Ejecutar la aplicación con configuración segura
|
| 124 |
+
app_chat.queue(api_open=True).launch(
|
| 125 |
+
server_name="0.0.0.0", # Permite acceso local
|
| 126 |
+
server_port=7860, # Puerto local
|
| 127 |
+
share=True, # Crea un enlace público
|
| 128 |
+
enable_queue=True # Activa la cola de procesamiento
|
| 129 |
+
)
|