Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 2 |
import gradio as gr
|
| 3 |
|
| 4 |
-
model_name = "
|
| 5 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 6 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
| 7 |
|
|
@@ -29,6 +29,6 @@ def predict(message, history):
|
|
| 29 |
gr.ChatInterface(
|
| 30 |
fn=predict,
|
| 31 |
title="💬 My Chatbot",
|
| 32 |
-
description="A simple CPU-friendly chatbot using
|
| 33 |
examples=["Hello!", "What's your name?", "Tell me a fun fact."],
|
| 34 |
).launch()
|
|
|
|
| 1 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 2 |
import gradio as gr
|
| 3 |
|
| 4 |
+
model_name = "Qwen/Qwen3-0.6B"
|
| 5 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 6 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
| 7 |
|
|
|
|
| 29 |
gr.ChatInterface(
|
| 30 |
fn=predict,
|
| 31 |
title="💬 My Chatbot",
|
| 32 |
+
description="A simple CPU-friendly chatbot using Qwen/Qwen3-0.6B.",
|
| 33 |
examples=["Hello!", "What's your name?", "Tell me a fun fact."],
|
| 34 |
).launch()
|