Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -7,8 +7,8 @@ import torch
|
|
| 7 |
tokenizer = AutoTokenizer.from_pretrained("unsloth/Qwen3-1.7B")
|
| 8 |
base_model = AutoModelForCausalLM.from_pretrained(
|
| 9 |
"unsloth/Qwen3-1.7B",
|
| 10 |
-
torch_dtype=torch.
|
| 11 |
-
device_map="
|
| 12 |
)
|
| 13 |
model = PeftModel.from_pretrained(base_model, "khazarai/BioGenesis-ToT")
|
| 14 |
|
|
@@ -26,12 +26,12 @@ def generate_response(user_input, chat_history):
|
|
| 26 |
)
|
| 27 |
|
| 28 |
# Tokenize and send to GPU
|
| 29 |
-
inputs = tokenizer(text, return_tensors="pt").to("
|
| 30 |
|
| 31 |
# Generate
|
| 32 |
output_tokens = model.generate(
|
| 33 |
**inputs,
|
| 34 |
-
max_new_tokens=
|
| 35 |
temperature=0.6,
|
| 36 |
top_p=0.95,
|
| 37 |
top_k=20,
|
|
|
|
| 7 |
tokenizer = AutoTokenizer.from_pretrained("unsloth/Qwen3-1.7B")
|
| 8 |
base_model = AutoModelForCausalLM.from_pretrained(
|
| 9 |
"unsloth/Qwen3-1.7B",
|
| 10 |
+
torch_dtype=torch.float32,
|
| 11 |
+
device_map="cpu"
|
| 12 |
)
|
| 13 |
model = PeftModel.from_pretrained(base_model, "khazarai/BioGenesis-ToT")
|
| 14 |
|
|
|
|
| 26 |
)
|
| 27 |
|
| 28 |
# Tokenize and send to GPU
|
| 29 |
+
inputs = tokenizer(text, return_tensors="pt").to("cpu")
|
| 30 |
|
| 31 |
# Generate
|
| 32 |
output_tokens = model.generate(
|
| 33 |
**inputs,
|
| 34 |
+
max_new_tokens=2200,
|
| 35 |
temperature=0.6,
|
| 36 |
top_p=0.95,
|
| 37 |
top_k=20,
|