Rustamshry commited on
Commit
15d7736
·
verified ·
1 Parent(s): 53ab31c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -7,8 +7,8 @@ import torch
7
  tokenizer = AutoTokenizer.from_pretrained("unsloth/Qwen3-1.7B")
8
  base_model = AutoModelForCausalLM.from_pretrained(
9
  "unsloth/Qwen3-1.7B",
10
- torch_dtype=torch.float16,
11
- device_map="auto"
12
  )
13
  model = PeftModel.from_pretrained(base_model, "khazarai/BioGenesis-ToT")
14
 
@@ -26,12 +26,12 @@ def generate_response(user_input, chat_history):
26
  )
27
 
28
  # Tokenize and send to GPU
29
- inputs = tokenizer(text, return_tensors="pt").to("cuda")
30
 
31
  # Generate
32
  output_tokens = model.generate(
33
  **inputs,
34
- max_new_tokens=1200,
35
  temperature=0.6,
36
  top_p=0.95,
37
  top_k=20,
 
7
  tokenizer = AutoTokenizer.from_pretrained("unsloth/Qwen3-1.7B")
8
  base_model = AutoModelForCausalLM.from_pretrained(
9
  "unsloth/Qwen3-1.7B",
10
+ torch_dtype=torch.float32,
11
+ device_map="cpu"
12
  )
13
  model = PeftModel.from_pretrained(base_model, "khazarai/BioGenesis-ToT")
14
 
 
26
  )
27
 
28
  # Tokenize and send to GPU
29
+ inputs = tokenizer(text, return_tensors="pt").to("cpu")
30
 
31
  # Generate
32
  output_tokens = model.generate(
33
  **inputs,
34
+ max_new_tokens=2200,
35
  temperature=0.6,
36
  top_p=0.95,
37
  top_k=20,