RanjithaRuttala commited on
Commit
1987c4f
·
verified ·
1 Parent(s): c8723a7

Update handler.py

Browse files
Files changed (1) hide show
  1. handler.py +15 -6
handler.py CHANGED
@@ -1,6 +1,6 @@
1
  import torch
2
- from unsloth import FastLanguageModel
3
- from transformers import AutoTokenizer
4
 
5
  # === Config ===
6
  MODEL_DIR = "./" # Replace with path to your merged model folder on HF
@@ -12,11 +12,20 @@ TOP_P = 0.95
12
  # === Load merged model and tokenizer ===
13
  print("[Handler] Loading model and tokenizer...")
14
  # FastLanguageModel can load merged PEFT models directly
15
- model, tokenizer = FastLanguageModel.from_pretrained(
16
- model_name=MODEL_DIR,
17
- dtype=torch.float16,
18
- load_in_4bit=False
 
 
 
 
 
 
 
 
19
  )
 
20
  model = model.to(DEVICE)
21
  model.eval()
22
 
 
1
  import torch
2
+ # from unsloth import FastLanguageModel
3
+ from transformers import AutoTokenizer,AutoModelForCausalLM
4
 
5
  # === Config ===
6
  MODEL_DIR = "./" # Replace with path to your merged model folder on HF
 
12
  # === Load merged model and tokenizer ===
13
  print("[Handler] Loading model and tokenizer...")
14
  # FastLanguageModel can load merged PEFT models directly
15
+ # model, tokenizer = FastLanguageModel.from_pretrained(
16
+ # model_name=MODEL_DIR,
17
+ # dtype=torch.float16,
18
+ # load_in_4bit=False
19
+ # )
20
+
21
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_DIR)
22
+
23
+ model = AutoModelForCausalLM.from_pretrained(
24
+ MODEL_DIR,
25
+ torch_dtype=torch.float16,
26
+ device_map="auto"
27
  )
28
+
29
  model = model.to(DEVICE)
30
  model.eval()
31