NV9523 commited on
Commit
cef3fdf
·
verified ·
1 Parent(s): 079004b

Update model_loader.py

Browse files
Files changed (1) hide show
  1. model_loader.py +8 -11
model_loader.py CHANGED
@@ -1,28 +1,25 @@
1
  import os
2
  import torch
3
- from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
4
 
5
  HF_KEY = os.getenv("HF_Key")
6
- MODEL_NAME = "meta-llama/Llama-3.2-1B-Instruct" # bạn có thể đổi model nhỏ hơn nếu chậm
7
 
8
  def load_model():
9
  print("Đang load model:", MODEL_NAME)
10
 
11
- tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, token=HF_KEY)
 
 
 
12
 
13
  model = AutoModelForCausalLM.from_pretrained(
14
  MODEL_NAME,
15
- torch_dtype=torch.float32, # CPU nên dùng float32, nếu GPU có thể dùng float16
16
  low_cpu_mem_usage=True,
17
  device_map="cpu",
18
  token=HF_KEY
19
  )
20
 
21
- pipe = pipeline(
22
- "text-generation",
23
- model=model,
24
- tokenizer=tokenizer,
25
- )
26
-
27
  print("Model loaded thành công")
28
- return pipe
 
1
  import os
2
  import torch
3
+ from transformers import AutoModelForCausalLM, AutoTokenizer
4
 
5
  HF_KEY = os.getenv("HF_Key")
6
+ MODEL_NAME = "meta-llama/Llama-3.2-1B-Instruct"
7
 
8
  def load_model():
9
  print("Đang load model:", MODEL_NAME)
10
 
11
+ tokenizer = AutoTokenizer.from_pretrained(
12
+ MODEL_NAME,
13
+ token=HF_KEY
14
+ )
15
 
16
  model = AutoModelForCausalLM.from_pretrained(
17
  MODEL_NAME,
18
+ torch_dtype=torch.float32, # CPU nên để float32
19
  low_cpu_mem_usage=True,
20
  device_map="cpu",
21
  token=HF_KEY
22
  )
23
 
 
 
 
 
 
 
24
  print("Model loaded thành công")
25
+ return model, tokenizer