Ilke Ileri commited on
Commit
25839d0
·
1 Parent(s): 692ef6b

Add HF_TOKEN support for gated model access

Browse files
Files changed (1) hide show
  1. app.py +7 -3
app.py CHANGED
@@ -8,23 +8,27 @@ import os
8
  app = Flask(__name__)
9
  CORS(app)
10
 
 
 
 
11
  # Model adları
12
  MODEL_NAME = "ilkeileri/gemma-sales-comprehensive"
13
  BASE_MODEL = "google/gemma-1.1-2b-it"
14
 
15
  print("Loading tokenizer...")
16
- tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL, trust_remote_code=True)
17
 
18
  print("Loading base model...")
19
  base_model = AutoModelForCausalLM.from_pretrained(
20
  BASE_MODEL,
21
  dtype=torch.float16,
22
  low_cpu_mem_usage=True,
23
- trust_remote_code=True
 
24
  )
25
 
26
  print("Loading LoRA adapters...")
27
- model = PeftModel.from_pretrained(base_model, MODEL_NAME)
28
  model.eval()
29
 
30
  print("Model loaded successfully!")
 
8
  app = Flask(__name__)
9
  CORS(app)
10
 
11
+ # Hugging Face token'ı environment variable'dan al
12
+ HF_TOKEN = os.environ.get("HF_TOKEN")
13
+
14
  # Model adları
15
  MODEL_NAME = "ilkeileri/gemma-sales-comprehensive"
16
  BASE_MODEL = "google/gemma-1.1-2b-it"
17
 
18
  print("Loading tokenizer...")
19
+ tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL, trust_remote_code=True, token=HF_TOKEN)
20
 
21
  print("Loading base model...")
22
  base_model = AutoModelForCausalLM.from_pretrained(
23
  BASE_MODEL,
24
  dtype=torch.float16,
25
  low_cpu_mem_usage=True,
26
+ trust_remote_code=True,
27
+ token=HF_TOKEN
28
  )
29
 
30
  print("Loading LoRA adapters...")
31
+ model = PeftModel.from_pretrained(base_model, MODEL_NAME, token=HF_TOKEN)
32
  model.eval()
33
 
34
  print("Model loaded successfully!")