bsny commited on
Commit
b5434da
·
verified ·
1 Parent(s): eccc79e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -2
app.py CHANGED
@@ -1,15 +1,17 @@
 
1
  from transformers import AutoTokenizer, AutoModelForCausalLM
2
  import torch
3
 
4
  model_id = "hugging-quants/Meta-Llama-3.1-8B-Instruct-GPTQ-INT4"
 
5
 
6
- tokenizer = AutoTokenizer.from_pretrained(model_id, use_auth_token=True)
7
  model = AutoModelForCausalLM.from_pretrained(
8
  model_id,
9
  device_map="auto",
10
  torch_dtype=torch.float16,
11
  low_cpu_mem_usage=True,
12
- use_auth_token=True # ⬅️ ensures gated model access
13
  )
14
 
15
  def generate(prompt):
 
1
+ import os
2
  from transformers import AutoTokenizer, AutoModelForCausalLM
3
  import torch
4
 
5
  model_id = "hugging-quants/Meta-Llama-3.1-8B-Instruct-GPTQ-INT4"
6
+ hf_token = os.environ.get("HF_TOKEN")
7
 
8
+ tokenizer = AutoTokenizer.from_pretrained(model_id, token=hf_token)
9
  model = AutoModelForCausalLM.from_pretrained(
10
  model_id,
11
  device_map="auto",
12
  torch_dtype=torch.float16,
13
  low_cpu_mem_usage=True,
14
+ token=hf_token
15
  )
16
 
17
  def generate(prompt):