Spaces:
Sleeping
Sleeping
tokens
Browse files- app/model_loader.py +3 -3
app/model_loader.py
CHANGED
|
@@ -17,7 +17,7 @@ def load_model():
|
|
| 17 |
|
| 18 |
base_model = AutoModelForCausalLM.from_pretrained(
|
| 19 |
"meta-llama/Llama-2-7b-chat-hf",
|
| 20 |
-
|
| 21 |
cache_dir="/tmp/hf_cache",
|
| 22 |
torch_dtype="auto",
|
| 23 |
device_map="auto"
|
|
@@ -25,12 +25,12 @@ def load_model():
|
|
| 25 |
model = PeftModel.from_pretrained(
|
| 26 |
base_model,
|
| 27 |
"BrainGPT/BrainGPT-7B-v0.1",
|
| 28 |
-
|
| 29 |
cache_dir="/tmp/hf_cache"
|
| 30 |
)
|
| 31 |
tokenizer = AutoTokenizer.from_pretrained(
|
| 32 |
"meta-llama/Llama-2-7b-chat-hf",
|
| 33 |
-
|
| 34 |
cache_dir="/tmp/hf_cache"
|
| 35 |
)
|
| 36 |
return model, tokenizer
|
|
|
|
| 17 |
|
| 18 |
base_model = AutoModelForCausalLM.from_pretrained(
|
| 19 |
"meta-llama/Llama-2-7b-chat-hf",
|
| 20 |
+
token=hf_token,
|
| 21 |
cache_dir="/tmp/hf_cache",
|
| 22 |
torch_dtype="auto",
|
| 23 |
device_map="auto"
|
|
|
|
| 25 |
model = PeftModel.from_pretrained(
|
| 26 |
base_model,
|
| 27 |
"BrainGPT/BrainGPT-7B-v0.1",
|
| 28 |
+
token=hf_token,
|
| 29 |
cache_dir="/tmp/hf_cache"
|
| 30 |
)
|
| 31 |
tokenizer = AutoTokenizer.from_pretrained(
|
| 32 |
"meta-llama/Llama-2-7b-chat-hf",
|
| 33 |
+
token=hf_token,
|
| 34 |
cache_dir="/tmp/hf_cache"
|
| 35 |
)
|
| 36 |
return model, tokenizer
|