Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -3,14 +3,12 @@ from ctransformers import AutoModelForCausalLM
|
|
| 3 |
from huggingface_hub import hf_hub_download
|
| 4 |
|
| 5 |
# 1. Download the file with the CORRECT name
|
| 6 |
-
# We fixed the filename here to match your screenshot
|
| 7 |
model_path = hf_hub_download(
|
| 8 |
repo_id="XY26/dual-frame-llama-3",
|
| 9 |
filename="meta-llama-3.1-8b-instruct.Q4_K_M.gguf"
|
| 10 |
)
|
| 11 |
|
| 12 |
# 2. Load the model
|
| 13 |
-
# We set model_type="llama3" because you are using the new Llama 3.1
|
| 14 |
llm = AutoModelForCausalLM.from_pretrained(
|
| 15 |
"XY26/dual-frame-llama-3",
|
| 16 |
model_file="meta-llama-3.1-8b-instruct.Q4_K_M.gguf",
|
|
|
|
| 3 |
from huggingface_hub import hf_hub_download
|
| 4 |
|
| 5 |
# 1. Download the file with the CORRECT name
|
|
|
|
| 6 |
model_path = hf_hub_download(
|
| 7 |
repo_id="XY26/dual-frame-llama-3",
|
| 8 |
filename="meta-llama-3.1-8b-instruct.Q4_K_M.gguf"
|
| 9 |
)
|
| 10 |
|
| 11 |
# 2. Load the model
|
|
|
|
| 12 |
llm = AutoModelForCausalLM.from_pretrained(
|
| 13 |
"XY26/dual-frame-llama-3",
|
| 14 |
model_file="meta-llama-3.1-8b-instruct.Q4_K_M.gguf",
|