Update app.py
Browse files
app.py
CHANGED
|
@@ -35,8 +35,8 @@ import subprocess
|
|
| 35 |
|
| 36 |
subprocess.run([
|
| 37 |
"huggingface-cli", "download",
|
| 38 |
-
"
|
| 39 |
-
"
|
| 40 |
"--local-dir", "./models",
|
| 41 |
"--local-dir-use-symlinks", "False"
|
| 42 |
], check=True)
|
|
@@ -58,7 +58,7 @@ embeddings = HuggingFaceEmbeddings(
|
|
| 58 |
# Load Mistral GGUF via llama.cpp (CPU optimized)
|
| 59 |
# ------------------------------
|
| 60 |
llm_cpp = Llama(
|
| 61 |
-
model_path="./models/
|
| 62 |
n_ctx=2048,
|
| 63 |
n_threads=4, # Adjust based on your CPU cores
|
| 64 |
n_gpu_layers=0, # Force CPU-only
|
|
|
|
| 35 |
|
| 36 |
subprocess.run([
|
| 37 |
"huggingface-cli", "download",
|
| 38 |
+
"microsoft/Phi-3-mini-4k-instruct-gguf",
|
| 39 |
+
"Phi-3-mini-4k-instruct-gguf",
|
| 40 |
"--local-dir", "./models",
|
| 41 |
"--local-dir-use-symlinks", "False"
|
| 42 |
], check=True)
|
|
|
|
| 58 |
# Load Mistral GGUF via llama.cpp (CPU optimized)
|
| 59 |
# ------------------------------
|
| 60 |
llm_cpp = Llama(
|
| 61 |
+
model_path="./models/Phi-3-mini-4k-instruct-gguf",
|
| 62 |
n_ctx=2048,
|
| 63 |
n_threads=4, # Adjust based on your CPU cores
|
| 64 |
n_gpu_layers=0, # Force CPU-only
|