Upload app.py with huggingface_hub
Browse files
app.py
CHANGED
|
@@ -47,17 +47,17 @@ def _get_llm_and_path() -> Dict[str, Any]:
|
|
| 47 |
repo_id=MODEL_REPO_ID, filename=MODEL_FILE, repo_type="model"
|
| 48 |
)
|
| 49 |
|
| 50 |
-
init_kwargs: Dict[str, Any] = {
|
| 51 |
-
"model_path": model_path,
|
| 52 |
-
"n_ctx": N_CTX,
|
| 53 |
-
"n_threads": N_THREADS,
|
| 54 |
-
"n_batch": N_BATCH,
|
| 55 |
-
"n_gpu_layers": 0,
|
| 56 |
-
"verbose": False,
|
| 57 |
-
"use_mmap": USE_MMAP,
|
| 58 |
-
}
|
| 59 |
-
if CHAT_FORMAT:
|
| 60 |
-
init_kwargs["chat_format"] = CHAT_FORMAT
|
| 61 |
|
| 62 |
llm = Llama(**init_kwargs)
|
| 63 |
return {"llm": llm, "model_path": model_path}
|
|
|
|
| 47 |
repo_id=MODEL_REPO_ID, filename=MODEL_FILE, repo_type="model"
|
| 48 |
)
|
| 49 |
|
| 50 |
+
init_kwargs: Dict[str, Any] = {
|
| 51 |
+
"model_path": model_path,
|
| 52 |
+
"n_ctx": N_CTX,
|
| 53 |
+
"n_threads": N_THREADS,
|
| 54 |
+
"n_batch": N_BATCH,
|
| 55 |
+
"n_gpu_layers": 0,
|
| 56 |
+
"verbose": False,
|
| 57 |
+
"use_mmap": USE_MMAP,
|
| 58 |
+
}
|
| 59 |
+
if CHAT_FORMAT:
|
| 60 |
+
init_kwargs["chat_format"] = CHAT_FORMAT
|
| 61 |
|
| 62 |
llm = Llama(**init_kwargs)
|
| 63 |
return {"llm": llm, "model_path": model_path}
|