Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -26,10 +26,6 @@ MODELS = [
|
|
| 26 |
"url": "https://huggingface.co/Qwen/Qwen2.5-3B-Instruct-GGUF/resolve/main/qwen2.5-3b-instruct-q4_k_m.gguf",
|
| 27 |
"name": "qwen2.5-3b-instruct"
|
| 28 |
},
|
| 29 |
-
{
|
| 30 |
-
"url": "https://huggingface.co/Novaciano/Llama-3.2_1b_Uncensored_RP_Aesir_GGUF/resolve/main/Llama-3.2_1b_Uncensored_RP_Aesir.gguf",
|
| 31 |
-
"name": "llama-3.2-1b-rp"
|
| 32 |
-
},
|
| 33 |
{
|
| 34 |
"url": "https://huggingface.co/PurpleAILAB/Llama-3.2-3B-Instruct-uncensored-LoRA_final-Q4_K_M-GGUF/resolve/main/unsloth.Q4_K_M.gguf",
|
| 35 |
"name": "llama-3.2-3b-instruct-uncensored-lora"
|
|
@@ -67,7 +63,7 @@ class LLMManager:
|
|
| 67 |
llm_instance = Llama(
|
| 68 |
model_path=temp_path,
|
| 69 |
n_ctx=MAX_CONTEXT_TOKENS,
|
| 70 |
-
n_batch=
|
| 71 |
n_threads=2,
|
| 72 |
n_threads_batch=2,
|
| 73 |
use_mlock=True,
|
|
|
|
| 26 |
"url": "https://huggingface.co/Qwen/Qwen2.5-3B-Instruct-GGUF/resolve/main/qwen2.5-3b-instruct-q4_k_m.gguf",
|
| 27 |
"name": "qwen2.5-3b-instruct"
|
| 28 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
| 29 |
{
|
| 30 |
"url": "https://huggingface.co/PurpleAILAB/Llama-3.2-3B-Instruct-uncensored-LoRA_final-Q4_K_M-GGUF/resolve/main/unsloth.Q4_K_M.gguf",
|
| 31 |
"name": "llama-3.2-3b-instruct-uncensored-lora"
|
|
|
|
| 63 |
llm_instance = Llama(
|
| 64 |
model_path=temp_path,
|
| 65 |
n_ctx=MAX_CONTEXT_TOKENS,
|
| 66 |
+
n_batch=512,
|
| 67 |
n_threads=2,
|
| 68 |
n_threads_batch=2,
|
| 69 |
use_mlock=True,
|