Update app.py
Browse files
app.py
CHANGED
|
@@ -48,15 +48,25 @@ huggingface_token = os.getenv("HUGGINGFACE_TOKEN")
|
|
| 48 |
# local_dir="./models",
|
| 49 |
# token=huggingface_token
|
| 50 |
# )
|
| 51 |
-
|
| 52 |
-
|
| 53 |
-
|
| 54 |
-
repo_id="unsloth/
|
| 55 |
-
|
| 56 |
-
local_dir="./models
|
| 57 |
-
allow_patterns=["UD-IQ1_S/*"], # 👈 folder inside repo
|
| 58 |
-
token=huggingface_token # only if gated/private
|
| 59 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 60 |
|
| 61 |
|
| 62 |
# llm = Llama.from_pretrained(
|
|
@@ -204,10 +214,10 @@ demo = gr.ChatInterface(
|
|
| 204 |
"gemma-2-27b-it-Q5_K_M.gguf",
|
| 205 |
# "2b_it_v2.gguf",
|
| 206 |
"GLM-4.7-Flash-UD-Q8_K_XL.gguf",
|
| 207 |
-
"
|
| 208 |
"gpt-oss-20b-Q4_K_M.gguf"
|
| 209 |
],
|
| 210 |
-
value="
|
| 211 |
label="Model",
|
| 212 |
),
|
| 213 |
gr.Textbox(
|
|
|
|
| 48 |
# local_dir="./models",
|
| 49 |
# token=huggingface_token
|
| 50 |
# )
|
| 51 |
+
|
| 52 |
+
|
| 53 |
+
hf_hub_download(
|
| 54 |
+
repo_id="unsloth/Qwen3-Coder-Next-GGUF",
|
| 55 |
+
filename="Qwen3-Coder-Next-Q4_K_S.gguf",
|
| 56 |
+
local_dir="./models"
|
|
|
|
|
|
|
| 57 |
)
|
| 58 |
+
# from huggingface_hub import snapshot_download
|
| 59 |
+
|
| 60 |
+
# snapshot_download(
|
| 61 |
+
# repo_id="unsloth/MiniMax-M2.5-GGUF",
|
| 62 |
+
# repo_type="model",
|
| 63 |
+
# local_dir="./models/minmax",
|
| 64 |
+
# allow_patterns=["UD-IQ1_S/*"], # 👈 folder inside repo
|
| 65 |
+
# token=huggingface_token # only if gated/private
|
| 66 |
+
# )
|
| 67 |
+
|
| 68 |
+
|
| 69 |
+
|
| 70 |
|
| 71 |
|
| 72 |
# llm = Llama.from_pretrained(
|
|
|
|
| 214 |
"gemma-2-27b-it-Q5_K_M.gguf",
|
| 215 |
# "2b_it_v2.gguf",
|
| 216 |
"GLM-4.7-Flash-UD-Q8_K_XL.gguf",
|
| 217 |
+
"Qwen3-Coder-Next-Q4_K_S.gguf",
|
| 218 |
"gpt-oss-20b-Q4_K_M.gguf"
|
| 219 |
],
|
| 220 |
+
value="Qwen3-Coder-Next-Q4_K_S.gguf",
|
| 221 |
label="Model",
|
| 222 |
),
|
| 223 |
gr.Textbox(
|