Update app.py
Browse files
app.py
CHANGED
|
@@ -33,6 +33,17 @@ internals.LlamaModel.__del__ = safe_del
|
|
| 33 |
|
| 34 |
|
| 35 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 36 |
|
| 37 |
|
| 38 |
|
|
@@ -96,12 +107,12 @@ huggingface_token = os.getenv("HUGGINGFACE_TOKEN")
|
|
| 96 |
|
| 97 |
|
| 98 |
|
| 99 |
-
|
| 100 |
-
|
| 101 |
-
|
| 102 |
-
|
| 103 |
-
|
| 104 |
-
|
| 105 |
|
| 106 |
|
| 107 |
# hf_hub_download(
|
|
@@ -126,13 +137,17 @@ from huggingface_hub import snapshot_download
|
|
| 126 |
# token=huggingface_token # only if gated/private
|
| 127 |
# )
|
| 128 |
|
| 129 |
-
|
| 130 |
-
|
| 131 |
-
|
| 132 |
-
|
| 133 |
-
|
| 134 |
-
|
| 135 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 136 |
|
| 137 |
|
| 138 |
|
|
@@ -302,7 +317,7 @@ demo = gr.ChatInterface(
|
|
| 302 |
"Qwen3-VL-32B-Thinking-Q8_0.gguf",
|
| 303 |
"Q8_0/gpt-oss-120b-Q8_0-00001-of-00002.gguf"
|
| 304 |
],
|
| 305 |
-
value="
|
| 306 |
label="Model",
|
| 307 |
),
|
| 308 |
gr.Textbox(
|
|
|
|
| 33 |
|
| 34 |
|
| 35 |
|
| 36 |
+
##### final verdict
|
| 37 |
+
|
| 38 |
+
# GLM 4.7 flash fast infrence
|
| 39 |
+
#qwen 3 VL
|
| 40 |
+
#mini max 2,5
|
| 41 |
+
# qwen 3 coder next
|
| 42 |
+
#gpt oss 120B
|
| 43 |
+
#qwen 3 next 80b
|
| 44 |
+
|
| 45 |
+
|
| 46 |
+
|
| 47 |
|
| 48 |
|
| 49 |
|
|
|
|
| 107 |
|
| 108 |
|
| 109 |
|
| 110 |
+
hf_hub_download(
|
| 111 |
+
repo_id="unsloth/Qwen3-Next-80B-A3B-Instruct-GGUF",
|
| 112 |
+
filename="Qwen3-Next-80B-A3B-Instruct-Q4_K_M.gguf",
|
| 113 |
+
local_dir="./models",
|
| 114 |
+
token=huggingface_token
|
| 115 |
+
)
|
| 116 |
|
| 117 |
|
| 118 |
# hf_hub_download(
|
|
|
|
| 137 |
# token=huggingface_token # only if gated/private
|
| 138 |
# )
|
| 139 |
|
| 140 |
+
|
| 141 |
+
|
| 142 |
+
|
| 143 |
+
#### Deploy Minimax 2.5 insplace of gpt oss 120b its larger . and better and more recet leeases
|
| 144 |
+
# snapshot_download(
|
| 145 |
+
# repo_id="unsloth/gpt-oss-120b-GGUF",
|
| 146 |
+
# repo_type="model",
|
| 147 |
+
# local_dir="./models/",
|
| 148 |
+
# allow_patterns=["Q8_0/*"], # 👈 folder inside repo
|
| 149 |
+
# token=huggingface_token # only if gated/private
|
| 150 |
+
# )
|
| 151 |
|
| 152 |
|
| 153 |
|
|
|
|
| 317 |
"Qwen3-VL-32B-Thinking-Q8_0.gguf",
|
| 318 |
"Q8_0/gpt-oss-120b-Q8_0-00001-of-00002.gguf"
|
| 319 |
],
|
| 320 |
+
value="Qwen3-Next-80B-A3B-Instruct-Q4_K_M.gguf",
|
| 321 |
label="Model",
|
| 322 |
),
|
| 323 |
gr.Textbox(
|