rphrp1985 commited on
Commit
eb42bac
·
verified ·
1 Parent(s): af2801b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +29 -14
app.py CHANGED
@@ -33,6 +33,17 @@ internals.LlamaModel.__del__ = safe_del
33
 
34
 
35
 
 
 
 
 
 
 
 
 
 
 
 
36
 
37
 
38
 
@@ -96,12 +107,12 @@ huggingface_token = os.getenv("HUGGINGFACE_TOKEN")
96
 
97
 
98
 
99
- # hf_hub_download(
100
- # repo_id="unsloth/Qwen3-Next-80B-A3B-Instruct-GGUF",
101
- # filename="Qwen3-Next-80B-A3B-Instruct-Q4_K_M.gguf",
102
- # local_dir="./models",
103
- # token=huggingface_token
104
- # )
105
 
106
 
107
  # hf_hub_download(
@@ -126,13 +137,17 @@ from huggingface_hub import snapshot_download
126
  # token=huggingface_token # only if gated/private
127
  # )
128
 
129
- snapshot_download(
130
- repo_id="unsloth/gpt-oss-120b-GGUF",
131
- repo_type="model",
132
- local_dir="./models/",
133
- allow_patterns=["Q8_0/*"], # 👈 folder inside repo
134
- token=huggingface_token # only if gated/private
135
- )
 
 
 
 
136
 
137
 
138
 
@@ -302,7 +317,7 @@ demo = gr.ChatInterface(
302
  "Qwen3-VL-32B-Thinking-Q8_0.gguf",
303
  "Q8_0/gpt-oss-120b-Q8_0-00001-of-00002.gguf"
304
  ],
305
- value="Q8_0/gpt-oss-120b-Q8_0-00001-of-00002.gguf",
306
  label="Model",
307
  ),
308
  gr.Textbox(
 
33
 
34
 
35
 
36
+ ##### final verdict
37
+
38
+ # GLM 4.7 flash fast infrence
39
+ #qwen 3 VL
40
+ #mini max 2,5
41
+ # qwen 3 coder next
42
+ #gpt oss 120B
43
+ #qwen 3 next 80b
44
+
45
+
46
+
47
 
48
 
49
 
 
107
 
108
 
109
 
110
+ hf_hub_download(
111
+ repo_id="unsloth/Qwen3-Next-80B-A3B-Instruct-GGUF",
112
+ filename="Qwen3-Next-80B-A3B-Instruct-Q4_K_M.gguf",
113
+ local_dir="./models",
114
+ token=huggingface_token
115
+ )
116
 
117
 
118
  # hf_hub_download(
 
137
  # token=huggingface_token # only if gated/private
138
  # )
139
 
140
+
141
+
142
+
143
+ #### Deploy Minimax 2.5 insplace of gpt oss 120b its larger . and better and more recet leeases
144
+ # snapshot_download(
145
+ # repo_id="unsloth/gpt-oss-120b-GGUF",
146
+ # repo_type="model",
147
+ # local_dir="./models/",
148
+ # allow_patterns=["Q8_0/*"], # 👈 folder inside repo
149
+ # token=huggingface_token # only if gated/private
150
+ # )
151
 
152
 
153
 
 
317
  "Qwen3-VL-32B-Thinking-Q8_0.gguf",
318
  "Q8_0/gpt-oss-120b-Q8_0-00001-of-00002.gguf"
319
  ],
320
+ value="Qwen3-Next-80B-A3B-Instruct-Q4_K_M.gguf",
321
  label="Model",
322
  ),
323
  gr.Textbox(