Update app.py
Browse files
app.py
CHANGED
|
@@ -50,22 +50,22 @@ huggingface_token = os.getenv("HUGGINGFACE_TOKEN")
|
|
| 50 |
# )
|
| 51 |
# from huggingface_hub import snapshot_download
|
| 52 |
|
| 53 |
-
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
|
| 57 |
-
#
|
| 58 |
-
|
| 59 |
-
|
| 60 |
|
| 61 |
|
| 62 |
|
| 63 |
-
llm = Llama.from_pretrained(
|
| 64 |
-
|
| 65 |
-
|
| 66 |
-
|
| 67 |
-
|
| 68 |
-
)
|
| 69 |
|
| 70 |
|
| 71 |
|
|
@@ -170,7 +170,7 @@ demo = gr.ChatInterface(
|
|
| 170 |
"gemma-2-27b-it-Q5_K_M.gguf",
|
| 171 |
# "2b_it_v2.gguf",
|
| 172 |
"GLM-4.7-Flash-UD-Q8_K_XL.gguf",
|
| 173 |
-
"stepfun
|
| 174 |
],
|
| 175 |
value="GLM-4.7-Flash-UD-Q8_K_XL.gguf",
|
| 176 |
label="Model",
|
|
|
|
| 50 |
# )
|
| 51 |
# from huggingface_hub import snapshot_download
|
| 52 |
|
| 53 |
+
snapshot_download(
|
| 54 |
+
repo_id="stepfun-ai/Step-3.5-Flash-GGUF-Q4_K_S",
|
| 55 |
+
repo_type="model",
|
| 56 |
+
local_dir="./models/stepfun",
|
| 57 |
+
# allow_patterns=["UD-TQ1_0/*"], # 👈 folder inside repo
|
| 58 |
+
token=huggingface_token # only if gated/private
|
| 59 |
+
)
|
| 60 |
|
| 61 |
|
| 62 |
|
| 63 |
+
# llm = Llama.from_pretrained(
|
| 64 |
+
# repo_id="stepfun-ai/Step-3.5-Flash-GGUF-Q4_K_S",
|
| 65 |
+
# filename="step3p5_flash_Q4_K_S-00001-of-00012.gguf",
|
| 66 |
+
# allow_patterns=["UD-TQ1_0/*.gguf"],
|
| 67 |
+
# verbose=False
|
| 68 |
+
# )
|
| 69 |
|
| 70 |
|
| 71 |
|
|
|
|
| 170 |
"gemma-2-27b-it-Q5_K_M.gguf",
|
| 171 |
# "2b_it_v2.gguf",
|
| 172 |
"GLM-4.7-Flash-UD-Q8_K_XL.gguf",
|
| 173 |
+
"stepfun"
|
| 174 |
],
|
| 175 |
value="GLM-4.7-Flash-UD-Q8_K_XL.gguf",
|
| 176 |
label="Model",
|