rphrp1985 commited on
Commit
19b7e36
·
verified ·
1 Parent(s): 0bfe063

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +14 -14
app.py CHANGED
@@ -50,22 +50,22 @@ huggingface_token = os.getenv("HUGGINGFACE_TOKEN")
50
  # )
51
  # from huggingface_hub import snapshot_download
52
 
53
- # snapshot_download(
54
- # repo_id="stepfun-ai/Step-3.5-Flash-GGUF-Q4_K_S",
55
- # repo_type="model",
56
- # local_dir="./models/stepfun",
57
- # # allow_patterns=["UD-TQ1_0/*"], # 👈 folder inside repo
58
- # token=huggingface_token # only if gated/private
59
- # )
60
 
61
 
62
 
63
- llm = Llama.from_pretrained(
64
- repo_id="stepfun-ai/Step-3.5-Flash-GGUF-Q4_K_S",
65
- filename="step3p5_flash_Q4_K_S-00001-of-00012.gguf",
66
- allow_patterns=["UD-TQ1_0/*.gguf"],
67
- verbose=False
68
- )
69
 
70
 
71
 
@@ -170,7 +170,7 @@ demo = gr.ChatInterface(
170
  "gemma-2-27b-it-Q5_K_M.gguf",
171
  # "2b_it_v2.gguf",
172
  "GLM-4.7-Flash-UD-Q8_K_XL.gguf",
173
- "stepfun/step3p5_flash_Q4_K_S-00001-of-00012.gguf"
174
  ],
175
  value="GLM-4.7-Flash-UD-Q8_K_XL.gguf",
176
  label="Model",
 
50
  # )
51
  # from huggingface_hub import snapshot_download
52
 
53
+ snapshot_download(
54
+ repo_id="stepfun-ai/Step-3.5-Flash-GGUF-Q4_K_S",
55
+ repo_type="model",
56
+ local_dir="./models/stepfun",
57
+ # allow_patterns=["UD-TQ1_0/*"], # 👈 folder inside repo
58
+ token=huggingface_token # only if gated/private
59
+ )
60
 
61
 
62
 
63
+ # llm = Llama.from_pretrained(
64
+ # repo_id="stepfun-ai/Step-3.5-Flash-GGUF-Q4_K_S",
65
+ # filename="step3p5_flash_Q4_K_S-00001-of-00012.gguf",
66
+ # allow_patterns=["UD-TQ1_0/*.gguf"],
67
+ # verbose=False
68
+ # )
69
 
70
 
71
 
 
170
  "gemma-2-27b-it-Q5_K_M.gguf",
171
  # "2b_it_v2.gguf",
172
  "GLM-4.7-Flash-UD-Q8_K_XL.gguf",
173
+ "stepfun"
174
  ],
175
  value="GLM-4.7-Flash-UD-Q8_K_XL.gguf",
176
  label="Model",