rphrp1985 commited on
Commit
64f5b61
·
verified ·
1 Parent(s): e4343e5

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +23 -16
app.py CHANGED
@@ -16,11 +16,11 @@ print(llama_cpp.__version__)
16
 
17
  huggingface_token = os.getenv("HUGGINGFACE_TOKEN")
18
 
19
- hf_hub_download(
20
- repo_id="bartowski/gemma-2-9b-it-GGUF",
21
- filename="gemma-2-9b-it-Q5_K_M.gguf",
22
- local_dir="./models"
23
- )
24
 
25
  hf_hub_download(
26
  repo_id="bartowski/gemma-2-27b-it-GGUF",
@@ -28,16 +28,23 @@ hf_hub_download(
28
  local_dir="./models"
29
  )
30
 
31
- hf_hub_download(
32
- repo_id="google/gemma-2-2b-it-GGUF",
33
- filename="2b_it_v2.gguf",
34
- local_dir="./models",
35
- token=huggingface_token
36
- )
 
 
 
 
 
 
 
37
 
38
  hf_hub_download(
39
- repo_id="unsloth/GLM-4.7-Flash-GGUF",
40
- filename="GLM-4.7-Flash-Q2_K.gguf",
41
  local_dir="./models",
42
  token=huggingface_token
43
  )
@@ -147,10 +154,10 @@ demo = gr.ChatInterface(
147
  additional_inputs=[
148
  gr.Dropdown(
149
  [
150
- "gemma-2-9b-it-Q5_K_M.gguf",
151
  "gemma-2-27b-it-Q5_K_M.gguf",
152
- "2b_it_v2.gguf",
153
- "GLM-4.7-Flash-Q2_K.gguf"
154
  ],
155
  value="2b_it_v2.gguf",
156
  label="Model",
 
16
 
17
  huggingface_token = os.getenv("HUGGINGFACE_TOKEN")
18
 
19
+ # hf_hub_download(
20
+ # repo_id="bartowski/gemma-2-9b-it-GGUF",
21
+ # filename="gemma-2-9b-it-Q5_K_M.gguf",
22
+ # local_dir="./models"
23
+ # )
24
 
25
  hf_hub_download(
26
  repo_id="bartowski/gemma-2-27b-it-GGUF",
 
28
  local_dir="./models"
29
  )
30
 
31
+ # hf_hub_download(
32
+ # repo_id="google/gemma-2-2b-it-GGUF",
33
+ # filename="2b_it_v2.gguf",
34
+ # local_dir="./models",
35
+ # token=huggingface_token
36
+ # )
37
+
38
+ # hf_hub_download(
39
+ # repo_id="unsloth/GLM-4.7-Flash-GGUF",
40
+ # filename="GLM-4.7-Flash-Q2_K.gguf",
41
+ # local_dir="./models",
42
+ # token=huggingface_token
43
+ # )
44
 
45
  hf_hub_download(
46
+ repo_id="TeichAI/GLM-4.7-Flash-Claude-Opus-4.5-High-Reasoning-Distill-GGUF",
47
+ filename="glm-4.7-flash-claude-4.5-opus.q8_0.gguf",
48
  local_dir="./models",
49
  token=huggingface_token
50
  )
 
154
  additional_inputs=[
155
  gr.Dropdown(
156
  [
157
+ # "gemma-2-9b-it-Q5_K_M.gguf",
158
  "gemma-2-27b-it-Q5_K_M.gguf",
159
+ # "2b_it_v2.gguf",
160
+ "glm-4.7-flash-claude-4.5-opus.q8_0.gguf"
161
  ],
162
  value="2b_it_v2.gguf",
163
  label="Model",