Beibars003 commited on
Commit
2bb85c9
·
verified ·
1 Parent(s): 9649a70

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -12
app.py CHANGED
@@ -26,13 +26,8 @@ if not os.path.exists("./models"):
26
  os.makedirs("./models")
27
 
28
  hf_hub_download(
29
- repo_id="bartowski/google_gemma-3-1b-it-GGUF",
30
- filename="google_gemma-3-1b-it-Q4_K_M.gguf",
31
- local_dir="./models",
32
- )
33
- hf_hub_download(
34
- repo_id="bartowski/google_gemma-3-1b-it-GGUF",
35
- filename="google_gemma-3-1b-it-Q5_K_M.gguf",
36
  local_dir="./models",
37
  )
38
 
@@ -72,7 +67,7 @@ llm_model = None
72
  def respond(
73
  message: str,
74
  history: List[Tuple[str, str]],
75
- model: str = "google_gemma-3-1b-it-Q4_K_M.gguf", # Set default model
76
  system_message: str = "You are a helpful assistant.",
77
  max_tokens: int = 1024,
78
  temperature: float = 0.7,
@@ -102,7 +97,7 @@ def respond(
102
 
103
  # Ensure model is not None
104
  if model is None:
105
- model = "google_gemma-3-1b-it-Q4_K_M.gguf"
106
 
107
  # Load the model
108
  if llm is None or llm_model != model:
@@ -184,10 +179,9 @@ demo = gr.ChatInterface(
184
  additional_inputs=[
185
  gr.Dropdown(
186
  choices=[
187
- "google_gemma-3-1b-it-Q4_K_M.gguf",
188
- "google_gemma-3-1b-it-Q5_K_M.gguf",
189
  ],
190
- value="google_gemma-3-1b-it-Q4_K_M.gguf",
191
  label="Model",
192
  info="Select the AI model to use for chat",
193
  ),
 
26
  os.makedirs("./models")
27
 
28
  hf_hub_download(
29
+ repo_id="SRP-base-model-training/gemma_3_800M_sft_v2_translation-kazparc_latest",
30
+ filename="gemma_3_800M_sft_v2_translation-kazparc_latest.gguf",
 
 
 
 
 
31
  local_dir="./models",
32
  )
33
 
 
67
  def respond(
68
  message: str,
69
  history: List[Tuple[str, str]],
70
+ model: str = "gemma_3_800M_sft_v2_translation-kazparc_latest.gguf", # Set default model
71
  system_message: str = "You are a helpful assistant.",
72
  max_tokens: int = 1024,
73
  temperature: float = 0.7,
 
97
 
98
  # Ensure model is not None
99
  if model is None:
100
+ model = "gemma_3_800M_sft_v2_translation-kazparc_latest.gguf"
101
 
102
  # Load the model
103
  if llm is None or llm_model != model:
 
179
  additional_inputs=[
180
  gr.Dropdown(
181
  choices=[
182
+ "gemma_3_800M_sft_v2_translation-kazparc_latest.gguf",
 
183
  ],
184
+ value="gemma_3_800M_sft_v2_translation-kazparc_latest.gguf",
185
  label="Model",
186
  info="Select the AI model to use for chat",
187
  ),