Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -26,13 +26,8 @@ if not os.path.exists("./models"):
|
|
| 26 |
os.makedirs("./models")
|
| 27 |
|
| 28 |
hf_hub_download(
|
| 29 |
-
repo_id="
|
| 30 |
-
filename="
|
| 31 |
-
local_dir="./models",
|
| 32 |
-
)
|
| 33 |
-
hf_hub_download(
|
| 34 |
-
repo_id="bartowski/google_gemma-3-1b-it-GGUF",
|
| 35 |
-
filename="google_gemma-3-1b-it-Q5_K_M.gguf",
|
| 36 |
local_dir="./models",
|
| 37 |
)
|
| 38 |
|
|
@@ -72,7 +67,7 @@ llm_model = None
|
|
| 72 |
def respond(
|
| 73 |
message: str,
|
| 74 |
history: List[Tuple[str, str]],
|
| 75 |
-
model: str = "
|
| 76 |
system_message: str = "You are a helpful assistant.",
|
| 77 |
max_tokens: int = 1024,
|
| 78 |
temperature: float = 0.7,
|
|
@@ -102,7 +97,7 @@ def respond(
|
|
| 102 |
|
| 103 |
# Ensure model is not None
|
| 104 |
if model is None:
|
| 105 |
-
model = "
|
| 106 |
|
| 107 |
# Load the model
|
| 108 |
if llm is None or llm_model != model:
|
|
@@ -184,10 +179,9 @@ demo = gr.ChatInterface(
|
|
| 184 |
additional_inputs=[
|
| 185 |
gr.Dropdown(
|
| 186 |
choices=[
|
| 187 |
-
"
|
| 188 |
-
"google_gemma-3-1b-it-Q5_K_M.gguf",
|
| 189 |
],
|
| 190 |
-
value="
|
| 191 |
label="Model",
|
| 192 |
info="Select the AI model to use for chat",
|
| 193 |
),
|
|
|
|
| 26 |
os.makedirs("./models")
|
| 27 |
|
| 28 |
hf_hub_download(
|
| 29 |
+
repo_id="SRP-base-model-training/gemma_3_800M_sft_v2_translation-kazparc_latest",
|
| 30 |
+
filename="gemma_3_800M_sft_v2_translation-kazparc_latest.gguf",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 31 |
local_dir="./models",
|
| 32 |
)
|
| 33 |
|
|
|
|
| 67 |
def respond(
|
| 68 |
message: str,
|
| 69 |
history: List[Tuple[str, str]],
|
| 70 |
+
model: str = "gemma_3_800M_sft_v2_translation-kazparc_latest.gguf", # Set default model
|
| 71 |
system_message: str = "You are a helpful assistant.",
|
| 72 |
max_tokens: int = 1024,
|
| 73 |
temperature: float = 0.7,
|
|
|
|
| 97 |
|
| 98 |
# Ensure model is not None
|
| 99 |
if model is None:
|
| 100 |
+
model = "gemma_3_800M_sft_v2_translation-kazparc_latest.gguf"
|
| 101 |
|
| 102 |
# Load the model
|
| 103 |
if llm is None or llm_model != model:
|
|
|
|
| 179 |
additional_inputs=[
|
| 180 |
gr.Dropdown(
|
| 181 |
choices=[
|
| 182 |
+
"gemma_3_800M_sft_v2_translation-kazparc_latest.gguf",
|
|
|
|
| 183 |
],
|
| 184 |
+
value="gemma_3_800M_sft_v2_translation-kazparc_latest.gguf",
|
| 185 |
label="Model",
|
| 186 |
info="Select the AI model to use for chat",
|
| 187 |
),
|