Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -69,12 +69,11 @@ final_answer = FinalAnswerTool()
|
|
| 69 |
# model_id='Qwen/Qwen2.5-Coder-32B-Instruct',# it is possible that this model may be overloaded
|
| 70 |
# custom_role_conversions=None,
|
| 71 |
# )
|
| 72 |
-
LiteLLMModel._turn_on_debug()
|
| 73 |
|
| 74 |
model = LiteLLMModel(
|
| 75 |
# model_id="ollama_chat/qwen2:7b", # Or try other Ollama-supported models
|
| 76 |
# model_id="ollama/qwen2:7b", # Or try other Ollama-supported models
|
| 77 |
-
model_id="
|
| 78 |
api_base="http://localhost:11434/api/generate", # Default Ollama local server
|
| 79 |
# api_base="http://0.0.0.0:11434",
|
| 80 |
num_ctx=8192,
|
|
|
|
| 69 |
# model_id='Qwen/Qwen2.5-Coder-32B-Instruct',# it is possible that this model may be overloaded
|
| 70 |
# custom_role_conversions=None,
|
| 71 |
# )
|
|
|
|
| 72 |
|
| 73 |
model = LiteLLMModel(
|
| 74 |
# model_id="ollama_chat/qwen2:7b", # Or try other Ollama-supported models
|
| 75 |
# model_id="ollama/qwen2:7b", # Or try other Ollama-supported models
|
| 76 |
+
model_id="ollama/gemma3:1b",
|
| 77 |
api_base="http://localhost:11434/api/generate", # Default Ollama local server
|
| 78 |
# api_base="http://0.0.0.0:11434",
|
| 79 |
num_ctx=8192,
|