Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -131,11 +131,11 @@ def image_generator(prompt: str) -> str:
|
|
| 131 |
return f"Image generation failed: {str(e)}"
|
| 132 |
|
| 133 |
# -------------------- Local LLM (Replaces HfApiModel) --------------------
|
| 134 |
-
from
|
| 135 |
|
| 136 |
-
class
|
| 137 |
"""
|
| 138 |
-
|
| 139 |
"""
|
| 140 |
def __init__(self):
|
| 141 |
self.pipeline = pipeline(
|
|
@@ -159,7 +159,7 @@ class TransformersModel(LocalModel):
|
|
| 159 |
result = self.pipeline(prompt, max_new_tokens=500, do_sample=True)
|
| 160 |
return result[0]['generated_text']
|
| 161 |
|
| 162 |
-
model =
|
| 163 |
|
| 164 |
# -------------------- Agent Setup --------------------
|
| 165 |
final_answer = FinalAnswerTool()
|
|
|
|
| 131 |
return f"Image generation failed: {str(e)}"
|
| 132 |
|
| 133 |
# -------------------- Local LLM (Replaces HfApiModel) --------------------
|
| 134 |
+
from transformers import pipeline
|
| 135 |
|
| 136 |
+
class LocalModel:
|
| 137 |
"""
|
| 138 |
+
Minimal local model interface compatible with smolagents CodeAgent.
|
| 139 |
"""
|
| 140 |
def __init__(self):
|
| 141 |
self.pipeline = pipeline(
|
|
|
|
| 159 |
result = self.pipeline(prompt, max_new_tokens=500, do_sample=True)
|
| 160 |
return result[0]['generated_text']
|
| 161 |
|
| 162 |
+
model = LocalModel()
|
| 163 |
|
| 164 |
# -------------------- Agent Setup --------------------
|
| 165 |
final_answer = FinalAnswerTool()
|