Spaces:
Build error
Build error
Update agent.py
Browse files
agent.py
CHANGED
|
@@ -1,6 +1,8 @@
|
|
| 1 |
from typing import Any, List, Optional
|
| 2 |
|
| 3 |
-
from smolagents import CodeAgent, InferenceClientModel, HfApiModel, TransformersModel
|
|
|
|
|
|
|
| 4 |
|
| 5 |
import os
|
| 6 |
hf_token = os.getenv("HF_TOKEN")
|
|
@@ -31,9 +33,12 @@ class Agent:
|
|
| 31 |
):
|
| 32 |
logger.info("Initializing Agent")
|
| 33 |
|
| 34 |
-
model_id = "meta-llama/Llama-3.3-70B-Instruct"
|
| 35 |
#self.model = HfApiModel(model_id=model_id, token=hf_token)
|
| 36 |
-
self.model =
|
|
|
|
|
|
|
|
|
|
| 37 |
|
| 38 |
self.tools = tools
|
| 39 |
self.imports = [
|
|
|
|
| 1 |
from typing import Any, List, Optional
|
| 2 |
|
| 3 |
+
from smolagents import CodeAgent, InferenceClientModel, HfApiModel, TransformersModel, LiteLLMModel
|
| 4 |
+
|
| 5 |
+
|
| 6 |
|
| 7 |
import os
|
| 8 |
hf_token = os.getenv("HF_TOKEN")
|
|
|
|
| 33 |
):
|
| 34 |
logger.info("Initializing Agent")
|
| 35 |
|
| 36 |
+
#model_id = "meta-llama/Llama-3.3-70B-Instruct"
|
| 37 |
#self.model = HfApiModel(model_id=model_id, token=hf_token)
|
| 38 |
+
self.model = LiteLLMModel(
|
| 39 |
+
model_id="groq/deepseek-r1-distill-llama-70b",
|
| 40 |
+
temperature=0.2,
|
| 41 |
+
api_key=os.environ["GROQ_API_KEY"])
|
| 42 |
|
| 43 |
self.tools = tools
|
| 44 |
self.imports = [
|