Spaces:
Sleeping
Sleeping
Upload folder using huggingface_hub
Browse files
agent.py
CHANGED
|
@@ -1,9 +1,6 @@
|
|
| 1 |
from smolagents import CodeAgent, InferenceClientModel, MCPClient
|
| 2 |
from mcp import StdioServerParameters
|
| 3 |
import os
|
| 4 |
-
from dotenv import load_dotenv
|
| 5 |
-
|
| 6 |
-
load_dotenv()
|
| 7 |
|
| 8 |
def create_agent_with_mcp():
|
| 9 |
# Path to server
|
|
@@ -33,7 +30,6 @@ def create_agent_with_mcp():
|
|
| 33 |
# Using InferenceClientModel with Llama 4 Scout
|
| 34 |
model = InferenceClientModel(
|
| 35 |
model_id="meta-llama/Llama-4-Scout-17B-16E-Instruct",
|
| 36 |
-
api_key=os.getenv("HF_TOKEN"),
|
| 37 |
)
|
| 38 |
|
| 39 |
agent = CodeAgent(
|
|
|
|
| 1 |
from smolagents import CodeAgent, InferenceClientModel, MCPClient
|
| 2 |
from mcp import StdioServerParameters
|
| 3 |
import os
|
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
def create_agent_with_mcp():
|
| 6 |
# Path to server
|
|
|
|
| 30 |
# Using InferenceClientModel with Llama 4 Scout
|
| 31 |
model = InferenceClientModel(
|
| 32 |
model_id="meta-llama/Llama-4-Scout-17B-16E-Instruct",
|
|
|
|
| 33 |
)
|
| 34 |
|
| 35 |
agent = CodeAgent(
|