Spaces:
Runtime error
Runtime error
Update agent.py
Browse files
agent.py
CHANGED
|
@@ -125,7 +125,7 @@ tools = [
|
|
| 125 |
]
|
| 126 |
|
| 127 |
# Build graph function
|
| 128 |
-
def build_graph(provider: str = "
|
| 129 |
"""Build the graph"""
|
| 130 |
# Load environment variables from .env file
|
| 131 |
if provider == "google":
|
|
@@ -135,7 +135,6 @@ def build_graph(provider: str = "groq"):
|
|
| 135 |
# Groq https://console.groq.com/docs/models
|
| 136 |
llm = ChatGroq(model="qwen-qwq-32b", temperature=0) # optional : qwen-qwq-32b gemma2-9b-it
|
| 137 |
elif provider == "huggingface":
|
| 138 |
-
# TODO: Add huggingface endpoint
|
| 139 |
llm = ChatHuggingFace(
|
| 140 |
llm=HuggingFaceEndpoint(
|
| 141 |
url="https://api-inference.huggingface.co/models/Meta-DeepLearning/llama-2-7b-chat-hf",
|
|
|
|
| 125 |
]
|
| 126 |
|
| 127 |
# Build graph function
|
| 128 |
+
def build_graph(provider: str = "huggingface"):
|
| 129 |
"""Build the graph"""
|
| 130 |
# Load environment variables from .env file
|
| 131 |
if provider == "google":
|
|
|
|
| 135 |
# Groq https://console.groq.com/docs/models
|
| 136 |
llm = ChatGroq(model="qwen-qwq-32b", temperature=0) # optional : qwen-qwq-32b gemma2-9b-it
|
| 137 |
elif provider == "huggingface":
|
|
|
|
| 138 |
llm = ChatHuggingFace(
|
| 139 |
llm=HuggingFaceEndpoint(
|
| 140 |
url="https://api-inference.huggingface.co/models/Meta-DeepLearning/llama-2-7b-chat-hf",
|