Update my_agent.py
Browse files- my_agent.py +10 -1
my_agent.py
CHANGED
|
@@ -25,13 +25,22 @@ class AgentResponseState(AgentState):
|
|
| 25 |
class BasicAgent:
|
| 26 |
def __init__(self):
|
| 27 |
model = HuggingFaceEndpoint(
|
| 28 |
-
repo_id="
|
| 29 |
task="text-generation",
|
| 30 |
max_new_tokens=512,
|
| 31 |
do_sample=False,
|
| 32 |
repetition_penalty=1.03,
|
| 33 |
)
|
| 34 |
llm = ChatHuggingFace(llm=model, verbose=True)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
# llm = ChatOllama(
|
| 36 |
# model="qwen3:0.6b",
|
| 37 |
# api_base="http://localhost:11434", # replace with
|
|
|
|
| 25 |
class BasicAgent:
|
| 26 |
def __init__(self):
|
| 27 |
model = HuggingFaceEndpoint(
|
| 28 |
+
repo_id="qwen3:0.6b",
|
| 29 |
task="text-generation",
|
| 30 |
max_new_tokens=512,
|
| 31 |
do_sample=False,
|
| 32 |
repetition_penalty=1.03,
|
| 33 |
)
|
| 34 |
llm = ChatHuggingFace(llm=model, verbose=True)
|
| 35 |
+
|
| 36 |
+
model = HuggingFaceEndpoint(
|
| 37 |
+
repo_id="TinyLlama/TinyLlama-1.1B-Chat-v1.0",
|
| 38 |
+
task="text-generation",
|
| 39 |
+
max_new_tokens=512,
|
| 40 |
+
do_sample=False,
|
| 41 |
+
repetition_penalty=1.03,
|
| 42 |
+
)
|
| 43 |
+
llm_qa = ChatHuggingFace(llm=model, verbose=True)
|
| 44 |
# llm = ChatOllama(
|
| 45 |
# model="qwen3:0.6b",
|
| 46 |
# api_base="http://localhost:11434", # replace with
|