updated nodes
Browse files- hype_pack/utils/nodes.py +3 -3
hype_pack/utils/nodes.py
CHANGED
|
@@ -29,7 +29,7 @@ def build_reference_material_node(interview_state: InterviewState) -> InterviewS
|
|
| 29 |
# temperature=0.1
|
| 30 |
# ).with_structured_output(ReferenceMaterial)
|
| 31 |
llm = ChatGroq(
|
| 32 |
-
model="llama-3.
|
| 33 |
temperature=0.1
|
| 34 |
).with_structured_output(ReferenceMaterial)
|
| 35 |
|
|
@@ -91,7 +91,7 @@ def generate_questions_node(interview_state: InterviewState) -> InterviewState:
|
|
| 91 |
# temperature=0.35
|
| 92 |
# ).with_structured_output(QuestionList)
|
| 93 |
llm = ChatGroq(
|
| 94 |
-
model="llama-3.
|
| 95 |
temperature=0.35
|
| 96 |
).with_structured_output(QuestionList)
|
| 97 |
|
|
@@ -151,7 +151,7 @@ def generate_transcript_node(interview_state: InterviewState, speaker_profile: d
|
|
| 151 |
# temperature=0.6
|
| 152 |
# ).with_structured_output(HypeCastTranscript)
|
| 153 |
llm = ChatGroq(
|
| 154 |
-
model="llama-3.
|
| 155 |
temperature=0.6
|
| 156 |
).with_structured_output(HypeCastTranscript)
|
| 157 |
|
|
|
|
| 29 |
# temperature=0.1
|
| 30 |
# ).with_structured_output(ReferenceMaterial)
|
| 31 |
llm = ChatGroq(
|
| 32 |
+
model="llama-3.1-70b-versatile",
|
| 33 |
temperature=0.1
|
| 34 |
).with_structured_output(ReferenceMaterial)
|
| 35 |
|
|
|
|
| 91 |
# temperature=0.35
|
| 92 |
# ).with_structured_output(QuestionList)
|
| 93 |
llm = ChatGroq(
|
| 94 |
+
model="llama-3.1-70b-versatile",
|
| 95 |
temperature=0.35
|
| 96 |
).with_structured_output(QuestionList)
|
| 97 |
|
|
|
|
| 151 |
# temperature=0.6
|
| 152 |
# ).with_structured_output(HypeCastTranscript)
|
| 153 |
llm = ChatGroq(
|
| 154 |
+
model="llama-3.1-70b-versatile",
|
| 155 |
temperature=0.6
|
| 156 |
).with_structured_output(HypeCastTranscript)
|
| 157 |
|