Spaces:
Sleeping
Sleeping
Update utils.py
Browse files
utils.py
CHANGED
|
@@ -97,7 +97,7 @@ def query_llm_for_additional_info(topic: str, existing_text: str) -> str:
|
|
| 97 |
try:
|
| 98 |
response = groq_client.chat.completions.create(
|
| 99 |
messages=[{"role": "system", "content": system_prompt}],
|
| 100 |
-
model="llama-3.
|
| 101 |
max_tokens=1024,
|
| 102 |
temperature=0.7
|
| 103 |
)
|
|
@@ -305,7 +305,7 @@ def generate_script(system_prompt: str, input_text: str, tone: str, target_lengt
|
|
| 305 |
try:
|
| 306 |
response = groq_client.chat.completions.create(
|
| 307 |
messages=[{"role": "system", "content": prompt}],
|
| 308 |
-
model="llama-3.
|
| 309 |
max_tokens=2048,
|
| 310 |
temperature=0.7
|
| 311 |
)
|
|
|
|
| 97 |
try:
|
| 98 |
response = groq_client.chat.completions.create(
|
| 99 |
messages=[{"role": "system", "content": system_prompt}],
|
| 100 |
+
model="llama-3.3-70b-versatile",
|
| 101 |
max_tokens=1024,
|
| 102 |
temperature=0.7
|
| 103 |
)
|
|
|
|
| 305 |
try:
|
| 306 |
response = groq_client.chat.completions.create(
|
| 307 |
messages=[{"role": "system", "content": prompt}],
|
| 308 |
+
model="llama-3.3-70b-versatile",
|
| 309 |
max_tokens=2048,
|
| 310 |
temperature=0.7
|
| 311 |
)
|