Update backend.py
Browse files- backend.py +1 -1
backend.py
CHANGED
|
@@ -86,7 +86,7 @@ def generate_testcases(user_story):
|
|
| 86 |
try:
|
| 87 |
# Call the Nvidia llama API with the refined prompt
|
| 88 |
completion = client.chat.completions.create(
|
| 89 |
-
model="meta/llama-3.1-
|
| 90 |
messages=[
|
| 91 |
{"role": "user", "content": prompt}
|
| 92 |
],
|
|
|
|
| 86 |
try:
|
| 87 |
# Call the Nvidia llama API with the refined prompt
|
| 88 |
completion = client.chat.completions.create(
|
| 89 |
+
model="meta/llama-3.1-405b-instruct", # Using llama3.1 405b model
|
| 90 |
messages=[
|
| 91 |
{"role": "user", "content": prompt}
|
| 92 |
],
|