debug
Browse files
data/models/llama3-1-8b.py
CHANGED
|
@@ -26,7 +26,7 @@ def chat_completion():
|
|
| 26 |
data = request.json
|
| 27 |
|
| 28 |
try:
|
| 29 |
-
print(
|
| 30 |
outputs = pipeline(
|
| 31 |
user_input=data.get('messages', []),
|
| 32 |
max_new_tokens=data.get('max_tokens', 2048),
|
|
|
|
| 26 |
data = request.json
|
| 27 |
|
| 28 |
try:
|
| 29 |
+
print(data.get('messages', []))
|
| 30 |
outputs = pipeline(
|
| 31 |
user_input=data.get('messages', []),
|
| 32 |
max_new_tokens=data.get('max_tokens', 2048),
|