debug
Browse files
data/models/llama3-1-8b.py
CHANGED
|
@@ -28,7 +28,7 @@ def chat_completion():
|
|
| 28 |
try:
|
| 29 |
print(data.get('messages', []))
|
| 30 |
outputs = pipeline(
|
| 31 |
-
|
| 32 |
max_new_tokens=data.get('max_tokens', 2048),
|
| 33 |
temperature=data.get('temperature', 0.7),
|
| 34 |
top_p=data.get('top_p', 0.95)
|
|
|
|
| 28 |
try:
|
| 29 |
print(data.get('messages', []))
|
| 30 |
outputs = pipeline(
|
| 31 |
+
text_inputs=data.get('messages', []),
|
| 32 |
max_new_tokens=data.get('max_tokens', 2048),
|
| 33 |
temperature=data.get('temperature', 0.7),
|
| 34 |
top_p=data.get('top_p', 0.95)
|