debug
Browse files
data/models/llama3-1-8b.py
CHANGED
|
@@ -22,7 +22,7 @@ pipeline = transformers.pipeline(
|
|
| 22 |
)
|
| 23 |
|
| 24 |
@app.route('/chat', methods=['POST'])
|
| 25 |
-
@spaces.GPU()
|
| 26 |
def chat_completion():
|
| 27 |
data = request.json
|
| 28 |
|
|
|
|
| 22 |
)
|
| 23 |
|
| 24 |
@app.route('/chat', methods=['POST'])
|
| 25 |
+
@spaces.GPU(enable_queue=True)
|
| 26 |
def chat_completion():
|
| 27 |
data = request.json
|
| 28 |
|