Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -3,16 +3,17 @@ import os
|
|
| 3 |
import logging
|
| 4 |
import threading
|
| 5 |
import time
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
from llama_cpp import Llama
|
| 7 |
import requests
|
| 8 |
import tempfile
|
| 9 |
import json
|
| 10 |
import gc
|
| 11 |
from concurrent.futures import ThreadPoolExecutor
|
| 12 |
-
import subprocess
|
| 13 |
-
import sys
|
| 14 |
-
|
| 15 |
-
subprocess.check_call([sys.executable, "-m", "pip", "install", "llama-cpp-python==0.3.15"])
|
| 16 |
|
| 17 |
app = Flask(__name__)
|
| 18 |
logging.basicConfig(level=logging.INFO)
|
|
|
|
| 3 |
import logging
|
| 4 |
import threading
|
| 5 |
import time
|
| 6 |
+
|
| 7 |
+
import subprocess
|
| 8 |
+
import sys
|
| 9 |
+
subprocess.check_call([sys.executable, "-m", "pip", "install", "llama-cpp-python==0.3.15"])
|
| 10 |
+
|
| 11 |
from llama_cpp import Llama
|
| 12 |
import requests
|
| 13 |
import tempfile
|
| 14 |
import json
|
| 15 |
import gc
|
| 16 |
from concurrent.futures import ThreadPoolExecutor
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
|
| 18 |
app = Flask(__name__)
|
| 19 |
logging.basicConfig(level=logging.INFO)
|