Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -6,6 +6,7 @@ from llama_cpp import Llama
|
|
| 6 |
import requests
|
| 7 |
import tempfile
|
| 8 |
import json
|
|
|
|
| 9 |
|
| 10 |
app = Flask(__name__)
|
| 11 |
logging.basicConfig(level=logging.INFO)
|
|
@@ -20,6 +21,7 @@ class LLMManager:
|
|
| 20 |
def __init__(self, models_config):
|
| 21 |
self.models = {}
|
| 22 |
self.models_config = models_config
|
|
|
|
| 23 |
self.load_all_models()
|
| 24 |
|
| 25 |
def load_all_models(self):
|
|
|
|
| 6 |
import requests
|
| 7 |
import tempfile
|
| 8 |
import json
|
| 9 |
+
from concurrent.futures import ThreadPoolExecutor
|
| 10 |
|
| 11 |
app = Flask(__name__)
|
| 12 |
logging.basicConfig(level=logging.INFO)
|
|
|
|
| 21 |
def __init__(self, models_config):
|
| 22 |
self.models = {}
|
| 23 |
self.models_config = models_config
|
| 24 |
+
self.executor = ThreadPoolExecutor(max_workers=2)
|
| 25 |
self.load_all_models()
|
| 26 |
|
| 27 |
def load_all_models(self):
|