Spaces:
Sleeping
Sleeping
Commit
·
ba3ab1e
1
Parent(s):
6babfdb
Updated to new version llama-cpp
Browse files- requirements.txt +1 -1
- scripts/llm_service.py +4 -4
requirements.txt
CHANGED
|
@@ -3,4 +3,4 @@ requests>=2.31.0
|
|
| 3 |
pathlib2>=2.3.7
|
| 4 |
pinecone>=3.0.0
|
| 5 |
numpy>=1.21.0
|
| 6 |
-
llama-cpp-python==0.
|
|
|
|
| 3 |
pathlib2>=2.3.7
|
| 4 |
pinecone>=3.0.0
|
| 5 |
numpy>=1.21.0
|
| 6 |
+
llama-cpp-python==0.3.16
|
scripts/llm_service.py
CHANGED
|
@@ -199,12 +199,12 @@ Always be helpful, accurate, and focused on the user's specific needs."""
|
|
| 199 |
# )
|
| 200 |
self.llm = Llama(
|
| 201 |
model_path=self.model_path,
|
| 202 |
-
|
| 203 |
seed=42,
|
| 204 |
n_ctx=self.n_ctx,
|
| 205 |
-
|
| 206 |
-
|
| 207 |
-
|
| 208 |
)
|
| 209 |
|
| 210 |
self.is_loaded = True
|
|
|
|
| 199 |
# )
|
| 200 |
self.llm = Llama(
|
| 201 |
model_path=self.model_path,
|
| 202 |
+
cache_dir=Path('models'),
|
| 203 |
seed=42,
|
| 204 |
n_ctx=self.n_ctx,
|
| 205 |
+
verbose=False,
|
| 206 |
+
n_gpu_layers=self.n_gpu_layers,
|
| 207 |
+
n_threads=self.n_threads,
|
| 208 |
)
|
| 209 |
|
| 210 |
self.is_loaded = True
|