Commit
·
8d8b9ba
1
Parent(s):
b80eb73
Update src/backend/chatbot.py
Browse files- src/backend/chatbot.py +7 -6
src/backend/chatbot.py
CHANGED
|
@@ -18,7 +18,7 @@ from llama_index.llms.llama_utils import (
|
|
| 18 |
messages_to_prompt,
|
| 19 |
completion_to_prompt,
|
| 20 |
)
|
| 21 |
-
import
|
| 22 |
|
| 23 |
# set version
|
| 24 |
# st.session_state.demo_lite = False
|
|
@@ -33,11 +33,12 @@ print("BP 4 ")
|
|
| 33 |
def init_llm(model, demo_lite):
|
| 34 |
# st.write("BP 4.1: model: ", model)
|
| 35 |
if demo_lite == False:
|
| 36 |
-
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
|
|
|
| 41 |
print("BP 5 : running full demo")
|
| 42 |
if model == "Llama2-7b_CPP":
|
| 43 |
model_path = "src/models/llama-2-7b-chat.Q4_K_M.gguf"
|
|
|
|
| 18 |
messages_to_prompt,
|
| 19 |
completion_to_prompt,
|
| 20 |
)
|
| 21 |
+
from huggingface_hub import HfApi, SpaceHardware
|
| 22 |
|
| 23 |
# set version
|
| 24 |
# st.session_state.demo_lite = False
|
|
|
|
| 33 |
def init_llm(model, demo_lite):
|
| 34 |
# st.write("BP 4.1: model: ", model)
|
| 35 |
if demo_lite == False:
|
| 36 |
+
api = HfApi(token=HF_TOKEN)
|
| 37 |
+
runtime = api.get_space_runtime(repo_id=TRAINING_SPACE_ID)
|
| 38 |
+
# Check if Space is loaded with a GPU.
|
| 39 |
+
if runtime.hardware == SpaceHardware.T4_SMALL:
|
| 40 |
+
print("using nvidia T4 small")
|
| 41 |
+
|
| 42 |
print("BP 5 : running full demo")
|
| 43 |
if model == "Llama2-7b_CPP":
|
| 44 |
model_path = "src/models/llama-2-7b-chat.Q4_K_M.gguf"
|