ai: Ship internal training data.
Browse files
jarvis.py
CHANGED
|
@@ -39,6 +39,8 @@ from pptx import Presentation
|
|
| 39 |
|
| 40 |
os.system("apt-get update -q -y && apt-get install -q -y tesseract-ocr tesseract-ocr-eng tesseract-ocr-ind libleptonica-dev libtesseract-dev")
|
| 41 |
|
|
|
|
|
|
|
| 42 |
LINUX_SERVER_HOSTS = [host for host in json.loads(os.getenv("LINUX_SERVER_HOST", "[]")) if host]
|
| 43 |
LINUX_SERVER_HOSTS_MARKED = set()
|
| 44 |
LINUX_SERVER_HOSTS_ATTEMPTS = {}
|
|
@@ -158,7 +160,6 @@ async def fetch_response_async(host, provider_key, selected_model, messages, mod
|
|
| 158 |
raise
|
| 159 |
|
| 160 |
async def chat_with_model_async(history, user_input, selected_model_display, sess):
|
| 161 |
-
global ACTIVE_CANDIDATE
|
| 162 |
if not get_available_items(LINUX_SERVER_PROVIDER_KEYS, LINUX_SERVER_PROVIDER_KEYS_MARKED) or not get_available_items(LINUX_SERVER_HOSTS, LINUX_SERVER_HOSTS_MARKED):
|
| 163 |
return RESPONSES["RESPONSE_3"]
|
| 164 |
if not hasattr(sess, "session_id"):
|
|
@@ -167,7 +168,10 @@ async def chat_with_model_async(history, user_input, selected_model_display, ses
|
|
| 167 |
model_config = MODEL_CONFIG.get(selected_model, DEFAULT_CONFIG)
|
| 168 |
messages = [{"role": "user", "content": user} for user, _ in history]
|
| 169 |
messages += [{"role": "assistant", "content": assistant} for _, assistant in history if assistant]
|
|
|
|
|
|
|
| 170 |
messages.append({"role": "user", "content": user_input})
|
|
|
|
| 171 |
if ACTIVE_CANDIDATE is not None:
|
| 172 |
try:
|
| 173 |
return await fetch_response_async(ACTIVE_CANDIDATE[0], ACTIVE_CANDIDATE[1], selected_model, messages, model_config, sess.session_id)
|
|
|
|
| 39 |
|
| 40 |
os.system("apt-get update -q -y && apt-get install -q -y tesseract-ocr tesseract-ocr-eng tesseract-ocr-ind libleptonica-dev libtesseract-dev")
|
| 41 |
|
| 42 |
+
INTERNAL_TRAINING_DATA = os.getenv("INTERNAL_TRAINING_DATA", "")
|
| 43 |
+
|
| 44 |
LINUX_SERVER_HOSTS = [host for host in json.loads(os.getenv("LINUX_SERVER_HOST", "[]")) if host]
|
| 45 |
LINUX_SERVER_HOSTS_MARKED = set()
|
| 46 |
LINUX_SERVER_HOSTS_ATTEMPTS = {}
|
|
|
|
| 160 |
raise
|
| 161 |
|
| 162 |
async def chat_with_model_async(history, user_input, selected_model_display, sess):
|
|
|
|
| 163 |
if not get_available_items(LINUX_SERVER_PROVIDER_KEYS, LINUX_SERVER_PROVIDER_KEYS_MARKED) or not get_available_items(LINUX_SERVER_HOSTS, LINUX_SERVER_HOSTS_MARKED):
|
| 164 |
return RESPONSES["RESPONSE_3"]
|
| 165 |
if not hasattr(sess, "session_id"):
|
|
|
|
| 168 |
model_config = MODEL_CONFIG.get(selected_model, DEFAULT_CONFIG)
|
| 169 |
messages = [{"role": "user", "content": user} for user, _ in history]
|
| 170 |
messages += [{"role": "assistant", "content": assistant} for _, assistant in history if assistant]
|
| 171 |
+
if INTERNAL_TRAINING_DATA:
|
| 172 |
+
messages.insert(0, {"role": "system", "content": INTERNAL_TRAINING_DATA})
|
| 173 |
messages.append({"role": "user", "content": user_input})
|
| 174 |
+
global ACTIVE_CANDIDATE
|
| 175 |
if ACTIVE_CANDIDATE is not None:
|
| 176 |
try:
|
| 177 |
return await fetch_response_async(ACTIVE_CANDIDATE[0], ACTIVE_CANDIDATE[1], selected_model, messages, model_config, sess.session_id)
|