Update app.py
Browse files
app.py
CHANGED
|
@@ -83,19 +83,21 @@ def setup_model(model_key, intial=False):
|
|
| 83 |
model_config = get_model_config(model_key)
|
| 84 |
log_debug(f"update_model() --> Model config: {model_config}")
|
| 85 |
|
| 86 |
-
|
| 87 |
-
|
| 88 |
-
|
| 89 |
-
|
| 90 |
-
|
| 91 |
-
|
|
|
|
|
|
|
| 92 |
|
| 93 |
openai_client = OpenAI(
|
| 94 |
-
api_key=
|
| 95 |
base_url=base_url
|
| 96 |
)
|
| 97 |
model_config['base_url'] = base_url
|
| 98 |
-
log_debug(f"Switched to model {model_key} using endpoint {base_url}")
|
| 99 |
|
| 100 |
_model_hf_name = model_config.get("MODEL_HF_URL").split('https://huggingface.co/')[1]
|
| 101 |
_link = f"<a href='{model_config.get('MODEL_HF_URL')}{COMMUNITY_POSTFIX_URL}' target='_blank'>{_model_hf_name}</a>"
|
|
@@ -134,7 +136,11 @@ def run_chat_inference(history, message, state):
|
|
| 134 |
state["is_streaming"] = True
|
| 135 |
state["stop_flag"] = False
|
| 136 |
error = None
|
| 137 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 138 |
temperature = model_config.get('TEMPERATURE', DEFAULT_MODEL_TEMPERATURE)
|
| 139 |
output_tag_start = model_config.get('OUTPUT_TAG_START', "[BEGIN FINAL RESPONSE]")
|
| 140 |
output_tag_end = model_config.get('OUTPUT_TAG_END', "[END FINAL RESPONSE]")
|
|
@@ -659,4 +665,4 @@ with gr.Blocks(theme=theme, css=custom_css) as demo:
|
|
| 659 |
)
|
| 660 |
|
| 661 |
demo.queue(default_concurrency_limit=2).launch(ssr_mode=False, show_api=False, max_file_size="10mb")
|
| 662 |
-
log_info("Gradio app launched")
|
|
|
|
| 83 |
model_config = get_model_config(model_key)
|
| 84 |
log_debug(f"update_model() --> Model config: {model_config}")
|
| 85 |
|
| 86 |
+
# ENVIRONMENT VARIABLE CONFIGURATION
|
| 87 |
+
base_url = os.environ.get("API_BASE_URL")
|
| 88 |
+
api_key = os.environ.get("API_KEY")
|
| 89 |
+
|
| 90 |
+
if not base_url:
|
| 91 |
+
raise ValueError("API_BASE_URL environment variable not set")
|
| 92 |
+
if not api_key:
|
| 93 |
+
raise ValueError("API_KEY environment variable not set")
|
| 94 |
|
| 95 |
openai_client = OpenAI(
|
| 96 |
+
api_key=api_key,
|
| 97 |
base_url=base_url
|
| 98 |
)
|
| 99 |
model_config['base_url'] = base_url
|
| 100 |
+
log_debug(f"Switched to model {model_key} using endpoint {base_url} (ENV VARS)")
|
| 101 |
|
| 102 |
_model_hf_name = model_config.get("MODEL_HF_URL").split('https://huggingface.co/')[1]
|
| 103 |
_link = f"<a href='{model_config.get('MODEL_HF_URL')}{COMMUNITY_POSTFIX_URL}' target='_blank'>{_model_hf_name}</a>"
|
|
|
|
| 136 |
state["is_streaming"] = True
|
| 137 |
state["stop_flag"] = False
|
| 138 |
error = None
|
| 139 |
+
# ENVIRONMENT VARIABLE MODEL
|
| 140 |
+
model_name = os.environ.get("API_MODEL")
|
| 141 |
+
if not model_name:
|
| 142 |
+
raise ValueError("API_MODEL environment variable not set")
|
| 143 |
+
# model_name = model_config.get('MODEL_NAME')
|
| 144 |
temperature = model_config.get('TEMPERATURE', DEFAULT_MODEL_TEMPERATURE)
|
| 145 |
output_tag_start = model_config.get('OUTPUT_TAG_START', "[BEGIN FINAL RESPONSE]")
|
| 146 |
output_tag_end = model_config.get('OUTPUT_TAG_END', "[END FINAL RESPONSE]")
|
|
|
|
| 665 |
)
|
| 666 |
|
| 667 |
demo.queue(default_concurrency_limit=2).launch(ssr_mode=False, show_api=False, max_file_size="10mb")
|
| 668 |
+
log_info("Gradio app launched")
|