Spaces:
Build error
Build error
| if [ ${CONFIG_PATH} ] || [[ ${MODEL_PATH} && ${CONFIG_PATH} ]]; then | |
| python -m llama_cpp.server --config_file ${CONFIG_PATH} | |
| exit 0 | |
| fi | |
| if [ "${MODEL_HF_PRETRAINED_MODEL}" == "None" ]; then | |
| MODEL_HF_PRETRAINED_MODEL="" | |
| fi | |
| if [ ${MODEL_PATH} ]; then | |
| python -m llama_cpp.server \ | |
| --model ${MODEL_PATH} \ | |
| --host ${HOST:=0.0.0.0} \ | |
| --port ${PORT:=8001} \ | |
| --n_gpu_layers ${GPU_LAYERS:=0} \ | |
| --clip_model_path ${CLIP_MODEL_PATH:=None} \ | |
| --chat_format ${MODEL_CHAT_FORMAT:=llama-2} \ | |
| ${PRETRAINED_MODEL_PATH:=} \ | |
| ${MODEL_HF_PRETRAINED_MODEL:+--hf_pretrained_model_name_or_path ${MODEL_HF_PRETRAINED_MODEL}} \ | |
| --interrupt_requests ${INTERRUPT_REQUESTS:=False} | |
| exit 0 | |
| fi | |
| echo "Please set either a CONFIG_PATH or a MODEL_PATH" | |
| exit 1 | |