# Copyright (c) 2025 Huawei Technologies Co., Ltd. All rights reserved. export ASCEND_RT_VISIBLE_DEVICES=0,1,2,3 export VLLM_USE_V1=1 export VLLM_WORKER_MULTIPROC_METHOD=fork export VLLM_ENABLE_MC2=0 export USING_LCCL_COM=0 export OMNI_USE_PANGU=1 export ENABLE_PREFILL_TND=1 export HCCL_OP_EXPANSION_MODE="AIV" export VLLM_ALLOW_LONG_MAX_MODEL_LEN=1 export PYTORCH_NPU_ALLOC_CONF=expandable_segments:True export HCCL_RDMA_TIMEOUT=5 export HCCL_DETERMINISTIC=False export ASCEND_GLOBAL_LOG_LEVEL=3 export CPU_AFFINITY_CONF=2 export VLLM_LOGGING_LEVEL=INFO export HCCL_BUFFSIZE=1000 export HCCL_CONNECT_TIMEOUT=1800 export HCCL_EXEC_TIMEOUT=1800 export HCCL_INTRA_ROCE_ENABLE=1 export HCCL_INTRA_PCIE_ENABLE=0 export FORCE_ENABLE_CHUNK_PREFILL=1 export USE_REASONING=${USE_REASONING:=1} export USE_TOOL=${USE_TOOL:=1} if [ "$USE_REASONING" = "1" ]; then reasoning="--reasoning-parser pangu" fi if [ "$USE_TOOL" = "1" ]; then tools="--enable-auto-tool-choice --tool-call-parser pangu" fi export PYTHONPATH=/path/to/omniinfer/:$PYTHONPATH rm -r -f .torchair_cache/ python start_api_servers.py \ --num-servers 1 \ --model-path /path/to/model/ \ --master-ip 0.0.0.0 \ --tp 4 \ --num-dp 1 \ --master-port 3512 \ --served-model-name openpangu_r_72b_2512 \ --log-dir apiserverlog_pangu72B_hybrid_chunk \ --extra-args "--max-num-batched-tokens 2048 --enforce-eager --no-enable-prefix-caching --enable-expert-parallel --max-num-seqs 32 --long-prefill-token-threshold 1024" \ --base-api-port 8000 \ --gpu-util 0.90 \ --no-enable-prefix-caching \ --max-model-len 131072 \ $reasoning \ $tools \ --additional-config '{"graph_model_compile_config":{"level":1, "use_ge_graph_cached":true, "decode_gear_list": [32]}, "enable_hybrid_graph_mode": false, "expert_parallel_size": 4, "expert_tensor_parallel_size": 1}' &