openPangu-R-72B-2512 / examples /start_serving_openpangu_r_72b_2512.sh
wangrongsheng's picture
Upload folder using huggingface_hub
dec3707 verified
# Copyright (c) 2025 Huawei Technologies Co., Ltd. All rights reserved.
export ASCEND_RT_VISIBLE_DEVICES=0,1,2,3
export VLLM_USE_V1=1
export VLLM_WORKER_MULTIPROC_METHOD=fork
export VLLM_ENABLE_MC2=0
export USING_LCCL_COM=0
export OMNI_USE_PANGU=1
export ENABLE_PREFILL_TND=1
export HCCL_OP_EXPANSION_MODE="AIV"
export VLLM_ALLOW_LONG_MAX_MODEL_LEN=1
export PYTORCH_NPU_ALLOC_CONF=expandable_segments:True
export HCCL_RDMA_TIMEOUT=5
export HCCL_DETERMINISTIC=False
export ASCEND_GLOBAL_LOG_LEVEL=3
export CPU_AFFINITY_CONF=2
export VLLM_LOGGING_LEVEL=INFO
export HCCL_BUFFSIZE=1000
export HCCL_CONNECT_TIMEOUT=1800
export HCCL_EXEC_TIMEOUT=1800
export HCCL_INTRA_ROCE_ENABLE=1
export HCCL_INTRA_PCIE_ENABLE=0
export FORCE_ENABLE_CHUNK_PREFILL=1
export USE_REASONING=${USE_REASONING:=1}
export USE_TOOL=${USE_TOOL:=1}
if [ "$USE_REASONING" = "1" ]; then
reasoning="--reasoning-parser pangu"
fi
if [ "$USE_TOOL" = "1" ]; then
tools="--enable-auto-tool-choice --tool-call-parser pangu"
fi
export PYTHONPATH=/path/to/omniinfer/:$PYTHONPATH
rm -r -f .torchair_cache/
python start_api_servers.py \
--num-servers 1 \
--model-path /path/to/model/ \
--master-ip 0.0.0.0 \
--tp 4 \
--num-dp 1 \
--master-port 3512 \
--served-model-name openpangu_r_72b_2512 \
--log-dir apiserverlog_pangu72B_hybrid_chunk \
--extra-args "--max-num-batched-tokens 2048 --enforce-eager --no-enable-prefix-caching --enable-expert-parallel --max-num-seqs 32 --long-prefill-token-threshold 1024" \
--base-api-port 8000 \
--gpu-util 0.90 \
--no-enable-prefix-caching \
--max-model-len 131072 \
$reasoning \
$tools \
--additional-config '{"graph_model_compile_config":{"level":1, "use_ge_graph_cached":true, "decode_gear_list": [32]}, "enable_hybrid_graph_mode": false, "expert_parallel_size": 4, "expert_tensor_parallel_size": 1}' &