File size: 1,943 Bytes
dec3707
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
# Copyright (c) 2025 Huawei Technologies Co., Ltd. All rights reserved.

export ASCEND_RT_VISIBLE_DEVICES=0,1,2,3
export VLLM_USE_V1=1
export VLLM_WORKER_MULTIPROC_METHOD=fork
export VLLM_ENABLE_MC2=0
export USING_LCCL_COM=0

export OMNI_USE_PANGU=1
export ENABLE_PREFILL_TND=1

export HCCL_OP_EXPANSION_MODE="AIV"
export VLLM_ALLOW_LONG_MAX_MODEL_LEN=1
export PYTORCH_NPU_ALLOC_CONF=expandable_segments:True

export HCCL_RDMA_TIMEOUT=5
export HCCL_DETERMINISTIC=False
export ASCEND_GLOBAL_LOG_LEVEL=3
export CPU_AFFINITY_CONF=2
export VLLM_LOGGING_LEVEL=INFO

export HCCL_BUFFSIZE=1000
export HCCL_CONNECT_TIMEOUT=1800
export HCCL_EXEC_TIMEOUT=1800
export HCCL_INTRA_ROCE_ENABLE=1
export HCCL_INTRA_PCIE_ENABLE=0
export FORCE_ENABLE_CHUNK_PREFILL=1

export USE_REASONING=${USE_REASONING:=1}
export USE_TOOL=${USE_TOOL:=1}
if [ "$USE_REASONING" = "1" ]; then
    reasoning="--reasoning-parser pangu"
fi
if [ "$USE_TOOL" = "1" ]; then
    tools="--enable-auto-tool-choice --tool-call-parser pangu"
fi

export PYTHONPATH=/path/to/omniinfer/:$PYTHONPATH
rm -r -f .torchair_cache/

python start_api_servers.py \
        --num-servers 1 \
        --model-path /path/to/model/ \
        --master-ip 0.0.0.0 \
        --tp 4 \
        --num-dp 1 \
        --master-port 3512 \
        --served-model-name openpangu_r_72b_2512 \
        --log-dir apiserverlog_pangu72B_hybrid_chunk \
        --extra-args "--max-num-batched-tokens 2048 --enforce-eager --no-enable-prefix-caching --enable-expert-parallel --max-num-seqs 32 --long-prefill-token-threshold 1024" \
        --base-api-port 8000 \
        --gpu-util 0.90 \
        --no-enable-prefix-caching \
        --max-model-len 131072 \
        $reasoning \
        $tools \
        --additional-config '{"graph_model_compile_config":{"level":1, "use_ge_graph_cached":true, "decode_gear_list": [32]}, "enable_hybrid_graph_mode": false, "expert_parallel_size": 4, "expert_tensor_parallel_size": 1}' &