applesilicon commited on
Commit
6370771
·
verified ·
1 Parent(s): 8d93622

docs: fix vLLM tensor parallel flag in deploy guide

Browse files

The vLLM CLI uses --tensor-parallel-size (or -tp) for tensor parallelism, not --tp. The current example in deploy_guidance.md fails with unrecognized arguments: --tp. This PR updates only the vLLM command example; the SGLang example remains unchanged.

Files changed (1) hide show
  1. docs/deploy_guidance.md +1 -1
docs/deploy_guidance.md CHANGED
@@ -15,7 +15,7 @@ uv pip install -U vllm \
15
 
16
  Here is the example to serve this model on a H200 single node with TP8 via vLLM:
17
  ```bash
18
- vllm serve $MODEL_PATH --tp 8 --trust-remote-code --tool-call-parser kimi_k2 --reasoning-parser kimi_k2
19
  ```
20
  **Key notes**
21
  - `--tool-call-parser kimi_k2`: Required for enabling tool calling
 
15
 
16
  Here is the example to serve this model on a H200 single node with TP8 via vLLM:
17
  ```bash
18
+ vllm serve $MODEL_PATH -tp 8 --trust-remote-code --tool-call-parser kimi_k2 --reasoning-parser kimi_k2
19
  ```
20
  **Key notes**
21
  - `--tool-call-parser kimi_k2`: Required for enabling tool calling