courage17340 applesilicon commited on
Commit
907b3ee
·
verified ·
1 Parent(s): 8d93622

docs: fix vLLM tensor parallel flag in deploy guide (#24)

Browse files

- docs: fix vLLM tensor parallel flag in deploy guide (6370771a8cdf0d76bc05a637d80846ef6dc127c1)


Co-authored-by: applesilicon <applesilicon@users.noreply.huggingface.co>

Files changed (1) hide show
  1. docs/deploy_guidance.md +1 -1
docs/deploy_guidance.md CHANGED
@@ -15,7 +15,7 @@ uv pip install -U vllm \
15
 
16
  Here is the example to serve this model on a H200 single node with TP8 via vLLM:
17
  ```bash
18
- vllm serve $MODEL_PATH --tp 8 --trust-remote-code --tool-call-parser kimi_k2 --reasoning-parser kimi_k2
19
  ```
20
  **Key notes**
21
  - `--tool-call-parser kimi_k2`: Required for enabling tool calling
 
15
 
16
  Here is the example to serve this model on a H200 single node with TP8 via vLLM:
17
  ```bash
18
+ vllm serve $MODEL_PATH -tp 8 --trust-remote-code --tool-call-parser kimi_k2 --reasoning-parser kimi_k2
19
  ```
20
  **Key notes**
21
  - `--tool-call-parser kimi_k2`: Required for enabling tool calling