Update README.md
Browse files
README.md
CHANGED
|
@@ -200,13 +200,13 @@ The minimum hardware requirements for deploying Intern-S1 series models are:
|
|
| 200 |
|
| 201 |
You can utilize one of the following LLM inference frameworks to create an OpenAI compatible server:
|
| 202 |
|
| 203 |
-
#### [lmdeploy(>=0.9.2)](https://github.com/InternLM/lmdeploy)
|
| 204 |
|
| 205 |
```bash
|
| 206 |
lmdeploy serve api_server internlm/Intern-S1 --reasoning-parser intern-s1 --tool-call-parser intern-s1 --tp 8
|
| 207 |
```
|
| 208 |
|
| 209 |
-
#### [vllm](https://github.com/vllm-project/vllm)
|
| 210 |
|
| 211 |
```bash
|
| 212 |
vllm serve internlm/Intern-S1 --tensor-parallel-size 8 --trust-remote-code
|
|
|
|
| 200 |
|
| 201 |
You can utilize one of the following LLM inference frameworks to create an OpenAI compatible server:
|
| 202 |
|
| 203 |
+
#### [lmdeploy (>=0.9.2)](https://github.com/InternLM/lmdeploy)
|
| 204 |
|
| 205 |
```bash
|
| 206 |
lmdeploy serve api_server internlm/Intern-S1 --reasoning-parser intern-s1 --tool-call-parser intern-s1 --tp 8
|
| 207 |
```
|
| 208 |
|
| 209 |
+
#### [vllm (>=0.10.1)](https://github.com/vllm-project/vllm)
|
| 210 |
|
| 211 |
```bash
|
| 212 |
vllm serve internlm/Intern-S1 --tensor-parallel-size 8 --trust-remote-code
|