Update README.md
Browse files
README.md
CHANGED
|
@@ -122,13 +122,13 @@ model_name = "IQuest/IQuest-Coder-V1-40B-Thinking"
|
|
| 122 |
For production deployment, you can use vLLM to create an OpenAI-compatible API endpoint. Please refer to the [vLLM PR](https://github.com/vllm-project/vllm/pull/31575/files) for implementation details.
|
| 123 |
|
| 124 |
```bash
|
| 125 |
-
vllm serve
|
| 126 |
```
|
| 127 |
|
| 128 |
For Thinking models with reasoning support:
|
| 129 |
|
| 130 |
```bash
|
| 131 |
-
vllm serve
|
| 132 |
```
|
| 133 |
|
| 134 |
|
|
|
|
| 122 |
For production deployment, you can use vLLM to create an OpenAI-compatible API endpoint. Please refer to the [vLLM PR](https://github.com/vllm-project/vllm/pull/31575/files) for implementation details.
|
| 123 |
|
| 124 |
```bash
|
| 125 |
+
vllm serve IQuestLab/IQuest-Coder-V1-40B-Instruct --tensor-parallel-size 8
|
| 126 |
```
|
| 127 |
|
| 128 |
For Thinking models with reasoning support:
|
| 129 |
|
| 130 |
```bash
|
| 131 |
+
vllm serve IQuestLab/IQuest-Coder-V1-40B-Thinking --reasoning-parser qwen3 --tensor-parallel-size 8
|
| 132 |
```
|
| 133 |
|
| 134 |
|