Update README.md
Browse files
README.md
CHANGED
|
@@ -120,13 +120,13 @@ model_name = "IQuest/IQuest-Coder-V1-40B-Thinking"
|
|
| 120 |
For production deployment, you can use vLLM to create an OpenAI-compatible API endpoint. Please refer to the [vLLM PR](https://github.com/vllm-project/vllm/pull/31575/files) for implementation details.
|
| 121 |
|
| 122 |
```bash
|
| 123 |
-
vllm serve
|
| 124 |
```
|
| 125 |
|
| 126 |
For Thinking models with reasoning support:
|
| 127 |
|
| 128 |
```bash
|
| 129 |
-
vllm serve
|
| 130 |
```
|
| 131 |
|
| 132 |
|
|
|
|
| 120 |
For production deployment, you can use vLLM to create an OpenAI-compatible API endpoint. Please refer to the [vLLM PR](https://github.com/vllm-project/vllm/pull/31575/files) for implementation details.
|
| 121 |
|
| 122 |
```bash
|
| 123 |
+
vllm serve IQuestLab/IQuest-Coder-V1-40B-Instruct --tensor-parallel-size 8
|
| 124 |
```
|
| 125 |
|
| 126 |
For Thinking models with reasoning support:
|
| 127 |
|
| 128 |
```bash
|
| 129 |
+
vllm serve IQuestLab/IQuest-Coder-V1-40B-Thinking --reasoning-parser qwen3 --tensor-parallel-size 8
|
| 130 |
```
|
| 131 |
|
| 132 |
|