Update README.md
Browse files
README.md
CHANGED
|
@@ -139,26 +139,26 @@ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
|
|
| 139 |
|
| 140 |
**Deployment**
|
| 141 |
|
| 142 |
-
For deployment, you can use `sglang>=0.5.2` or `vllm>=0.
|
| 143 |
- SGLang:
|
| 144 |
```shell
|
| 145 |
-
python -m sglang.launch_server --model-path tablegpt/TableGPT-R1 --served-model-name TableGPT-R1 --reasoning-parser qwen3
|
| 146 |
```
|
| 147 |
- vLLM:
|
| 148 |
```shell
|
| 149 |
-
vllm serve tablegpt/TableGPT-R1 --served-model-name TableGPT-R1 --
|
| 150 |
```
|
| 151 |
|
| 152 |
Then you can access the Chat API by:
|
| 153 |
|
| 154 |
```bash
|
| 155 |
-
curl http://localhost:
|
| 156 |
-H "Content-Type: application/json" \
|
| 157 |
-d '{
|
| 158 |
"model": "TableGPT-R1",
|
| 159 |
"messages": [
|
| 160 |
{"role": "system", "content": "You are a helpful assistant."},
|
| 161 |
-
{"role": "user", "content": "
|
| 162 |
]
|
| 163 |
}'
|
| 164 |
|
|
|
|
| 139 |
|
| 140 |
**Deployment**
|
| 141 |
|
| 142 |
+
For deployment, you can use `sglang>=0.5.2` or `vllm>=0.10.2` or to create an OpenAI-compatible API endpoint:
|
| 143 |
- SGLang:
|
| 144 |
```shell
|
| 145 |
+
python -m sglang.launch_server --model-path tablegpt/TableGPT-R1 --port 8080 --served-model-name TableGPT-R1 --reasoning-parser qwen3
|
| 146 |
```
|
| 147 |
- vLLM:
|
| 148 |
```shell
|
| 149 |
+
vllm serve tablegpt/TableGPT-R1 --port 8080 --served-model-name TableGPT-R1 --reasoning-parser deepseek_r1
|
| 150 |
```
|
| 151 |
|
| 152 |
Then you can access the Chat API by:
|
| 153 |
|
| 154 |
```bash
|
| 155 |
+
curl http://localhost:8080/v1/chat/completions \
|
| 156 |
-H "Content-Type: application/json" \
|
| 157 |
-d '{
|
| 158 |
"model": "TableGPT-R1",
|
| 159 |
"messages": [
|
| 160 |
{"role": "system", "content": "You are a helpful assistant."},
|
| 161 |
+
{"role": "user", "content": "Give me a short introduction to large language model."}
|
| 162 |
]
|
| 163 |
}'
|
| 164 |
|