TagRouter: Learning Route to LLMs through Tags for Open-Domain Text Generation Tasks
Paper • 2506.12473 • Published • 1
docker run --gpus all \
--shm-size 32g \
-p 30000:30000 \
-v ~/.cache/huggingface:/root/.cache/huggingface \
--env "HF_TOKEN=<secret>" \
--ipc=host \
lmsysorg/sglang:latest \
python3 -m sglang.launch_server \
--model-path "itpossible/ClimateChat" \
--host 0.0.0.0 \
--port 30000# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/chat/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "itpossible/ClimateChat",
"messages": [
{
"role": "user",
"content": "What is the capital of France?"
}
]
}'| Model Series | Model | Download Link | Description |
|---|---|---|---|
| JiuZhou | JiuZhou-base | Huggingface | Base model (Rich in geoscience knowledge) |
| JiuZhou | JiuZhou-Instruct-v0.1 | Huggingface | Instruct model (Instruction alignment caused a loss of some geoscience knowledge, but it has instruction-following ability) LoRA fine-tuned on Alpaca_GPT4 in both Chinese and English and GeoSignal |
| JiuZhou | JiuZhou-Instruct-v0.2 | HuggingFace Wisemodel |
Instruct model (Instruction alignment caused a loss of some geoscience knowledge, but it has instruction-following ability) Fine-tuned with high-quality general instruction data |
| ClimateChat | ClimateChat | HuggingFace Wisemodel |
Instruct model Fine-tuned on JiuZhou-base for instruction following |
| Chinese-Mistral | Chinese-Mistral-7B | HuggingFace Wisemodel ModelScope |
Base model |
| Chinese-Mistral | Chinese-Mistral-7B-Instruct-v0.1 | HuggingFace Wisemodel ModelScope |
Instruct model LoRA fine-tuned with Alpaca_GPT4 in both Chinese and English |
| Chinese-Mistral | Chinese-Mistral-7B-Instruct-v0.2 | HuggingFace Wisemodel |
Instruct model LoRA fine-tuned with a million high-quality instructions |
| PreparedLLM | Prepared-Llama | Huggingface Wisemodel |
Base model Continual pretraining with a small number of geoscience data Recommended to use JiuZhou |
Install from pip and serve model
# Install SGLang from pip: pip install sglang# Start the SGLang server: python3 -m sglang.launch_server \ --model-path "itpossible/ClimateChat" \ --host 0.0.0.0 \ --port 30000# Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "itpossible/ClimateChat", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'