TagRouter: Learning Route to LLMs through Tags for Open-Domain Text Generation Tasks
Paper • 2506.12473 • Published • 1
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("itpossible/ClimateChat")
model = AutoModelForCausalLM.from_pretrained("itpossible/ClimateChat")
messages = [
{"role": "user", "content": "Who are you?"},
]
inputs = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
tokenize=True,
return_dict=True,
return_tensors="pt",
).to(model.device)
outputs = model.generate(**inputs, max_new_tokens=40)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))| Model Series | Model | Download Link | Description |
|---|---|---|---|
| JiuZhou | JiuZhou-base | Huggingface | Base model (Rich in geoscience knowledge) |
| JiuZhou | JiuZhou-Instruct-v0.1 | Huggingface | Instruct model (Instruction alignment caused a loss of some geoscience knowledge, but it has instruction-following ability) LoRA fine-tuned on Alpaca_GPT4 in both Chinese and English and GeoSignal |
| JiuZhou | JiuZhou-Instruct-v0.2 | HuggingFace Wisemodel |
Instruct model (Instruction alignment caused a loss of some geoscience knowledge, but it has instruction-following ability) Fine-tuned with high-quality general instruction data |
| ClimateChat | ClimateChat | HuggingFace Wisemodel |
Instruct model Fine-tuned on JiuZhou-base for instruction following |
| Chinese-Mistral | Chinese-Mistral-7B | HuggingFace Wisemodel ModelScope |
Base model |
| Chinese-Mistral | Chinese-Mistral-7B-Instruct-v0.1 | HuggingFace Wisemodel ModelScope |
Instruct model LoRA fine-tuned with Alpaca_GPT4 in both Chinese and English |
| Chinese-Mistral | Chinese-Mistral-7B-Instruct-v0.2 | HuggingFace Wisemodel |
Instruct model LoRA fine-tuned with a million high-quality instructions |
| PreparedLLM | Prepared-Llama | Huggingface Wisemodel |
Base model Continual pretraining with a small number of geoscience data Recommended to use JiuZhou |
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="itpossible/ClimateChat") messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)