Meet25M ChatML

ChatML fine-tuned version of Meet25M Base.

Model

  • Base: Meet25M Base
  • Approx size: ~25M parameters
  • Context length: 1024
  • Tokenizer: custom byte-level BPE with ChatML tokens
  • Chat format: ChatML

ChatML Format

Use:

<|im_start|>system
You are a helpful assistant.<|im_end|>
<|im_start|>user
Your message here.<|im_end|>
<|im_start|>assistant

The model should generate the assistant response until <|im_end|>.

SFT Dataset

Fine-tuned on:

  • OpenLeecher/lmsys_chat_1m_clean

Files

  • model.safetensors โ€” safetensors checkpoint
  • config.json โ€” model config
  • tokenizer/ โ€” tokenizer files
  • safetensors_info.json โ€” checkpoint metadata

Loading

This is not a standard Transformers AutoModelForCausalLM checkpoint.
Use the custom GPT class from the training/SFT script and load model.safetensors.

Downloads last month
15
Safetensors
Model size
26.7M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support