Kite

🎉 You are looking at Kite 2.5, which is now trained using pika 2!

Kite is a small, trained, 13 million parameter language model, without any special optimizations.

Training

It was trained on 50K rows of this dataset using 12500 steps, 1 epoch, 4 batch size, 5e-4 learning rate, and the pika 2 tokenizer.

Limitations

Due to its size, the model is not suitable for production workloads.

Loss

Downloads last month
39
Safetensors
Model size
14.6M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for qikp/kite-2.5-13m

Finetunes
1 model

Dataset used to train qikp/kite-2.5-13m