A newer version of this model is available: qikp/kite-2.6-13m

Kite

Kite is a small, trained, 1 million parameter language model, without any special optimizations.

Training

It was trained on this dataset using 3500 steps, 1 epoch, 1 batch size, and the GPT-2 tokenizer.

Limitations

Due to its size, the model will produce near-nonsensical output and is not suitable for production workloads.

Downloads last month
2
Safetensors
Model size
1.65M params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train qikp/kite-1m