How to use from the
Use from the
Transformers library
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-generation", model="unsloth/tinyllama")
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("unsloth/tinyllama")
model = AutoModelForCausalLM.from_pretrained("unsloth/tinyllama")
Quick Links

Finetune Mistral, Gemma, Llama 2-5x faster with 70% less memory via Unsloth!

A reupload from https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T

We have a Google Colab Tesla T4 notebook for TinyLlama with 4096 max sequence length RoPE Scaling here: https://colab.research.google.com/drive/1AZghoNBQaMDgWJpi4RbffGM1h6raLUj9?usp=sharing

โœจ Finetune for Free

All notebooks are beginner friendly! Add your dataset, click "Run All", and you'll get a 2x faster finetuned model which can be exported to GGUF, vLLM or uploaded to Hugging Face.

Unsloth supports Free Notebooks Performance Memory use
Gemma 7b โ–ถ๏ธ Start on Colab 2.4x faster 58% less
Mistral 7b โ–ถ๏ธ Start on Colab 2.2x faster 62% less
Llama-2 7b โ–ถ๏ธ Start on Colab 2.2x faster 43% less
TinyLlama โ–ถ๏ธ Start on Colab 3.9x faster 74% less
CodeLlama 34b A100 โ–ถ๏ธ Start on Colab 1.9x faster 27% less
Mistral 7b 1xT4 โ–ถ๏ธ Start on Kaggle 5x faster* 62% less
DPO - Zephyr โ–ถ๏ธ Start on Colab 1.9x faster 19% less
Downloads last month
768
Safetensors
Model size
1B params
Tensor type
BF16
ยท
Inference Providers NEW

Model tree for unsloth/tinyllama

Adapters
298 models
Finetunes
23 models
Quantizations
17 models

Spaces using unsloth/tinyllama 2