TinyStories: How Small Can Language Models Be and Still Speak Coherent English?
Paper • 2305.07759 • Published • 45
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("igorktech/TinyIG")
model = AutoModelForCausalLM.from_pretrained("igorktech/TinyIG")This model is a pre-trained version of LLama2 on the nampdn-ai/tiny-textbooks dataset.
Inspired by TinyStories.
It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
| Training Loss | Epoch | Step | Validation Loss |
|---|---|---|---|
| 3.4647 | 2.02 | 30000 | 3.5677 |
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="igorktech/TinyIG")