TinyStories: How Small Can Language Models Be and Still Speak Coherent English?
Paper
•
2305.07759
•
Published
•
38
A small GPT-2 style language model trained on TinyStories dataset for generating children's stories.
from transformers import pipeline
# Load the model
generator = pipeline("text-generation", model="endurasolution/ronmicro-llm-story")
# Generate a story
story = generator(
"Once upon a time",
max_new_tokens=150,
temperature=0.7,
repetition_penalty=1.3,
no_repeat_ngram_size=3,
do_sample=True
)
print(story[0]["generated_text"])
Prompt: "Once upon a time" Output: "Once upon a time, there was a little boy named Timmy. He loved to play with his toy cars and trucks all day long..."
Phase 2 training in progress with 20% data and 5 epochs for improved quality.
Built using TinyStories dataset:
@article{eldan2023tinystories,
title={TinyStories: How Small Can Language Models Be and Still Speak Coherent English?},
author={Eldan, Ronen and Li, Yuanzhi},
journal={arXiv preprint arXiv:2305.07759},
year={2023}
}
MIT License