Safetensors
English
gpt2

GPT-51M_step_6668

A 51M-parameter GPT-style model trained from scratch by Tralalabs.

Training Data

  • TinyStories (50%)
  • OpenWebText (30%)
  • Wikipedia 20231101.en (20%)

Details

  • Context: 768 tokens
  • Tokenizer: GPT-2 BPE
  • Training steps: 6668
  • ~60M tokens seen

Use

Best for storytelling, simple explanations, and structured text generation.

Downloads last month
76
Safetensors
Model size
51.3M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Datasets used to train Tralalabs/GPT-51M_step_6668