gpt-physics
A small GPT trained from scratch to predict 2D rigid body physics trajectories. Part of an ICML-2026 study on whether language models can learn physical dynamics from text-encoded simulation data.
Model details
- Architecture: 6-layer GPT, learned positional embeddings, tied LM head
- Tokenizer: digit-level
PhysicsTokenizer(custom) - Scaling: muP for hyperparameter transfer
- Training: curriculum learning over 5 difficulty stages
- Task: autoregressive next-frame prediction over 200-frame rigid-body scenes
- Domain: 2D rigid body dynamics simulated with Pymunk / Chipmunk2D
Files
best_model.ptโ best validation checkpoint (~69 MB)checkpoint_latest.ptโ latest training step (~158 MB)checkpoint_epoch0_step500.ptโ early checkpoint (~158 MB)
State dicts contain raw transformer.* and lm_head.* keys for a stock 6-layer GPT โ load with the project's src/scratch/gpt.py model class.
Training data
Trained on ~900K scenes across 24 "seen" scenario types (collisions, stacking, ramps, constraints, mini-games, complex). See physics-scenarios-packed and physics-scenarios-raw.
Intended use
Research on whether autoregressive LMs can internalize physical dynamics. Not intended for production physics simulation โ use Pymunk for that.
Citation
ICML-2026 submission (in progress).
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support