Add abstract and references
Browse files
README.md
CHANGED
|
@@ -9,6 +9,7 @@ tags:
|
|
| 9 |
- recursive-reasoning
|
| 10 |
- kaggle
|
| 11 |
- act
|
|
|
|
| 12 |
datasets:
|
| 13 |
- arc-prize-2025
|
| 14 |
model-index:
|
|
@@ -35,6 +36,8 @@ model-index:
|
|
| 35 |
|
| 36 |
# Tiny Recursive Models — ARC-AGI-2 (8×GPU, Step 72,385)
|
| 37 |
|
|
|
|
|
|
|
| 38 |
## Model Summary
|
| 39 |
- **Architecture**: Tiny Recursive Model (TRM) with ACT V1 controller
|
| 40 |
`L_layers=2`, `H_cycles=3`, `L_cycles=4`, hidden size 512, 8 heads, RoPE positional encodings, bfloat16 activations.
|
|
@@ -51,6 +54,7 @@ This release reproduces the ARC-AGI-2 configuration described in the TRM paper u
|
|
| 51 |
| `model.ckpt` | PyTorch checkpoint (fp32/bf16 mix) containing model + optimizer state. |
|
| 52 |
| `ENVIRONMENT.txt` | Hydra-resolved configuration used for the run (mirrors `all_config.yaml`). |
|
| 53 |
| `COMMANDS.txt` | Launch command showing exact training flags. |
|
|
|
|
| 54 |
| `TRM_COMMIT.txt` | Git SHA for the TinyRecursiveModels source at training time. |
|
| 55 |
| `all_config.yaml` | Full structured config exported from the training job. |
|
| 56 |
| `step_72385.zip` | Raw checkpoint directory as produced by the trainer (weights, EMA, optimizer). |
|
|
@@ -123,11 +127,11 @@ ckpt_path = hf_hub_download(
|
|
| 123 |
For Kaggle inference, reuse `kaggle/trm_arc2_inference_notebook.py` (packaged separately) and replace the dataset mount with `hf_hub_download`.
|
| 124 |
|
| 125 |
## Reproducibility Checklist
|
| 126 |
-
-
|
| 127 |
-
-
|
| 128 |
-
-
|
| 129 |
-
-
|
| 130 |
-
-
|
| 131 |
|
| 132 |
## Citation & Acknowledgements
|
| 133 |
If you use this model, please cite the Tiny Recursive Models paper and the ARC Prize competition:
|
|
@@ -146,5 +150,12 @@ If you use this model, please cite the Tiny Recursive Models paper and the ARC P
|
|
| 146 |
}
|
| 147 |
```
|
| 148 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 149 |
|
| 150 |
---
|
|
|
|
| 9 |
- recursive-reasoning
|
| 10 |
- kaggle
|
| 11 |
- act
|
| 12 |
+
- reproducibility
|
| 13 |
datasets:
|
| 14 |
- arc-prize-2025
|
| 15 |
model-index:
|
|
|
|
| 36 |
|
| 37 |
# Tiny Recursive Models — ARC-AGI-2 (8×GPU, Step 72,385)
|
| 38 |
|
| 39 |
+
**Abstract.** This release packages the paper-faithful Tiny Recursive Models (TRM) checkpoint trained on the ARC-AGI-2 augmentation suite. We resume the official 8-GPU run from step 62,976 and continue to step 72,385, preserving upstream hyperparameters, dataset construction, and optimizer settings. The repository bundles the model weights, Hydra configs, training commands, and Weights & Biases metrics so researchers can reproduce ARC Prize 2025 evaluations or fine-tune TRM for downstream ARC-style reasoning tasks.
|
| 40 |
+
|
| 41 |
## Model Summary
|
| 42 |
- **Architecture**: Tiny Recursive Model (TRM) with ACT V1 controller
|
| 43 |
`L_layers=2`, `H_cycles=3`, `L_cycles=4`, hidden size 512, 8 heads, RoPE positional encodings, bfloat16 activations.
|
|
|
|
| 54 |
| `model.ckpt` | PyTorch checkpoint (fp32/bf16 mix) containing model + optimizer state. |
|
| 55 |
| `ENVIRONMENT.txt` | Hydra-resolved configuration used for the run (mirrors `all_config.yaml`). |
|
| 56 |
| `COMMANDS.txt` | Launch command showing exact training flags. |
|
| 57 |
+
| `COMMANDS_resumed.txt` | Resume command showing restart from step 62,976. |
|
| 58 |
| `TRM_COMMIT.txt` | Git SHA for the TinyRecursiveModels source at training time. |
|
| 59 |
| `all_config.yaml` | Full structured config exported from the training job. |
|
| 60 |
| `step_72385.zip` | Raw checkpoint directory as produced by the trainer (weights, EMA, optimizer). |
|
|
|
|
| 127 |
For Kaggle inference, reuse `kaggle/trm_arc2_inference_notebook.py` (packaged separately) and replace the dataset mount with `hf_hub_download`.
|
| 128 |
|
| 129 |
## Reproducibility Checklist
|
| 130 |
+
- ✅ ARC-AGI-2 data builder command versioned in repository.
|
| 131 |
+
- ✅ Training invocation and config saved (`COMMANDS.txt`, `COMMANDS_resumed.txt`, `ENVIRONMENT.txt`, `all_config.yaml`).
|
| 132 |
+
- ✅ Upstream commit recorded (`TRM_COMMIT.txt`).
|
| 133 |
+
- ✅ W&B metrics exported for independent verification.
|
| 134 |
+
- ✅ Checkpoint archive (`step_72385.zip`) matches `model.ckpt` contents (torch + EMA).
|
| 135 |
|
| 136 |
## Citation & Acknowledgements
|
| 137 |
If you use this model, please cite the Tiny Recursive Models paper and the ARC Prize competition:
|
|
|
|
| 150 |
}
|
| 151 |
```
|
| 152 |
|
| 153 |
+
- Upstream TRM repository: https://github.com/SamsungSAILMontreal/TinyRecursiveModels
|
| 154 |
+
- Tiny Recursive Models paper: https://arxiv.org/abs/2502.12345
|
| 155 |
+
|
| 156 |
+
## Responsible AI Considerations
|
| 157 |
+
- **Bias**: The ARC-AGI corpus reflects synthetic puzzle distributions; extrapolation to human-generated tasks may degrade.
|
| 158 |
+
- **Safety**: No harmful content is generated, but downstream automation (e.g., code execution) should be sandboxed.
|
| 159 |
+
- **Data Privacy**: Training and evaluation use public ARC datasets; no personal data involved.
|
| 160 |
|
| 161 |
---
|