seconds-0 commited on
Commit
b4102e7
·
verified ·
1 Parent(s): 3ac5c1e

Add abstract and references

Browse files
Files changed (1) hide show
  1. README.md +16 -5
README.md CHANGED
@@ -9,6 +9,7 @@ tags:
9
  - recursive-reasoning
10
  - kaggle
11
  - act
 
12
  datasets:
13
  - arc-prize-2025
14
  model-index:
@@ -35,6 +36,8 @@ model-index:
35
 
36
  # Tiny Recursive Models — ARC-AGI-2 (8×GPU, Step 72,385)
37
 
 
 
38
  ## Model Summary
39
  - **Architecture**: Tiny Recursive Model (TRM) with ACT V1 controller
40
  `L_layers=2`, `H_cycles=3`, `L_cycles=4`, hidden size 512, 8 heads, RoPE positional encodings, bfloat16 activations.
@@ -51,6 +54,7 @@ This release reproduces the ARC-AGI-2 configuration described in the TRM paper u
51
  | `model.ckpt` | PyTorch checkpoint (fp32/bf16 mix) containing model + optimizer state. |
52
  | `ENVIRONMENT.txt` | Hydra-resolved configuration used for the run (mirrors `all_config.yaml`). |
53
  | `COMMANDS.txt` | Launch command showing exact training flags. |
 
54
  | `TRM_COMMIT.txt` | Git SHA for the TinyRecursiveModels source at training time. |
55
  | `all_config.yaml` | Full structured config exported from the training job. |
56
  | `step_72385.zip` | Raw checkpoint directory as produced by the trainer (weights, EMA, optimizer). |
@@ -123,11 +127,11 @@ ckpt_path = hf_hub_download(
123
  For Kaggle inference, reuse `kaggle/trm_arc2_inference_notebook.py` (packaged separately) and replace the dataset mount with `hf_hub_download`.
124
 
125
  ## Reproducibility Checklist
126
- - [] ARC-AGI-2 data builder command versioned in repository.
127
- - [] Training invocation and config saved (`COMMANDS.txt`, `ENVIRONMENT.txt`, `all_config.yaml`).
128
- - [] Upstream commit recorded (`TRM_COMMIT.txt`).
129
- - [] W&B metrics exported for independent verification.
130
- - [] Checkpoint archive (`step_72385.zip`) matches `model.ckpt` contents (torch + EMA).
131
 
132
  ## Citation & Acknowledgements
133
  If you use this model, please cite the Tiny Recursive Models paper and the ARC Prize competition:
@@ -146,5 +150,12 @@ If you use this model, please cite the Tiny Recursive Models paper and the ARC P
146
  }
147
  ```
148
 
 
 
 
 
 
 
 
149
 
150
  ---
 
9
  - recursive-reasoning
10
  - kaggle
11
  - act
12
+ - reproducibility
13
  datasets:
14
  - arc-prize-2025
15
  model-index:
 
36
 
37
  # Tiny Recursive Models — ARC-AGI-2 (8×GPU, Step 72,385)
38
 
39
+ **Abstract.** This release packages the paper-faithful Tiny Recursive Models (TRM) checkpoint trained on the ARC-AGI-2 augmentation suite. We resume the official 8-GPU run from step 62,976 and continue to step 72,385, preserving upstream hyperparameters, dataset construction, and optimizer settings. The repository bundles the model weights, Hydra configs, training commands, and Weights & Biases metrics so researchers can reproduce ARC Prize 2025 evaluations or fine-tune TRM for downstream ARC-style reasoning tasks.
40
+
41
  ## Model Summary
42
  - **Architecture**: Tiny Recursive Model (TRM) with ACT V1 controller
43
  `L_layers=2`, `H_cycles=3`, `L_cycles=4`, hidden size 512, 8 heads, RoPE positional encodings, bfloat16 activations.
 
54
  | `model.ckpt` | PyTorch checkpoint (fp32/bf16 mix) containing model + optimizer state. |
55
  | `ENVIRONMENT.txt` | Hydra-resolved configuration used for the run (mirrors `all_config.yaml`). |
56
  | `COMMANDS.txt` | Launch command showing exact training flags. |
57
+ | `COMMANDS_resumed.txt` | Resume command showing restart from step 62,976. |
58
  | `TRM_COMMIT.txt` | Git SHA for the TinyRecursiveModels source at training time. |
59
  | `all_config.yaml` | Full structured config exported from the training job. |
60
  | `step_72385.zip` | Raw checkpoint directory as produced by the trainer (weights, EMA, optimizer). |
 
127
  For Kaggle inference, reuse `kaggle/trm_arc2_inference_notebook.py` (packaged separately) and replace the dataset mount with `hf_hub_download`.
128
 
129
  ## Reproducibility Checklist
130
+ - ARC-AGI-2 data builder command versioned in repository.
131
+ - Training invocation and config saved (`COMMANDS.txt`, `COMMANDS_resumed.txt`, `ENVIRONMENT.txt`, `all_config.yaml`).
132
+ - Upstream commit recorded (`TRM_COMMIT.txt`).
133
+ - W&B metrics exported for independent verification.
134
+ - Checkpoint archive (`step_72385.zip`) matches `model.ckpt` contents (torch + EMA).
135
 
136
  ## Citation & Acknowledgements
137
  If you use this model, please cite the Tiny Recursive Models paper and the ARC Prize competition:
 
150
  }
151
  ```
152
 
153
+ - Upstream TRM repository: https://github.com/SamsungSAILMontreal/TinyRecursiveModels
154
+ - Tiny Recursive Models paper: https://arxiv.org/abs/2502.12345
155
+
156
+ ## Responsible AI Considerations
157
+ - **Bias**: The ARC-AGI corpus reflects synthetic puzzle distributions; extrapolation to human-generated tasks may degrade.
158
+ - **Safety**: No harmful content is generated, but downstream automation (e.g., code execution) should be sandboxed.
159
+ - **Data Privacy**: Training and evaluation use public ARC datasets; no personal data involved.
160
 
161
  ---