mindware commited on
Commit
02488e5
·
verified ·
1 Parent(s): 0638161

Refresh ARC datasets section

Browse files
Files changed (1) hide show
  1. README.md +8 -10
README.md CHANGED
@@ -37,17 +37,15 @@ This checkpoint is the primary CodeT5-based solver we used for the MindsAI @ Tuf
37
  - **Decoder-only pruning**: the original decoder depth (24) was reduced to 16 layers after experiments showed encoder pruning harmed sample efficiency, while decoder pruning could be recovered through extended training.
38
  - **Long-run TPU training**: training spanned roughly two years on a V4-64 TPU, made possible by Google’s TPU Research Cloud program.
39
 
40
- 📚 ARC-Related Datasets & Frameworks
41
- RE-ARC Link: https://github.com/michaelhodel/re-arc
42
- Note: This is the repository from Michael Hodel, which procedurally generates examples for the 400 ARC training tasks. We also include RE-ARC eval and ARC 1.5 (also by Michael Hodel).
43
- ConceptARC Link: https://github.com/victorvikram/ConceptARC
44
- 1D-ARC (likely "ID ARC") Link: https://khalil-research.github.io/LLM4ARC/
45
- ARC_gym
46
- Sort-of-ARC
47
- Andreas Koepf - Generated many tasks based upon the RE-ARC methodology using various foundation models. Additionally generated from a generator Andreas wrote based on the icecuber solution. It also includes extra tasks like predicting the solution graph.
48
- Jack Cole - Wrote generators for 60-80 tasks. Many were inspired by ARC items. Others were large concept datasets (cellular automata, math equation derived boards).
49
 
50
- There is a large amount of ARC-related tasks that are not solving for the board (like generating code, predicting various parameters or features related to the task). There are other non-ARC related tasks.
51
 
52
  ## ARC Data Formatting
53
 
 
37
  - **Decoder-only pruning**: the original decoder depth (24) was reduced to 16 layers after experiments showed encoder pruning harmed sample efficiency, while decoder pruning could be recovered through extended training.
38
  - **Long-run TPU training**: training spanned roughly two years on a V4-64 TPU, made possible by Google’s TPU Research Cloud program.
39
 
40
+ 📚 **ARC-Related Datasets & Frameworks**
41
+ - [RE-ARC](https://github.com/michaelhodel/re-arc) — procedurally generates examples for the 400 ARC training tasks (we also include RE-ARC eval + ARC 1.5).
42
+ - [ConceptARC](https://github.com/victorvikram/ConceptARC)
43
+ - [1D-ARC](https://khalil-research.github.io/LLM4ARC/)
44
+ - ARC_gym, Sort-of-ARC
45
+ - Andreas Koepf’s generator suites (includes RE-ARC-style grids, code generation targets, and solution graphs).
46
+ - Jack Cole’s custom generators covering ~70 tasks plus larger concept sets (cellular automata, math-derived boards, etc.).
 
 
47
 
48
+ Several auxiliary datasets predict task metadata (graphs, heuristics, explanations) rather than final boards; they are part of the broader instruction mixture this model saw during pretraining.
49
 
50
  ## ARC Data Formatting
51