Datasets:

Modalities:
Text
Formats:
parquet
Size:
< 1K
ArXiv:
License:
Sinjhin commited on
Commit
0559e0a
·
verified ·
1 Parent(s): de87c48

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +205 -196
README.md CHANGED
@@ -1,196 +1,205 @@
1
- ---
2
- license: apache-2.0
3
- ---
4
-
5
- # ARC-AGI-V1 Dataset (A Take On Format)
6
-
7
- This dataset is a reorganized version of the [ARC-AGI v1](https://github.com/fchollet/ARC-AGI) (Abstraction and Reasoning Corpus) benchmark, formatted for HuggingFace Datasets.
8
-
9
- ## Dataset Structure
10
-
11
- The original ARC-AGI dataset has been transformed from its file-based JSON structure into a standardized HuggingFace dataset with two splits:
12
-
13
- - **train** (400 examples): Tasks from the original `training` directory
14
- - **test** (400 examples): Tasks from the original `evaluation` directory
15
-
16
- ### Original Structure
17
-
18
- The original ARC-AGI dataset consisted of:
19
- - A `training` directory with JSON files (one per task)
20
- - An `evaluation` directory with JSON files (one per task)
21
- - Each JSON file named with a task ID (e.g., `007bbfb7.json`)
22
- - Each file containing:
23
- - `train`: Array of input/output example pairs for learning the pattern
24
- - `test`: Array of input/output pairs representing the actual task to solve
25
-
26
- ### Transformed Structure
27
-
28
- Each row in this dataset represents a single ARC-AGI task with the following schema:
29
-
30
- ```
31
- {
32
- "id": string, // Task ID from the original filename
33
- "list": [ // Combined training examples and test inputs
34
- [ // Training example inputs (from original 'train')
35
- [[int]], [[int]], ...
36
- ],
37
- [ // Training example outputs (from original 'train')
38
- [[int]], [[int]], ...
39
- ],
40
- [ // Test inputs (from original 'test')
41
- [[int]], [[int]], ...
42
- ]
43
- ],
44
- "label": [ // Test outputs (from original 'test')
45
- [[int]], [[int]], ...
46
- ]
47
- }
48
- ```
49
-
50
- #### Field Descriptions
51
-
52
- - **`id`**: The unique task identifier from the original filename
53
- - **`list`**: A nested list containing three components in order:
54
- 1. **Example inputs** (`list[0]`): All input grids from the original `train` array
55
- 2. **Example outputs** (`list[1]`): All output grids from the original `train` array (paired with example inputs)
56
- 3. **Test inputs** (`list[2]`): All input grids from the original `test` array
57
- - **`label`**: The correct output grids for the test inputs (from original `test` array outputs)
58
-
59
- ### Data Format
60
-
61
- Each grid is represented as a 2D array of integers (0-9), where:
62
- - Values range from 0 to 9 (representing different colors/states)
63
- - Grid dimensions vary from 1×1 to 30×30
64
- - Each integer represents a colored cell in the grid
65
-
66
- ### Example
67
-
68
- ```json
69
- {
70
- "id": "007bbfb7",
71
- "list": [
72
- [
73
- [[0, 7, 7], // Example input 1
74
- [7, 7, 7], //
75
- [0, 7, 7]], //
76
- [[4, 0, 4], [0, 0, 0], [0, 4, 0]], // Example input 2
77
- [[0, 0, 0], [0, 0, 2], [2, 0, 2]] // Example input 3
78
- ],
79
- [
80
- [[0, 0, 0, 0, 7, 7, 0, 7, 7], // Example output 1
81
- [0, 0, 0, 7, 7, 7, 7, 7, 7],
82
- [0, 0, 0, 0, 7, 7, 0, 7, 7],
83
- [0, 7, 7, 0, 7, 7, 0, 7, 7],
84
- [7, 7, 7, 7, 7, 7, 7, 7, 7],
85
- [0, 7, 7, 0, 7, 7, 0, 7, 7],
86
- [0, 0, 0, 0, 7, 7, 0, 7, 7],
87
- [0, 0, 0, 7, 7, 7, 7, 7, 7],
88
- [0, 0, 0, 0, 7, 7, 0, 7, 7]],
89
- [[], [], [], [], [], [], [], [], []], // etc..
90
- ],
91
- [
92
- [[7, 0, 7], [7, 0, 7], [7, 7, 0]] // Test input 1
93
- ]
94
- ],
95
- "label": [
96
- [[7, 0, 7, 0, 0, 0, 7, 0, 7], // Test output 1 (ground truth)
97
- [7, 0, 7, 0, 0, 0, 7, 0, 7],
98
- [7, 7, 0, 0, 0, 0, 7, 7, 0],
99
- [7, 0, 7, 0, 0, 0, 7, 0, 7],
100
- [7, 0, 7, 0, 0, 0, 7, 0, 7],
101
- [7, 7, 0, 0, 0, 0, 7, 7, 0],
102
- [7, 0, 7, 7, 0, 7, 0, 0, 0],
103
- [7, 0, 7, 7, 0, 7, 0, 0, 0],
104
- [7, 7, 0, 7, 7, 0, 0, 0, 0]]
105
- ]
106
- }
107
- ```
108
-
109
- ## Usage Philosophy
110
-
111
- pprint(dataset['train']['list'][0][0][0])
112
- pprint(dataset['train']['list'][0][1][0])
113
- print('')
114
- pprint(dataset['train']['list'][0][2][0])
115
- pprint(dataset['train']['label'][0][0])
116
-
117
- This ARC-AGI dataset format allows (me at least) to think about the tasks in this way:
118
- 1. **Learn from examples**: Study the input/output pairs:
119
- - input: `dataset['train']['list'][0][0][0]`
120
- - output: `dataset['train']['list'][0][1][0]`
121
- - input: `dataset['train']['list'][0][0][1]`
122
- - output: `dataset['train']['list'][0][1][1]`
123
- - where:
124
- - 1st num: `task number`
125
- - 2nd num: `either 0: example input || 1: example output`
126
- - 3rd num: `which example?`
127
- 2. **Then 'Get the tests'**:
128
- - `dataset['train']['list'][0][2][0]`
129
- 3. **Apply the pattern**: Use the learned rule to make your two guesses
130
- 4. **Evaluate performance**: Compare model predictions against the `label` field
131
- - `dataset['train']['label'][0][0]`
132
-
133
- ### Training Split
134
- - Contains all tasks from the original `training` directory
135
- - Intended for model training and development
136
- - Both example pairs and test solutions are provided
137
-
138
- ### Test Split
139
- - Contains all tasks from the original `evaluation` directory
140
- - Intended for final model evaluation
141
- - In competition settings, test labels may be withheld
142
-
143
- ## Dataset Features
144
-
145
- ```python
146
- Features({
147
- 'id': Value('string'),
148
- 'list': List(List(List(List(Value('int64'))))),
149
- 'label': List(List(List(Value('int64'))))
150
- })
151
- ```
152
-
153
- ## Loading the Dataset
154
-
155
- ```python
156
- from datasets import load_dataset
157
-
158
- dataset = load_dataset("ardea/arc_agi_v1")
159
-
160
- # Access splits
161
- train_data = dataset['train']
162
- test_data = dataset['test']
163
-
164
- # Example: Get a single task
165
- task = train_data[0]
166
- task_id = task['id']
167
- example_inputs = task['list'][0]
168
- example_outputs = task['list'][1]
169
- test_inputs = task['list'][2]
170
- test_outputs = task['label']
171
-
172
- # Example: Get a task by id
173
- task = list(filter(lambda t: t['id'] == '007bbfb7', train_data))
174
- ```
175
-
176
- ## Transparency
177
- I've left the script I used on the original dataset here as `arc_to_my_hf.py`
178
-
179
- ## Citation
180
-
181
- If you use this dataset, please cite the original ARC-AGI work:
182
-
183
- ```bibtex
184
- @misc{chollet2019measure,
185
- title={On the Measure of Intelligence},
186
- author={François Chollet},
187
- year={2019},
188
- eprint={1911.01547},
189
- archivePrefix={arXiv},
190
- primaryClass={cs.AI}
191
- }
192
- ```
193
-
194
- ## License
195
-
196
- This dataset maintains the Apache 2.0 license from the original ARC-AGI corpus.
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ task_categories:
4
+ - table-question-answering
5
+ tags:
6
+ - arc
7
+ - agi
8
+ - arc-agi
9
+ pretty_name: ARC AGI v1
10
+ size_categories:
11
+ - 1K<n<10K
12
+ ---
13
+
14
+ # ARC-AGI-V1 Dataset (A Take On Format)
15
+
16
+ This dataset is a reorganized version of the [ARC-AGI v1](https://github.com/fchollet/ARC-AGI) (Abstraction and Reasoning Corpus) benchmark, formatted for HuggingFace Datasets.
17
+
18
+ ## Dataset Structure
19
+
20
+ The original ARC-AGI dataset has been transformed from its file-based JSON structure into a standardized HuggingFace dataset with two splits:
21
+
22
+ - **train** (400 examples): Tasks from the original `training` directory
23
+ - **test** (400 examples): Tasks from the original `evaluation` directory
24
+
25
+ ### Original Structure
26
+
27
+ The original ARC-AGI dataset consisted of:
28
+ - A `training` directory with JSON files (one per task)
29
+ - An `evaluation` directory with JSON files (one per task)
30
+ - Each JSON file named with a task ID (e.g., `007bbfb7.json`)
31
+ - Each file containing:
32
+ - `train`: Array of input/output example pairs for learning the pattern
33
+ - `test`: Array of input/output pairs representing the actual task to solve
34
+
35
+ ### Transformed Structure
36
+
37
+ Each row in this dataset represents a single ARC-AGI task with the following schema:
38
+
39
+ ```
40
+ {
41
+ "id": string, // Task ID from the original filename
42
+ "list": [ // Combined training examples and test inputs
43
+ [ // Training example inputs (from original 'train')
44
+ [[int]], [[int]], ...
45
+ ],
46
+ [ // Training example outputs (from original 'train')
47
+ [[int]], [[int]], ...
48
+ ],
49
+ [ // Test inputs (from original 'test')
50
+ [[int]], [[int]], ...
51
+ ]
52
+ ],
53
+ "label": [ // Test outputs (from original 'test')
54
+ [[int]], [[int]], ...
55
+ ]
56
+ }
57
+ ```
58
+
59
+ #### Field Descriptions
60
+
61
+ - **`id`**: The unique task identifier from the original filename
62
+ - **`list`**: A nested list containing three components in order:
63
+ 1. **Example inputs** (`list[0]`): All input grids from the original `train` array
64
+ 2. **Example outputs** (`list[1]`): All output grids from the original `train` array (paired with example inputs)
65
+ 3. **Test inputs** (`list[2]`): All input grids from the original `test` array
66
+ - **`label`**: The correct output grids for the test inputs (from original `test` array outputs)
67
+
68
+ ### Data Format
69
+
70
+ Each grid is represented as a 2D array of integers (0-9), where:
71
+ - Values range from 0 to 9 (representing different colors/states)
72
+ - Grid dimensions vary from 1×1 to 30×30
73
+ - Each integer represents a colored cell in the grid
74
+
75
+ ### Example
76
+
77
+ ```json
78
+ {
79
+ "id": "007bbfb7",
80
+ "list": [
81
+ [
82
+ [[0, 7, 7], // Example input 1
83
+ [7, 7, 7], //
84
+ [0, 7, 7]], //
85
+ [[4, 0, 4], [0, 0, 0], [0, 4, 0]], // Example input 2
86
+ [[0, 0, 0], [0, 0, 2], [2, 0, 2]] // Example input 3
87
+ ],
88
+ [
89
+ [[0, 0, 0, 0, 7, 7, 0, 7, 7], // Example output 1
90
+ [0, 0, 0, 7, 7, 7, 7, 7, 7],
91
+ [0, 0, 0, 0, 7, 7, 0, 7, 7],
92
+ [0, 7, 7, 0, 7, 7, 0, 7, 7],
93
+ [7, 7, 7, 7, 7, 7, 7, 7, 7],
94
+ [0, 7, 7, 0, 7, 7, 0, 7, 7],
95
+ [0, 0, 0, 0, 7, 7, 0, 7, 7],
96
+ [0, 0, 0, 7, 7, 7, 7, 7, 7],
97
+ [0, 0, 0, 0, 7, 7, 0, 7, 7]],
98
+ [[], [], [], [], [], [], [], [], []], // etc..
99
+ ],
100
+ [
101
+ [[7, 0, 7], [7, 0, 7], [7, 7, 0]] // Test input 1
102
+ ]
103
+ ],
104
+ "label": [
105
+ [[7, 0, 7, 0, 0, 0, 7, 0, 7], // Test output 1 (ground truth)
106
+ [7, 0, 7, 0, 0, 0, 7, 0, 7],
107
+ [7, 7, 0, 0, 0, 0, 7, 7, 0],
108
+ [7, 0, 7, 0, 0, 0, 7, 0, 7],
109
+ [7, 0, 7, 0, 0, 0, 7, 0, 7],
110
+ [7, 7, 0, 0, 0, 0, 7, 7, 0],
111
+ [7, 0, 7, 7, 0, 7, 0, 0, 0],
112
+ [7, 0, 7, 7, 0, 7, 0, 0, 0],
113
+ [7, 7, 0, 7, 7, 0, 0, 0, 0]]
114
+ ]
115
+ }
116
+ ```
117
+
118
+ ## Usage Philosophy
119
+
120
+ pprint(dataset['train']['list'][0][0][0])
121
+ pprint(dataset['train']['list'][0][1][0])
122
+ print('')
123
+ pprint(dataset['train']['list'][0][2][0])
124
+ pprint(dataset['train']['label'][0][0])
125
+
126
+ This ARC-AGI dataset format allows (me at least) to think about the tasks in this way:
127
+ 1. **Learn from examples**: Study the input/output pairs:
128
+ - input: `dataset['train']['list'][0][0][0]`
129
+ - output: `dataset['train']['list'][0][1][0]`
130
+ - input: `dataset['train']['list'][0][0][1]`
131
+ - output: `dataset['train']['list'][0][1][1]`
132
+ - where:
133
+ - 1st num: `task number`
134
+ - 2nd num: `either 0: example input || 1: example output`
135
+ - 3rd num: `which example?`
136
+ 2. **Then 'Get the tests'**:
137
+ - `dataset['train']['list'][0][2][0]`
138
+ 3. **Apply the pattern**: Use the learned rule to make your two guesses
139
+ 4. **Evaluate performance**: Compare model predictions against the `label` field
140
+ - `dataset['train']['label'][0][0]`
141
+
142
+ ### Training Split
143
+ - Contains all tasks from the original `training` directory
144
+ - Intended for model training and development
145
+ - Both example pairs and test solutions are provided
146
+
147
+ ### Test Split
148
+ - Contains all tasks from the original `evaluation` directory
149
+ - Intended for final model evaluation
150
+ - In competition settings, test labels may be withheld
151
+
152
+ ## Dataset Features
153
+
154
+ ```python
155
+ Features({
156
+ 'id': Value('string'),
157
+ 'list': List(List(List(List(Value('int64'))))),
158
+ 'label': List(List(List(Value('int64'))))
159
+ })
160
+ ```
161
+
162
+ ## Loading the Dataset
163
+
164
+ ```python
165
+ from datasets import load_dataset
166
+
167
+ dataset = load_dataset("ardea/arc_agi_v1")
168
+
169
+ # Access splits
170
+ train_data = dataset['train']
171
+ test_data = dataset['test']
172
+
173
+ # Example: Get a single task
174
+ task = train_data[0]
175
+ task_id = task['id']
176
+ example_inputs = task['list'][0]
177
+ example_outputs = task['list'][1]
178
+ test_inputs = task['list'][2]
179
+ test_outputs = task['label']
180
+
181
+ # Example: Get a task by id
182
+ task = list(filter(lambda t: t['id'] == '007bbfb7', train_data))
183
+ ```
184
+
185
+ ## Transparency
186
+ I've left the script I used on the original dataset here as `arc_to_my_hf.py`
187
+
188
+ ## Citation
189
+
190
+ If you use this dataset, please cite the original ARC-AGI work:
191
+
192
+ ```bibtex
193
+ @misc{chollet2019measure,
194
+ title={On the Measure of Intelligence},
195
+ author={François Chollet},
196
+ year={2019},
197
+ eprint={1911.01547},
198
+ archivePrefix={arXiv},
199
+ primaryClass={cs.AI}
200
+ }
201
+ ```
202
+
203
+ ## License
204
+
205
+ This dataset maintains the Apache 2.0 license from the original ARC-AGI corpus.