P-ARC / README.md
PotARCin's picture
Update README.md
b0c5381 verified
---
language:
- en
license: other
pretty_name: P-ARC (PotARCin Test Set) tabular export
configs:
- config_name: default
data_files:
- split: test
path: p_arc_dataset.csv
size_categories:
- n<1K
tags:
- arc
- arc-agi
- potarcin
- program-synthesis
task_categories:
- other
---
# P-ARC CSV export (PotARCin Test2)
One CSV file in UTF-8. Each row is one of the fifty P-ARC tasks from PotARCin (`t1.json` through `t50.json`). Besides the usual train/test grids, each row includes the fifty-sample bundle from `t<n>_samples_50.json` as compact JSON (same structure as the file, without the extra whitespace from pretty-printing), plus the `generator.py` and `verifier.py` sources from the matching task folder.
## Files
| File | Description |
|------|-------------|
| `p_arc_dataset.csv` | One row per task (`t1``t50`); column details below and in `SCHEMA.json` |
| `README.md` | Dataset card (what you are reading now) |
| `SCHEMA.json` | Same column layout in JSON for scripts |
## Columns
| Column | Description |
|--------|-------------|
| `task_id` | `t1``t50` |
| `train_demonstrations_json` | JSON array of training pairs `{input, output}`; grids are nested lists of integers |
| `test_input_json` | JSON grid for `test[0].input` |
| `test_output_json` | JSON grid for `test[0].output` |
| `stable_instances_50_json` | Compact JSON for the object in `t<n>_samples_50.json` |
| `generator_py` | Full `generator.py` source |
| `verifier_py` | Full `verifier.py` source |
Grids follow the usual ARC convention: each row is a list of cell integers.
## Size and reading the CSV
The file is fairly large because big JSON blobs and full Python files sit inside cells. In Python, the standard library `csv` module caps field length by default; bump it before you read:
```python
csv.field_size_limit(sys.maxsize)
```
### Stdlib example
```python
import csv, json, sys
csv.field_size_limit(sys.maxsize)
with open("p_arc_dataset.csv", encoding="utf-8", newline="") as f:
for row in csv.DictReader(f):
train = json.loads(row["train_demonstrations_json"])
test_in = json.loads(row["test_input_json"])
test_out = json.loads(row["test_output_json"])
samples = json.loads(row["stable_instances_50_json"])
# row["generator_py"], row["verifier_py"]
```
### pandas
```python
import sys, pandas as pd
import csv as _csv
_csv.field_size_limit(sys.maxsize)
df = pd.read_csv("p_arc_dataset.csv", encoding="utf-8")
```