--- language: - en license: other pretty_name: P-ARC (PotARCin Test Set) tabular export configs: - config_name: default data_files: - split: test path: p_arc_dataset.csv size_categories: - n<1K tags: - arc - arc-agi - potarcin - program-synthesis task_categories: - other --- # P-ARC CSV export (PotARCin Test2) One CSV file in UTF-8. Each row is one of the fifty P-ARC tasks from PotARCin (`t1.json` through `t50.json`). Besides the usual train/test grids, each row includes the fifty-sample bundle from `t_samples_50.json` as compact JSON (same structure as the file, without the extra whitespace from pretty-printing), plus the `generator.py` and `verifier.py` sources from the matching task folder. ## Files | File | Description | |------|-------------| | `p_arc_dataset.csv` | One row per task (`t1`–`t50`); column details below and in `SCHEMA.json` | | `README.md` | Dataset card (what you are reading now) | | `SCHEMA.json` | Same column layout in JSON for scripts | ## Columns | Column | Description | |--------|-------------| | `task_id` | `t1` … `t50` | | `train_demonstrations_json` | JSON array of training pairs `{input, output}`; grids are nested lists of integers | | `test_input_json` | JSON grid for `test[0].input` | | `test_output_json` | JSON grid for `test[0].output` | | `stable_instances_50_json` | Compact JSON for the object in `t_samples_50.json` | | `generator_py` | Full `generator.py` source | | `verifier_py` | Full `verifier.py` source | Grids follow the usual ARC convention: each row is a list of cell integers. ## Size and reading the CSV The file is fairly large because big JSON blobs and full Python files sit inside cells. In Python, the standard library `csv` module caps field length by default; bump it before you read: ```python csv.field_size_limit(sys.maxsize) ``` ### Stdlib example ```python import csv, json, sys csv.field_size_limit(sys.maxsize) with open("p_arc_dataset.csv", encoding="utf-8", newline="") as f: for row in csv.DictReader(f): train = json.loads(row["train_demonstrations_json"]) test_in = json.loads(row["test_input_json"]) test_out = json.loads(row["test_output_json"]) samples = json.loads(row["stable_instances_50_json"]) # row["generator_py"], row["verifier_py"] ``` ### pandas ```python import sys, pandas as pd import csv as _csv _csv.field_size_limit(sys.maxsize) df = pd.read_csv("p_arc_dataset.csv", encoding="utf-8") ```