File size: 10,422 Bytes
1db6f31 7fe653c 1db6f31 7fe653c 1db6f31 7fe653c 1db6f31 7fe653c 1db6f31 7fe653c 1db6f31 7fe653c 1db6f31 0bf7cd8 b135aa1 0bf7cd8 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 | ---
license: other
pretty_name: ConnectomeBench2
tags:
- connectomics
- proofreading
- 3d
- electron-microscopy
- mesh
size_categories:
- 100K<n<1M
configs:
- config_name: default
data_files:
- split: train
path: train/train-*.parquet
- split: validation
path: val/val-*.parquet
- split: test
path: test/test-*.parquet
---
# ConnectomeBench2
ConnectomeBench2 is a unified benchmark for **automated proofreading of connectomic neural-segmentation data**. **401,170 samples** across 4 species (mouse, fly, human, zebrafish) and 5 sample types (real merge edits, real split edits, synthetic adjacent / junction / synapse controls), with the associated mesh geometry and electron-microscopy (EM) renderings.
Downstream trainers should treat this dataset as the single source of truth for sample identity, labels, train/validation/test split, and which task(s) a row is valid for.
## Context: Connectomic Proofreading
**Connectomics** scans and automatically segments neurons to create large-scale brain maps at cellular resolution. Two types of **segmentation errors** can occur in this process, which need to be corrected (= **proofreading**):
- **False Splits** — corrected via merge corrections
- **False Merges** — corrected via split corrections
*Merge corrections* (of false splits) are applied to *multiple segments* that need to be correctly merged together. *Split corrections* (of false merges) are applied to *single segments* that need to be correctly split apart.
For this reason, this dataset contains renderings of both *single-segment* (pre-split or post-merge) and *dual-segment* (post-split or pre-merge) mesh geometry, where possible. EM data is provided in dual format only — segmentation on imaging level is contiguous, so the single-version can be derived from the union of the dual.
## Renderings (geometry and EM imaging data)

(top: synapse merge-pair — both masks populated; bottom: junction control — single-mask only, mask B / seg B empty)
**Geometry files** (the `geometry` and `geometry_single` columns) are compressed `.npz` payloads that decode to `(3, 7, 224, 224) float16` arrays — three 2D views (front, side, top) × seven channels:
| ch | content |
|---|---|
| 0 | silhouette |
| 1 | depth |
| 2 | normal_x |
| 3 | normal_y |
| 4 | normal_z |
| 5 | mask A |
| 6 | mask B (empty in single-segment renders) |
Note that single and dual segment renders differ not only in mask channels, but also subtly differ in all other channels, due to slight differences in mesh geometry from merging/splitting.
**Free split-mask labels.** For `split_edit` rows, the dual-segment render (post-split) provides ground-truth split-mask labels (Mask A / Mask B channels) for the corresponding single-segment render (pre-split) — split-mask-generation tasks get pixel-level supervision without extra labeling.
**EM coverage.** EM views are not present on every sample. Coverage by `sample_type` (full dataset):
| sample_type | rows | has_em |
|---|---|---|
| adjacent_control | 121,333 | 100% |
| junction_control | 38,272 | 100% |
| synapse_control | 18,182 | 100% |
| merge_edit | 146,461 | 38% |
| split_edit | 77,213 | 23% |
| **total** | **401,170** | **63% (37% null)** |
real human edits (merge_edit, split_edit) only got EM rendered on a stratified subset; synthetic controls all have EM. Filter by `has_em` if your task requires it.
**EM imaging files** (`em_xy` / `em_xz` / `em_yz` / `em_best` columns) are PNG-encoded 3-channel slices:
| ch | content |
|---|---|
| 0 | raw EM intensity |
| 1 | segment A mask |
| 2 | segment B mask |
Four imaging views per sample: three cardinal slices (xy, xz, yz) + a `best` slice at an oblique angle that maximizes the visible area of both segments (sum of their logs).
For single-segment tasks, segment A and B should be merged (and B zeroed). The `best` view may leak some dual-label information (it takes both labels into account); we advise against testing single-segment tasks on `em_best`.
## Loading
```python
from datasets import load_dataset
ds = load_dataset("jeffbbrown2/connectomebench2-smoke", split="train")
sample = ds[0]
# sample["em_xy"] is a PIL Image (HF auto-decodes)
# sample["geometry"] is bytes — decode with:
import io, numpy as np
geom = np.load(io.BytesIO(sample["geometry"]))["arr_0"] # shape (3, 7, 224, 224) float16
```
Or with raw `pyarrow`:
```python
import pyarrow.parquet as pq
import numpy as np, io
df = pq.read_table("train/train-00000.parquet").to_pandas()
geom = np.load(io.BytesIO(df.iloc[0]["geometry"]))["arr_0"]
```
The `metadata/{train,val,test}.parquet` sidecars contain identifier/label/modality columns only (no image bytes) — useful for fast filtering or inspection.
## Columns
### Identifiers
- **`combined_sample_hash`** — primary key (md5 hex 32-char of `f"{source_archive}|{source_archive_sample_hash}"`); guaranteed unique across the dataset.
- **`source_archive_sample_hash`** — legacy 12-char hex hash from upstream; kept for traceability, not unique alone.
- **`source_archive`** — name of the originating render archive (e.g. `edits_and_adj_controls_fly`, `junction_controls_mouse`, `synapse_controls_fly`). 10 distinct values (5 archives × species).
### Sample identity
- **`sample_type: str`** — single source of truth for what kind of sample this row is. Five values:
- `merge_edit` — positive merge-correction edit
- `split_edit` — positive split-correction edit
- `adjacent_control` — synthetic negative for merge-correction (segments adjacent to genuine correction)
- `junction_control` — putative junction in proofread neuron (negative merge-error-id sample)
- `synapse_control` — synapse pair across neurons (negative merge-correction)
- **`same_neuron: bool`** — derived from sample_type:
- `True` for `merge_edit`, `junction_control`
- `False` for `split_edit`, `adjacent_control`, `synapse_control`
- **`species: str`** — `fly` / `mouse` / `human` / `zebrafish`.
### Image content
- **`geometry`** — bytes; compressed npz (key `"arr_0"`) decoding to `(3, 7, 224, 224) float16`. Null when the sample has no dual-segment render.
- **`geometry_single`** — same shape/dtype, single-segment version. Null when not present.
- **`em_xy` / `em_xz` / `em_yz` / `em_best`** — PIL Images (3-channel PNG, `(224, 224, 3) uint8`). Null when the row has no EM views.
- **`has_single_mask: bool`** — convenience flag.
- **`has_dual_mask: bool`** — convenience flag.
- **`has_em: bool`** — true if any `em_*` column is non-null.
- **`present_slots: list[str]`** — modality tags actually present (e.g. `["geometry", "geometry_single", "em_xy", "em_xz", "em_yz", "em_best"]`).
### Task routing & labels
- **`task_routing: list[str]`** — which downstream task(s) this row can serve as training data for:
- `false_split_correction` — merge-correction task; fires when `sample_type ∈ {merge_edit, synapse_control, adjacent_control}` AND `has_dual_mask`.
- `false_merge_identification` — merge-error binary classification; fires when `sample_type ∈ {split_edit, junction_control}` AND `has_single_mask`.
- `split_mask_generation` — pixel-level split prediction; fires when `sample_type == split_edit` AND `has_single_mask`.
- **`false_split_correction_label: bool`** = `same_neuron`. Populated for all rows; trainers filter by `task_routing`.
- **`false_merge_identification_label: bool`** = `not same_neuron`. Populated for all rows; trainers filter by `task_routing`.
**Usage note.** Downstream training scripts must load the appropriate geometry render per task:
- **Merge Correction** of false splits should use **dual-segment** renders
- **Split Correction** of false merges should use **single-segment** renders
- Furthermore, fuse A/B channels of EM images and **discard `em_best`** (it sees both labels at oblique angle and can leak ground truth)
Otherwise, ground-truth task or label information may leak to the model and bias performance.
### Train/val/test split
- **`split: str`** — `train` / `validation` / `test`. ~80/10/10 split assigned by spatial location of the proofreading sample (`interface_point_nm`), matched via cube splits (50µm cubes tiling the volume and randomly split).
### Other
- **`metadata: str`** — JSON-stringified original metadata struct. Parse with `json.loads`. Useful keys: `operation_id`, `source_operation_id`, `strategy`, `image_types`, `interface_point_nm`, `before_root_ids`, `after_root_ids`, …
## Counts
- **401,170 rows** total · ~80/11/9 train (319,727) / validation (43,517) / test (37,926)
- 251,499 rows with EM views; all 401,170 have geometry
- **~2.2M model-level samples** (EM × 4 views + geom × 3 views), or **~2.8M** counting dual + single geom separately
- 506 parquet shards (~240 MB each)
## Layout
```
README.md
shards.csv metadata across shards (path, sha256, n_samples, size)
train/train-*.parquet WebDataset-style parquet shards with image bytes
val/val-*.parquet
test/test-*.parquet
metadata/ sidecar parquets with identifiers + labels (no bytes)
train.parquet
val.parquet
test.parquet
demo.parquet stratified mini-shard (one-line preview)
figures/
channel_decomposition.png
```
## Sources & License
Derived from the following upstream connectomic proofreading datasets:
- **MICrONS** (mouse cortex)
- **FlyWire** (Drosophila brain)
- **H01** (human cortex)
- Zebrafish larval connectome
License = `other`; users must comply with upstream licenses (which may differ across species/sources). Final outbound license will be set after upstream license review.
## Citation
If you use ConnectomeBench2, please cite:
```
Brown, J., Farkas, T., Razgar, G., Boyden, E. S.
ConnectomeBench2: A unified benchmark for automated connectomic proofreading.
(2026, in submission). Brown J. and Farkas T. contributed equally as first authors.
```
Please also cite the upstream connectome sources used by this dataset:
- MICrONS (mouse cortex): https://www.microns-explorer.org/cortical-mm3
- FlyWire (Drosophila): https://flywire.ai/
- H01 (human cortex): https://h01-release.storage.googleapis.com/landing.html
- Zebrafish (fish1): https://fish1-release.storage.googleapis.com/index.html
|