File size: 10,422 Bytes
1db6f31
 
7fe653c
1db6f31
 
 
 
 
 
 
7fe653c
1db6f31
 
 
 
 
 
 
 
 
 
 
 
 
7fe653c
1db6f31
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7fe653c
1db6f31
7fe653c
1db6f31
 
7fe653c
1db6f31
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0bf7cd8
 
 
 
 
 
b135aa1
0bf7cd8
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
---
license: other
pretty_name: ConnectomeBench2
tags:
  - connectomics
  - proofreading
  - 3d
  - electron-microscopy
  - mesh
size_categories:
  - 100K<n<1M
configs:
  - config_name: default
    data_files:
      - split: train
        path: train/train-*.parquet
      - split: validation
        path: val/val-*.parquet
      - split: test
        path: test/test-*.parquet
---

# ConnectomeBench2

ConnectomeBench2 is a unified benchmark for **automated proofreading of connectomic neural-segmentation data**. **401,170 samples** across 4 species (mouse, fly, human, zebrafish) and 5 sample types (real merge edits, real split edits, synthetic adjacent / junction / synapse controls), with the associated mesh geometry and electron-microscopy (EM) renderings.

Downstream trainers should treat this dataset as the single source of truth for sample identity, labels, train/validation/test split, and which task(s) a row is valid for.

## Context: Connectomic Proofreading

**Connectomics** scans and automatically segments neurons to create large-scale brain maps at cellular resolution. Two types of **segmentation errors** can occur in this process, which need to be corrected (= **proofreading**):

- **False Splits** — corrected via merge corrections
- **False Merges** — corrected via split corrections

*Merge corrections* (of false splits) are applied to *multiple segments* that need to be correctly merged together. *Split corrections* (of false merges) are applied to *single segments* that need to be correctly split apart.

For this reason, this dataset contains renderings of both *single-segment* (pre-split or post-merge) and *dual-segment* (post-split or pre-merge) mesh geometry, where possible. EM data is provided in dual format only — segmentation on imaging level is contiguous, so the single-version can be derived from the union of the dual.

## Renderings (geometry and EM imaging data)

![channel decomposition: synapse 2-mask vs junction single-mask](figures/channel_decomposition.png)

(top: synapse merge-pair — both masks populated; bottom: junction control — single-mask only, mask B / seg B empty)

**Geometry files** (the `geometry` and `geometry_single` columns) are compressed `.npz` payloads that decode to `(3, 7, 224, 224) float16` arrays — three 2D views (front, side, top) × seven channels:

| ch | content |
|---|---|
| 0 | silhouette |
| 1 | depth |
| 2 | normal_x |
| 3 | normal_y |
| 4 | normal_z |
| 5 | mask A |
| 6 | mask B (empty in single-segment renders) |

Note that single and dual segment renders differ not only in mask channels, but also subtly differ in all other channels, due to slight differences in mesh geometry from merging/splitting.

**Free split-mask labels.** For `split_edit` rows, the dual-segment render (post-split) provides ground-truth split-mask labels (Mask A / Mask B channels) for the corresponding single-segment render (pre-split) — split-mask-generation tasks get pixel-level supervision without extra labeling.

**EM coverage.** EM views are not present on every sample. Coverage by `sample_type` (full dataset):

| sample_type | rows | has_em |
|---|---|---|
| adjacent_control | 121,333 | 100% |
| junction_control | 38,272 | 100% |
| synapse_control | 18,182 | 100% |
| merge_edit | 146,461 | 38% |
| split_edit | 77,213 | 23% |
| **total** | **401,170** | **63% (37% null)** |

real human edits (merge_edit, split_edit) only got EM rendered on a stratified subset; synthetic controls all have EM. Filter by `has_em` if your task requires it.

**EM imaging files** (`em_xy` / `em_xz` / `em_yz` / `em_best` columns) are PNG-encoded 3-channel slices:

| ch | content |
|---|---|
| 0 | raw EM intensity |
| 1 | segment A mask |
| 2 | segment B mask |

Four imaging views per sample: three cardinal slices (xy, xz, yz) + a `best` slice at an oblique angle that maximizes the visible area of both segments (sum of their logs).

For single-segment tasks, segment A and B should be merged (and B zeroed). The `best` view may leak some dual-label information (it takes both labels into account); we advise against testing single-segment tasks on `em_best`.

## Loading

```python
from datasets import load_dataset

ds = load_dataset("jeffbbrown2/connectomebench2-smoke", split="train")
sample = ds[0]
# sample["em_xy"] is a PIL Image (HF auto-decodes)
# sample["geometry"] is bytes — decode with:
import io, numpy as np
geom = np.load(io.BytesIO(sample["geometry"]))["arr_0"]   # shape (3, 7, 224, 224) float16
```

Or with raw `pyarrow`:

```python
import pyarrow.parquet as pq
import numpy as np, io
df = pq.read_table("train/train-00000.parquet").to_pandas()
geom = np.load(io.BytesIO(df.iloc[0]["geometry"]))["arr_0"]
```

The `metadata/{train,val,test}.parquet` sidecars contain identifier/label/modality columns only (no image bytes) — useful for fast filtering or inspection.

## Columns

### Identifiers
- **`combined_sample_hash`** — primary key (md5 hex 32-char of `f"{source_archive}|{source_archive_sample_hash}"`); guaranteed unique across the dataset.
- **`source_archive_sample_hash`** — legacy 12-char hex hash from upstream; kept for traceability, not unique alone.
- **`source_archive`** — name of the originating render archive (e.g. `edits_and_adj_controls_fly`, `junction_controls_mouse`, `synapse_controls_fly`). 10 distinct values (5 archives × species).

### Sample identity
- **`sample_type: str`** — single source of truth for what kind of sample this row is. Five values:
  - `merge_edit` — positive merge-correction edit
  - `split_edit` — positive split-correction edit
  - `adjacent_control` — synthetic negative for merge-correction (segments adjacent to genuine correction)
  - `junction_control` — putative junction in proofread neuron (negative merge-error-id sample)
  - `synapse_control` — synapse pair across neurons (negative merge-correction)
- **`same_neuron: bool`** — derived from sample_type:
  - `True` for `merge_edit`, `junction_control`
  - `False` for `split_edit`, `adjacent_control`, `synapse_control`
- **`species: str`**`fly` / `mouse` / `human` / `zebrafish`.

### Image content
- **`geometry`** — bytes; compressed npz (key `"arr_0"`) decoding to `(3, 7, 224, 224) float16`. Null when the sample has no dual-segment render.
- **`geometry_single`** — same shape/dtype, single-segment version. Null when not present.
- **`em_xy` / `em_xz` / `em_yz` / `em_best`** — PIL Images (3-channel PNG, `(224, 224, 3) uint8`). Null when the row has no EM views.
- **`has_single_mask: bool`** — convenience flag.
- **`has_dual_mask: bool`** — convenience flag.
- **`has_em: bool`** — true if any `em_*` column is non-null.
- **`present_slots: list[str]`** — modality tags actually present (e.g. `["geometry", "geometry_single", "em_xy", "em_xz", "em_yz", "em_best"]`).

### Task routing & labels
- **`task_routing: list[str]`** — which downstream task(s) this row can serve as training data for:
  - `false_split_correction` — merge-correction task; fires when `sample_type ∈ {merge_edit, synapse_control, adjacent_control}` AND `has_dual_mask`.
  - `false_merge_identification` — merge-error binary classification; fires when `sample_type ∈ {split_edit, junction_control}` AND `has_single_mask`.
  - `split_mask_generation` — pixel-level split prediction; fires when `sample_type == split_edit` AND `has_single_mask`.
- **`false_split_correction_label: bool`** = `same_neuron`. Populated for all rows; trainers filter by `task_routing`.
- **`false_merge_identification_label: bool`** = `not same_neuron`. Populated for all rows; trainers filter by `task_routing`.

**Usage note.** Downstream training scripts must load the appropriate geometry render per task:
- **Merge Correction** of false splits should use **dual-segment** renders
- **Split Correction** of false merges should use **single-segment** renders
  - Furthermore, fuse A/B channels of EM images and **discard `em_best`** (it sees both labels at oblique angle and can leak ground truth)

Otherwise, ground-truth task or label information may leak to the model and bias performance.

### Train/val/test split
- **`split: str`** — `train` / `validation` / `test`. ~80/10/10 split assigned by spatial location of the proofreading sample (`interface_point_nm`), matched via cube splits (50µm cubes tiling the volume and randomly split).

### Other
- **`metadata: str`** — JSON-stringified original metadata struct. Parse with `json.loads`. Useful keys: `operation_id`, `source_operation_id`, `strategy`, `image_types`, `interface_point_nm`, `before_root_ids`, `after_root_ids`, …

## Counts

- **401,170 rows** total · ~80/11/9 train (319,727) / validation (43,517) / test (37,926)
- 251,499 rows with EM views; all 401,170 have geometry
- **~2.2M model-level samples** (EM × 4 views + geom × 3 views), or **~2.8M** counting dual + single geom separately
- 506 parquet shards (~240 MB each)

## Layout

```
README.md
shards.csv                    metadata across shards (path, sha256, n_samples, size)
train/train-*.parquet         WebDataset-style parquet shards with image bytes
val/val-*.parquet
test/test-*.parquet
metadata/                     sidecar parquets with identifiers + labels (no bytes)
  train.parquet
  val.parquet
  test.parquet
demo.parquet                  stratified mini-shard (one-line preview)
figures/
  channel_decomposition.png
```

## Sources & License

Derived from the following upstream connectomic proofreading datasets:
- **MICrONS** (mouse cortex)
- **FlyWire** (Drosophila brain)
- **H01** (human cortex)
- Zebrafish larval connectome

License = `other`; users must comply with upstream licenses (which may differ across species/sources). Final outbound license will be set after upstream license review.

## Citation

If you use ConnectomeBench2, please cite:

```
Brown, J., Farkas, T., Razgar, G., Boyden, E. S.
ConnectomeBench2: A unified benchmark for automated connectomic proofreading.
(2026, in submission). Brown J. and Farkas T. contributed equally as first authors.
```

Please also cite the upstream connectome sources used by this dataset:
- MICrONS (mouse cortex): https://www.microns-explorer.org/cortical-mm3
- FlyWire (Drosophila): https://flywire.ai/
- H01 (human cortex): https://h01-release.storage.googleapis.com/landing.html
- Zebrafish (fish1): https://fish1-release.storage.googleapis.com/index.html