Dataset Viewer
Auto-converted to Parquet Duplicate
even
array 3D
odd
array 3D
sample_ids
stringlengths
36
36
[[[0.008563272655010223,0.010466222651302814,0.011417697183787823,0.010466222651302814,0.01141769718(...TRUNCATED)
[[[0.009514748118817806,0.013320647180080414,0.014272121712565422,0.009514748118817806,0.00856327265(...TRUNCATED)
voicecoil_532nm_01_01_full_r000_c000
[[[0.011417697183787823,0.01522359624505043,0.012369172647595406,0.012369172647595406,0.008563272655(...TRUNCATED)
[[[0.013320647180080414,0.009514748118817806,0.007611798122525215,0.009514748118817806,0.00951474811(...TRUNCATED)
voicecoil_532nm_01_01_full_r000_c001
[[[0.014272121712565422,0.014272121712565422,0.014272121712565422,0.013320647180080414,0.01332064718(...TRUNCATED)
[[[0.009514748118817806,0.007611798122525215,0.008563272655010223,0.007611798122525215,0.00951474811(...TRUNCATED)
voicecoil_532nm_01_01_full_r000_c002
[[[0.011417697183787823,0.008563272655010223,0.007611798122525215,0.007611798122525215,0.00951474811(...TRUNCATED)
[[[0.006660323590040207,0.004757374059408903,0.005708848591893911,0.008563272655010223,0.01236917264(...TRUNCATED)
voicecoil_532nm_01_01_full_r000_c003
[[[0.008563272655010223,0.009514748118817806,0.008563272655010223,0.009514748118817806,0.01141769718(...TRUNCATED)
[[[0.007611798122525215,0.01617507077753544,0.017126545310020447,0.053282588720321655,0.053282588720(...TRUNCATED)
voicecoil_532nm_01_01_full_r001_c000
[[[0.012369172647595406,0.007611798122525215,0.007611798122525215,0.014272121712565422,0.03044719249(...TRUNCATED)
[[[0.03615604341030121,0.043767839670181274,0.039010465145111084,0.03139866888523102,0.0142721217125(...TRUNCATED)
voicecoil_532nm_01_01_full_r001_c001
[[[0.029495717957615852,0.05708848685026169,0.06945765763521194,0.05994291231036186,0.01046622265130(...TRUNCATED)
[[[0.012369172647595406,0.017126545310020447,0.029495717957615852,0.03520456701517105,0.039010465145(...TRUNCATED)
voicecoil_532nm_01_01_full_r001_c002
[[[0.03139866888523102,0.029495717957615852,0.01522359624505043,0.011417697183787823,0.0133206471800(...TRUNCATED)
[[[0.013320647180080414,0.012369172647595406,0.009514748118817806,0.011417697183787823,0.00761179812(...TRUNCATED)
voicecoil_532nm_01_01_full_r001_c003
[[[0.010466222651302814,0.011417697183787823,0.011417697183787823,0.01522359624505043,0.016175070777(...TRUNCATED)
[[[0.007611798122525215,0.012369172647595406,0.012369172647595406,0.013320647180080414,0.01998097077(...TRUNCATED)
voicecoil_532nm_01_01_full_r002_c000
[[[0.01998097077012062,0.019029496237635612,0.012369172647595406,0.008563272655010223,0.013320647180(...TRUNCATED)
[[[0.011417697183787823,0.011417697183787823,0.011417697183787823,0.011417697183787823,0.01046622265(...TRUNCATED)
voicecoil_532nm_01_01_full_r002_c001
End of preview. Expand in Data Studio

OR-PAM-Reg-4K: A Benchmark Dataset for Bidirectional OR-PAM Registration

License: CC BY-NC 4.0 Dataset DOI

Overview

OR-PAM-Reg-4K is a benchmark dataset for evaluating image registration methods in bidirectional optical-resolution photoacoustic microscopy (OR-PAM). The dataset contains 4,248 paired image samples (512 × 512 pixels) of in vivo mouse brain vasculature, with separated odd-line (forward scan) and even-line (backward scan) acquisitions.

Dataset Statistics

Split Samples Percentage
Train 3,396 80%
Val 420 10%
Test 432 10%
Total 4,248 100%

Data Format

OR-PAM-Reg-4K/
├── train.h5    # Training set (1.5 GB)
├── val.h5      # Validation set (193 MB)
├── test.h5     # Test set (200 MB)
└── README.md   # This file

HDF5 Structure

Each .h5 file contains:

Key Shape Dtype Description
odd (N, 1, 512, 256) float32 Odd-line (forward scan) images
even (N, 1, 512, 256) float32 Even-line (backward scan) images
sample_ids (N,) string Unique sample identifiers
  • Pixel values: Normalized to [0, 1]
  • Channel: Single-channel grayscale (photoacoustic signal intensity)
  • Dimensions: (batch, channel, height, width) in PyTorch convention

Quick Start

Loading Data (Python)

import h5py
import torch
from torch.utils.data import Dataset

class ORPAMRegDataset(Dataset):
    def __init__(self, h5_path):
        with h5py.File(h5_path, 'r') as f:
            self.odd = f['odd'][:]
            self.even = f['even'][:]

    def __len__(self):
        return len(self.odd)

    def __getitem__(self, idx):
        odd = torch.from_numpy(self.odd[idx]).float()
        even = torch.from_numpy(self.even[idx]).float()
        return odd, even

# Usage
train_dataset = ORPAMRegDataset('train.h5')
odd, even = train_dataset[0]
print(f'Odd shape: {odd.shape}, Even shape: {even.shape}')
# Output: Odd shape: torch.Size([1, 512, 256]), Even shape: torch.Size([1, 512, 256])

Memory-Efficient Loading

For large-scale training, use memory-mapped loading:

import numpy as np

class ORPAMRegDatasetMmap(Dataset):
    def __init__(self, data_dir, split='train'):
        # First convert h5 to npy if needed
        self.odd = np.load(f'{data_dir}/{split}_odd.npy', mmap_mode='r')
        self.even = np.load(f'{data_dir}/{split}_even.npy', mmap_mode='r')

    def __getitem__(self, idx):
        return (torch.from_numpy(self.odd[idx].copy()).float(),
                torch.from_numpy(self.even[idx].copy()).float())

Benchmark Task

Goal: Register even-line images to odd-line images, correcting both domain shift and geometric misalignment.

Evaluation Metrics

Metric Description Range
NCC Normalized Cross-Correlation [-1, 1], higher is better
SSIM Structural Similarity Index [0, 1], higher is better
PSNR Peak Signal-to-Noise Ratio dB, higher is better

Baseline Results (Author Implementation)

The following baselines are implemented and evaluated by the dataset authors on the test set (432 samples).

Traditional Methods

Method SSIM PSNR NCC
Unregistered 0.482 19.46 0.167
SIFT 0.679 24.14 0.723
Demons 0.579 20.35 0.323
Optical Flow 0.455 18.99 0.061
SyN (ANTs) 0.613 21.55 0.411

Deep Learning Methods

Method SSIM PSNR NCC
VoxelMorph 0.659 22.88 0.724
TransMorph 0.641 21.09 0.594

Published Methods (Third-Party Comparison)

Results reported in the original publications and our reproduction on OR-PAM-Reg-4K. We welcome community contributions to this leaderboard.

Method Paper Reported SSIM Reported NCC Reproduced SSIM Reproduced NCC
To be updated To be updated To be updated To be updated

If you have evaluated your method on OR-PAM-Reg-4K, please open an issue or contact us to add your results.

Citation

If you use this dataset, please cite:

@misc{tianyan_zhang_2026,
  author    = {Tianyan Zhang and Chengliu Yan and Xiangzhi Lan},
  title     = {OR-PAM-Reg-4K: A Benchmark Dataset for Bidirectional OR-PAM Registration},
  year      = {2026},
  url       = {https://huggingface.co/datasets/chengliuyan/OR-PAM-Reg-4K},
  doi       = {10.57967/hf/7721},
  publisher = {Hugging Face}
}

License

This dataset is released under the Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0).

Downloads last month
28