Datasets:
Modalities:
Image
Formats:
imagefolder
Languages:
English
Size:
1K - 10K
ArXiv:
Tags:
video-object-segmentation
computer-vision
angiography
x-ray
medical-image-segmentation
video-segmentation
License:
Update README.md
Browse files
README.md
CHANGED
|
@@ -13,4 +13,167 @@ tags:
|
|
| 13 |
- medical-image-segmentation
|
| 14 |
- video-segmentation
|
| 15 |
pretty_name: MOSXAV
|
| 16 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
- medical-image-segmentation
|
| 14 |
- video-segmentation
|
| 15 |
pretty_name: MOSXAV
|
| 16 |
+
size_categories:
|
| 17 |
+
- 10K<n<100K
|
| 18 |
+
---
|
| 19 |
+
|
| 20 |
+
# <p align="center">MOSXAV: A Benchmark Dataset for Multi-Object Segmentation in X-ray Angiography Videos</p>
|
| 21 |
+
|
| 22 |
+
<p align="center">
|
| 23 |
+
<a href="https://xilin-x.github.io/MOSXAV/" target="_blank">
|
| 24 |
+
<img alt="Static Badge" src="https://img.shields.io/badge/Homepage-MOSXAV-yellow?logo=databricks&logoColor=%23FF3621">
|
| 25 |
+
</a>
|
| 26 |
+
</p>
|
| 27 |
+
|
| 28 |
+
<p align="center">
|
| 29 |
+
<b>VOS Task:</b>
|
| 30 |
+
|
| 31 |
+
<a href="https://arxiv.org/abs/2601.00988" target="_blank">
|
| 32 |
+
<img src="https://img.shields.io/badge/arXiv-2601.00988-b31b1b?style=flat&logo=arXiv&logoColor=%23B31B1B">
|
| 33 |
+
</a>
|
| 34 |
+
|
|
| 35 |
+
<b>Semantic Segmentation Task:</b>
|
| 36 |
+
|
| 37 |
+
<a href="https://link.springer.com/chapter/10.1007/978-3-032-05472-2_2" target="_blank">
|
| 38 |
+
<img alt="Static Badge" src="https://img.shields.io/badge/SN-DGM4MICCAI 2025-f8f8f8?style=flat&labelColor=01324b">
|
| 39 |
+
</a>
|
| 40 |
+
|
| 41 |
+
<a href="https://arxiv.org/abs/2507.16429" target="_blank">
|
| 42 |
+
<img src="https://img.shields.io/badge/arXiv-2507.16429-b31b1b?style=flat&logo=arXiv&logoColor=%23B31B1B">
|
| 43 |
+
</a>
|
| 44 |
+
</p>
|
| 45 |
+
|
| 46 |
+
---
|
| 47 |
+
|
| 48 |
+
## 📖 1. Overview
|
| 49 |
+
|
| 50 |
+
**MOSXAV** is a benchmark dataset designed for **multi-object segmentation** in X-ray angiography videos. It provides **high-quality, manually annotated segmentation ground truth**, supporting the analysis of vascular structures in dynamic medical imaging. Each video contains **33∼70** frames at a resolution of **512×512 pixels**. Vascular regions are annotated by experienced radiologists, with annotations focused on **one or two key frames** where the contrast agent is most prominent.
|
| 51 |
+
|
| 52 |
+
- The **training and validation sets** include **30 sequences** (2,335 frames), with **annotations every 5 frames**.
|
| 53 |
+
- The **test set** consists of **12 sequences** (488 frames), with **frame-level annotations** throughout.
|
| 54 |
+
|
| 55 |
+
MOSXAV provides a valuable resource for the development and benchmarking of methods in X-ray angiography video segmentation.
|
| 56 |
+
|
| 57 |
+
---
|
| 58 |
+
|
| 59 |
+
## 🛠️ 2. Annotation Protocol
|
| 60 |
+
|
| 61 |
+
To ensure high-quality and anatomically accurate labels, we implemented a rigorous, multi-stage annotation workflow. This process combined the efficiency of deep learning with the precision of manual expert refinement. The protocol consisted of four primary phases:
|
| 62 |
+
|
| 63 |
+
* **Annotator Training:** Annotators were trained on a specialized subset of images to standardize their understanding of anatomical structures and specific labeling guidelines.
|
| 64 |
+
* **Semi-Automated Initialization:** We utilized a semi-automated approach to generate initial segmentation masks. This was powered by the [PaddleSeg](https://github.com/PaddlePaddle/PaddleSeg) framework, leveraging models pre-trained on extensive image and video datasets to provide a robust baseline.
|
| 65 |
+
* **Expert Revision:** Human annotators meticulously reviewed the AI-generated masks. This involved careful delineation of vessel boundaries and manual adjustments to correct any discrepancies in the automated output.
|
| 66 |
+
* **Consensus & Quality Assurance:** To maintain consistency, a final review and consensus-building phase were conducted, ensuring that all labels met our strict quality benchmarks.
|
| 67 |
+
|
| 68 |
+
---
|
| 69 |
+
|
| 70 |
+
## 📊 3. Object Categories and Statistics
|
| 71 |
+
|
| 72 |
+
The **MOSXAV dataset** is designed to support two distinct medical imaging challenges: Video Object Segmentation (VOS) and Multi-class Semantic Segmentation.
|
| 73 |
+
|
| 74 |
+
### 3.1 Video Object Segmentation (VOS)
|
| 75 |
+
|
| 76 |
+
The VOS task focuses on the temporal tracking and segmentation of coronary arteries as they are opacified by contrast agents. This task requires high temporal consistency across video sequences:
|
| 77 |
+
* **Objective**: Segmenting coronary arteries filled with contrast agents throughout the cardiac cycle.
|
| 78 |
+
* **Object Density**: **Train & Val Sets:** Up to 5 individual objects per sequence. **Test Set:** Increased complexity with up to 10 individual objects per sequence to evaluate model scalability and robustness.
|
| 79 |
+
|
| 80 |
+
### 3.2 Semantic Segmentation
|
| 81 |
+
|
| 82 |
+
The semantic segmentation task targets the simultaneous identification of critical intervention tools and anatomical features. We define four primary categories:
|
| 83 |
+
|
| 84 |
+
| Category | Label ID | Color Preview | RGB Color | Description |
|
| 85 |
+
| :--- | :---: | :---: | :---: | :--- |
|
| 86 |
+
| **Background** | 0 |  | `[0,0,0]` | All pixels not belonging to the below classes, including the spine, ribs, diaphragm, and image noise. |
|
| 87 |
+
| **Vessel** | 1 |  | `[85,170,255]` | The primary coronary anatomy under observation. |
|
| 88 |
+
| **Contrast Catheter** | 2 |  | `[170,255,0]` | The specific catheter used for dye injection. |
|
| 89 |
+
| **Catheter** | 3 |  | `[249,193,0]` | General-purpose intervention or diagnostic catheters. |
|
| 90 |
+
| **Balloon** | 4 |  | `[255,0,0]` | Angioplasty balloons used during interventional procedures. |
|
| 91 |
+
| **Others** | 5 |  | `[244,108,59]` | Other category. |
|
| 92 |
+
|
| 93 |
+
### 3.3 File Structure
|
| 94 |
+
|
| 95 |
+
The MOSXAV dataset is organized into a hierarchical directory structure to support both video-level (VOS) and frame-level (Semantic Segmentation) tasks. The data is split into ```trainval```, and ```test``` directories, each containing the original sequences and their corresponding pixel-level annotations.
|
| 96 |
+
|
| 97 |
+
```text
|
| 98 |
+
MOSXAV_Dataset/
|
| 99 |
+
├── trainval/
|
| 100 |
+
│ ├── Annotations/ # VOS instance masks (unique ID per branch)
|
| 101 |
+
│ │ └── v00/ # Sequence folder
|
| 102 |
+
│ │ ├── 00000.png # Frame-wise instance mask
|
| 103 |
+
│ │ └── ...
|
| 104 |
+
│ ├── Annotations_Semantic/ # Multi-class semantic masks (Label IDs 0-4)
|
| 105 |
+
│ │ └── v00/
|
| 106 |
+
│ │ ├── 00000.png
|
| 107 |
+
│ │ └── ...
|
| 108 |
+
│ ├── JPEGImages/ # Raw X-ray Angiography frames
|
| 109 |
+
│ │ └── v00/
|
| 110 |
+
│ │ ├── 00000.jpg
|
| 111 |
+
│ │ └── ...
|
| 112 |
+
│ ├── ImageSets/ # Split lists and first-frame metadata
|
| 113 |
+
│ │ ├── train.txt
|
| 114 |
+
│ │ ├── val.txt
|
| 115 |
+
│ │ └── val_first_mask.json # Each object's first appearance frame ID
|
| 116 |
+
│ └── labels.json # Global category metadata
|
| 117 |
+
└── test/
|
| 118 |
+
├── Annotations/
|
| 119 |
+
├── Annotations_Semantic/ # Multi-class semantic masks (Label IDs 0-5)
|
| 120 |
+
├── JPEGImages/
|
| 121 |
+
├── ImageSets/
|
| 122 |
+
│ ├── test.txt
|
| 123 |
+
│ └── test_first_mask.json # Each object's first appearance frame ID
|
| 124 |
+
└── labels.json
|
| 125 |
+
```
|
| 126 |
+
|
| 127 |
+
---
|
| 128 |
+
|
| 129 |
+
## 📥 4. Download
|
| 130 |
+
|
| 131 |
+
The MOSXAV Dataset is hosted across multiple cloud storage platforms to ensure accessibility and high download speeds globally.
|
| 132 |
+
|
| 133 |
+
| Source | Download Link | Extraction Code / Notes |
|
| 134 |
+
| :---: | :---: | :---: |
|
| 135 |
+
| **OneDrive** | [🌐 Click Here](https://1drv.ms/f/c/05b6df5b859ecdde/IgBLUOUDkt5bQ4S80HI2Gb_7AZW-uXoxuXuFPXCrwOUrBPo?e=6fBF22) |  |
|
| 136 |
+
| **Google Drive** | [🌐 Click Here](https://drive.google.com/drive/folders/1d-kOWF7TkXqkRAfmh0ugkEvT9NmXr3cP?usp=sharing) | - |
|
| 137 |
+
| **Baidu Pan** | [🌐 Click Here](https://pan.baidu.com/s/1-1i6nljdrdp90tMg30bZwQ) |  |
|
| 138 |
+
|
| 139 |
+
---
|
| 140 |
+
|
| 141 |
+
## ⚖️ 5. License
|
| 142 |
+
|
| 143 |
+
The dataset is licensed under a [Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License](https://creativecommons.org/licenses/by-nc-sa/4.0/). See [LICENSE](./LICENSE) for details.
|
| 144 |
+
|
| 145 |
+
<br>
|
| 146 |
+
<p align="left">
|
| 147 |
+
<a rel="license" href="http://creativecommons.org/licenses/by-nc-sa/4.0/">
|
| 148 |
+
<img alt="Creative Commons License" style="border-width:0" src="https://i.creativecommons.org/l/by-nc-sa/4.0/88x31.png" />
|
| 149 |
+
</a>
|
| 150 |
+
</p>
|
| 151 |
+
|
| 152 |
+
---
|
| 153 |
+
|
| 154 |
+
## 📝 Citation
|
| 155 |
+
|
| 156 |
+
Please consider to cite MOSXAV if it helps your research.
|
| 157 |
+
|
| 158 |
+
```bibtex
|
| 159 |
+
@InProceedings{10.1007/978-3-032-05472-2_2,
|
| 160 |
+
author={Xi, Lin and Ma, Yingliang and Wang, Cheng and Howell, Sandra and Rinaldi, Aldo and Rhode, Kawal S.},
|
| 161 |
+
title={Robust Noisy Pseudo-Label Learning for Semi-supervised Medical Image Segmentation Using Diffusion Model},
|
| 162 |
+
booktitle={Deep Generative Models Workshop, International Conference on Medical Image Computing and Computer-Assisted Intervention (DGM4MICCAI)},
|
| 163 |
+
year={2026},
|
| 164 |
+
pages={12--23}
|
| 165 |
+
}
|
| 166 |
+
```
|
| 167 |
+
|
| 168 |
+
---
|
| 169 |
+
|
| 170 |
+
## ✉️ Contact
|
| 171 |
+
|
| 172 |
+
For questions or feedback, please contact:
|
| 173 |
+
* Lin Xi: 
|
| 174 |
+
* Yingliang Ma: 
|
| 175 |
+
|
| 176 |
+
---
|
| 177 |
+
<p align="center">
|
| 178 |
+
Copyright © 2026 MOSXAV Project Team. All rights reserved.
|
| 179 |
+
</p>
|