File size: 6,649 Bytes
d7ab5b1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 |
# Warehouse Object Detection Dataset
## Overview
This is a synthetic warehouse object detection dataset generated using NVIDIA Omniverse Replicator. The dataset contains high-quality RGB images with comprehensive annotations including 2D bounding boxes, instance segmentation masks, depth maps, and 3D primitive paths.
**Version:** 1.0.0
**License:** CC-BY-4.0
**Total Images:** 6,234
**Total Annotations:** 78,441
**Image Resolution:** 512x512
**Number of Classes:** 25
## Dataset Statistics
### Split Distribution
- **Training Set:** 4,363 images (70.0%)
- **Validation Set:** 935 images (15.0%)
- **Test Set:** 936 images (15.0%)
### Annotation Statistics
- **Average Annotations per Image:** 12.6
- **Total Bounding Boxes:** 78,441
### Top 10 Most Common Classes
1. **wall**: 6,234 instances in 6,234 images
2. **floor**: 6,232 instances in 6,232 images
3. **sign**: 5,945 instances in 5,945 images
4. **floor_decal**: 5,945 instances in 5,945 images
5. **pillar**: 5,845 instances in 5,845 images
6. **rack**: 5,824 instances in 5,824 images
7. **box**: 5,789 instances in 5,789 images
8. **pallet**: 5,743 instances in 5,743 images
9. **bracket**: 4,857 instances in 4,857 images
10. **lamp**: 4,451 instances in 4,451 images
## Classes
The dataset includes 25 object classes organized into the following categories:
### Container
- **box**: 5,789 annotations
- **crate**: 1,270 annotations
- **barrel**: 1,229 annotations
- **bottle**: 665 annotations
- **bucket**: 457 annotations
### Storage
- **pallet**: 5,743 annotations
- **rack**: 5,824 annotations
### Infrastructure
- **bracket**: 4,857 annotations
- **pillar**: 5,845 annotations
- **emergency_board**: 569 annotations
### Equipment
- **lamp**: 4,451 annotations
- **sign**: 5,945 annotations
- **wire**: 3,985 annotations
- **fuse_box**: 1,154 annotations
- **fire_extinguisher**: 3,304 annotations
- **forklift**: 343 annotations
- **cart**: 427 annotations
- **cone**: 639 annotations
### Markers
- **floor_decal**: 5,945 annotations
- **barcode**: 1,137 annotations
- **paper_note**: 2,125 annotations
- **paper_shortcut**: 390 annotations
### Background
- **wall**: 6,234 annotations
- **ceiling**: 3,882 annotations
- **floor**: 6,232 annotations
## Dataset Formats
This dataset is provided in three formats to support different use cases:
### 1. Raw Format (`raw/`)
Preserves all original Omniverse data:
- RGB images (PNG)
- 2D bounding boxes (NumPy `.npy`)
- Class labels (JSON)
- 3D primitive paths (JSON)
- Instance segmentation masks (PNG)
- Instance ID mappings (JSON)
- Depth/distance maps (NumPy `.npy`)
**Use for:** Multi-task learning, depth estimation, 3D tasks, research
### 2. YOLO Format (`yolo/`)
Standard YOLO v5/v8 format:
- Images in `images/train/`, `images/val/`, `images/test/`
- Labels in `labels/train/`, `labels/val/`, `labels/test/`
- Each label file contains: `class_id x_center y_center width height` (normalized 0-1)
**Use for:** Object detection with YOLO models (Ultralytics, Darknet)
### 3. COCO Format (`coco/`)
COCO JSON format:
- Images in `images/train/`, `images/val/`, `images/test/`
- Annotations in `annotations/instances_{train,val,test}.json`
**Use for:** Object detection/segmentation with detectron2, MMDetection, etc.
## Directory Structure
```
warehouse_detection_dataset/
├── README.md # This file
├── data.yaml # YOLO configuration
├── dataset_info.json # HuggingFace metadata
├── class_mapping.json # Complete class information
├── raw/ # Original Omniverse format
│ ├── train/
│ │ ├── images/
│ │ ├── annotations/
│ │ ├── segmentation/
│ │ └── depth/
│ ├── val/
│ └── test/
├── yolo/ # YOLO format
│ ├── images/
│ │ ├── train/
│ │ ├── val/
│ │ └── test/
│ └── labels/
│ ├── train/
│ ├── val/
│ └── test/
└── coco/ # COCO format
├── images/
│ ├── train/
│ ├── val/
│ └── test/
└── annotations/
├── instances_train.json
├── instances_val.json
└── instances_test.json
```
## Usage Examples
### YOLO (Ultralytics)
```python
from ultralytics import YOLO
# Train a model
model = YOLO('yolov8n.pt')
model.train(data='warehouse_detection_dataset/data.yaml', epochs=100)
# Validate
metrics = model.val()
# Predict
results = model('path/to/image.jpg')
```
### COCO (detectron2)
```python
from detectron2.data import DatasetCatalog, MetadataCatalog
from detectron2.data.datasets import load_coco_json
# Register dataset
DatasetCatalog.register(
"warehouse_train",
lambda: load_coco_json(
"warehouse_detection_dataset/coco/annotations/instances_train.json",
"warehouse_detection_dataset/coco/images/train"
)
)
# Train your model
# ... (standard detectron2 training code)
```
### Raw Format (Custom)
```python
import numpy as np
import json
from PIL import Image
# Load image
img = Image.open('raw/train/images/warehouse_000001.png')
# Load bounding boxes
bboxes = np.load('raw/train/annotations/warehouse_000001_bbox.npy')
# Load labels
with open('raw/train/annotations/warehouse_000001_labels.json') as f:
labels = json.load(f)
# Load segmentation mask
seg_mask = Image.open('raw/train/segmentation/warehouse_000001_seg.png')
# Load depth map
depth = np.load('raw/train/depth/warehouse_000001_depth.npy')
```
## Citiation
If you use this dataset in your research, please cite:
```bibtex
@dataset{warehouse_detection_2025,
title={Warehouse Object Detection Dataset},
author={Howe, McCarthy and Phillips, Cassandra and Lee, Alfred and Sethi, Varun and Hall, Jada},
year={2025},
publisher={Clemson University Capstone},
note={In partnership with Capgemini Supply Chain},
version={1.0.0},
license={CC-BY-4.0}
}
```
## License
This dataset is released under the CC-BY-4.0 license.
## Team & Acknowledgments
**Project Context:**
This dataset was created as part of the **Clemson University Capstone 2025** project, in partnership with **Capgemini Supply Chain**.
**Contributors:**
- **McCarthy Howe** (mac@machowe.com)
- **Cassandra Phillips**
- **Alfred Lee**
- **Varun Sethi**
- **Jada Hall**
**Tooling:**
Generated using NVIDIA Omniverse Replicator.
## Contact
For questions, issues, or feedback, please contact **McCarthy Howe** at [mac@machowe.com](mailto:mac@machowe.com).
|