| # Warehouse Object Detection Dataset | |
| ## Overview | |
| This is a synthetic warehouse object detection dataset generated using NVIDIA Omniverse Replicator. The dataset contains high-quality RGB images with comprehensive annotations including 2D bounding boxes, instance segmentation masks, depth maps, and 3D primitive paths. | |
| **Version:** 1.0.0 | |
| **License:** CC-BY-4.0 | |
| **Total Images:** 6,234 | |
| **Total Annotations:** 78,441 | |
| **Image Resolution:** 512x512 | |
| **Number of Classes:** 25 | |
| ## Dataset Statistics | |
| ### Split Distribution | |
| - **Training Set:** 4,363 images (70.0%) | |
| - **Validation Set:** 935 images (15.0%) | |
| - **Test Set:** 936 images (15.0%) | |
| ### Annotation Statistics | |
| - **Average Annotations per Image:** 12.6 | |
| - **Total Bounding Boxes:** 78,441 | |
| ### Top 10 Most Common Classes | |
| 1. **wall**: 6,234 instances in 6,234 images | |
| 2. **floor**: 6,232 instances in 6,232 images | |
| 3. **sign**: 5,945 instances in 5,945 images | |
| 4. **floor_decal**: 5,945 instances in 5,945 images | |
| 5. **pillar**: 5,845 instances in 5,845 images | |
| 6. **rack**: 5,824 instances in 5,824 images | |
| 7. **box**: 5,789 instances in 5,789 images | |
| 8. **pallet**: 5,743 instances in 5,743 images | |
| 9. **bracket**: 4,857 instances in 4,857 images | |
| 10. **lamp**: 4,451 instances in 4,451 images | |
| ## Classes | |
| The dataset includes 25 object classes organized into the following categories: | |
| ### Container | |
| - **box**: 5,789 annotations | |
| - **crate**: 1,270 annotations | |
| - **barrel**: 1,229 annotations | |
| - **bottle**: 665 annotations | |
| - **bucket**: 457 annotations | |
| ### Storage | |
| - **pallet**: 5,743 annotations | |
| - **rack**: 5,824 annotations | |
| ### Infrastructure | |
| - **bracket**: 4,857 annotations | |
| - **pillar**: 5,845 annotations | |
| - **emergency_board**: 569 annotations | |
| ### Equipment | |
| - **lamp**: 4,451 annotations | |
| - **sign**: 5,945 annotations | |
| - **wire**: 3,985 annotations | |
| - **fuse_box**: 1,154 annotations | |
| - **fire_extinguisher**: 3,304 annotations | |
| - **forklift**: 343 annotations | |
| - **cart**: 427 annotations | |
| - **cone**: 639 annotations | |
| ### Markers | |
| - **floor_decal**: 5,945 annotations | |
| - **barcode**: 1,137 annotations | |
| - **paper_note**: 2,125 annotations | |
| - **paper_shortcut**: 390 annotations | |
| ### Background | |
| - **wall**: 6,234 annotations | |
| - **ceiling**: 3,882 annotations | |
| - **floor**: 6,232 annotations | |
| ## Dataset Formats | |
| This dataset is provided in three formats to support different use cases: | |
| ### 1. Raw Format (`raw/`) | |
| Preserves all original Omniverse data: | |
| - RGB images (PNG) | |
| - 2D bounding boxes (NumPy `.npy`) | |
| - Class labels (JSON) | |
| - 3D primitive paths (JSON) | |
| - Instance segmentation masks (PNG) | |
| - Instance ID mappings (JSON) | |
| - Depth/distance maps (NumPy `.npy`) | |
| **Use for:** Multi-task learning, depth estimation, 3D tasks, research | |
| ### 2. YOLO Format (`yolo/`) | |
| Standard YOLO v5/v8 format: | |
| - Images in `images/train/`, `images/val/`, `images/test/` | |
| - Labels in `labels/train/`, `labels/val/`, `labels/test/` | |
| - Each label file contains: `class_id x_center y_center width height` (normalized 0-1) | |
| **Use for:** Object detection with YOLO models (Ultralytics, Darknet) | |
| ### 3. COCO Format (`coco/`) | |
| COCO JSON format: | |
| - Images in `images/train/`, `images/val/`, `images/test/` | |
| - Annotations in `annotations/instances_{train,val,test}.json` | |
| **Use for:** Object detection/segmentation with detectron2, MMDetection, etc. | |
| ## Directory Structure | |
| ``` | |
| warehouse_detection_dataset/ | |
| ├── README.md # This file | |
| ├── data.yaml # YOLO configuration | |
| ├── dataset_info.json # HuggingFace metadata | |
| ├── class_mapping.json # Complete class information | |
| ├── raw/ # Original Omniverse format | |
| │ ├── train/ | |
| │ │ ├── images/ | |
| │ │ ├── annotations/ | |
| │ │ ├── segmentation/ | |
| │ │ └── depth/ | |
| │ ├── val/ | |
| │ └── test/ | |
| ├── yolo/ # YOLO format | |
| │ ├── images/ | |
| │ │ ├── train/ | |
| │ │ ├── val/ | |
| │ │ └── test/ | |
| │ └── labels/ | |
| │ ├── train/ | |
| │ ├── val/ | |
| │ └── test/ | |
| └── coco/ # COCO format | |
| ├── images/ | |
| │ ├── train/ | |
| │ ├── val/ | |
| │ └── test/ | |
| └── annotations/ | |
| ├── instances_train.json | |
| ├── instances_val.json | |
| └── instances_test.json | |
| ``` | |
| ## Usage Examples | |
| ### YOLO (Ultralytics) | |
| ```python | |
| from ultralytics import YOLO | |
| # Train a model | |
| model = YOLO('yolov8n.pt') | |
| model.train(data='warehouse_detection_dataset/data.yaml', epochs=100) | |
| # Validate | |
| metrics = model.val() | |
| # Predict | |
| results = model('path/to/image.jpg') | |
| ``` | |
| ### COCO (detectron2) | |
| ```python | |
| from detectron2.data import DatasetCatalog, MetadataCatalog | |
| from detectron2.data.datasets import load_coco_json | |
| # Register dataset | |
| DatasetCatalog.register( | |
| "warehouse_train", | |
| lambda: load_coco_json( | |
| "warehouse_detection_dataset/coco/annotations/instances_train.json", | |
| "warehouse_detection_dataset/coco/images/train" | |
| ) | |
| ) | |
| # Train your model | |
| # ... (standard detectron2 training code) | |
| ``` | |
| ### Raw Format (Custom) | |
| ```python | |
| import numpy as np | |
| import json | |
| from PIL import Image | |
| # Load image | |
| img = Image.open('raw/train/images/warehouse_000001.png') | |
| # Load bounding boxes | |
| bboxes = np.load('raw/train/annotations/warehouse_000001_bbox.npy') | |
| # Load labels | |
| with open('raw/train/annotations/warehouse_000001_labels.json') as f: | |
| labels = json.load(f) | |
| # Load segmentation mask | |
| seg_mask = Image.open('raw/train/segmentation/warehouse_000001_seg.png') | |
| # Load depth map | |
| depth = np.load('raw/train/depth/warehouse_000001_depth.npy') | |
| ``` | |
| ## Citiation | |
| If you use this dataset in your research, please cite: | |
| ```bibtex | |
| @dataset{warehouse_detection_2025, | |
| title={Warehouse Object Detection Dataset}, | |
| author={Howe, McCarthy and Phillips, Cassandra and Lee, Alfred and Sethi, Varun and Hall, Jada}, | |
| year={2025}, | |
| publisher={Clemson University Capstone}, | |
| note={In partnership with Capgemini Supply Chain}, | |
| version={1.0.0}, | |
| license={CC-BY-4.0} | |
| } | |
| ``` | |
| ## License | |
| This dataset is released under the CC-BY-4.0 license. | |
| ## Team & Acknowledgments | |
| **Project Context:** | |
| This dataset was created as part of the **Clemson University Capstone 2025** project, in partnership with **Capgemini Supply Chain**. | |
| **Contributors:** | |
| - **McCarthy Howe** (mac@machowe.com) | |
| - **Cassandra Phillips** | |
| - **Alfred Lee** | |
| - **Varun Sethi** | |
| - **Jada Hall** | |
| **Tooling:** | |
| Generated using NVIDIA Omniverse Replicator. | |
| ## Contact | |
| For questions, issues, or feedback, please contact **McCarthy Howe** at [mac@machowe.com](mailto:mac@machowe.com). | |