dataset / README.md
cmolinac's picture
Update README.md
5fe3a6c verified
---
language:
- en
license:
- cc-by-4.0
tags:
- semantic-segmentation
- scene-understanding
- 3d-point-clouds
- multimodal
- drone-imagery
pretty_name: NEST3D - Sociable Weaver Nest 3D Dataset
---
# NEST3D: A High-Resolution Multimodal Dataset of Sociable Weaver Tree Nests
![NEST3D Workflow](./workflow.png)
## Dataset Description
NEST3D is a multimodal dataset of 104 sociable weaver nests, combining drone-based RGB and multispectral imagery with a semantically annotated 3D RGB point cloud. It captures trees hosting these nests through drone-based remote sensing, providing rich spatial and spectral information to benchmark and advance scene-level semantic segmentation methods for computer vision and ecological monitoring applications.
### Key Characteristics
- **Modality**: Multimodal (RGB imagery, multispectral bands, 3D point clouds)
- **Task**: Scene-level semantic segmentation
- **Scale**: Multiple tree-nest scenes with consistent spatial and spectral coverage
- **Annotation**: Point-level semantic labels for 3D point clouds
- **Data Source**: Drone-based RGB and multispectral imagery
- **Application Domain**: Ecological monitoring, wildlife management, 3d semantic segmenation, 3d reconstruction.
## Dataset Organization
The dataset is organized into modality-specific directories to support flexible access and reuse:
### Directory Structure
```
NEST3D/
├── train/
│ ├── sample_001/
│ │ ├── RGB/ # RGB drone images
│ │ │ ├── sample001_RGB_001.JPG
│ │ │ └── ...
│ │ ├── MS/ # Multispectral imagery
│ │ │ ├── Green/
│ │ │ │ ├── sample001_G_001.TIF
│ │ │ │ └── ...
│ │ │ ├── Red/
│ │ │ │ ├── sample001_R_001.TIF
│ │ │ │ └── ...
│ │ │ ├── Red_Edge/
│ │ │ │ ├── sample001_RE_001.TIF
│ │ │ │ └── ...
│ │ │ └── NIR/
│ │ │ ├── sample001_NIR_001.TIF
│ │ │ └── ...
│ │ └── sample001.npy # 3D point cloud with labels
│ └── sample_002/
│ └── ...
└── test/
├── sample_084/
│ ├── RGB/
│ ├── MS/
│ └── sample084.npy
└── ...
```
### Data Modalities
#### 1. **RGB Imagery**
- Raw drone images from aerial acquisition
- Format: JPEG
- Organized by data split and scene identifier
- Example path: `train/sample_001/RGB/sample001_RGB_119.JPG`
#### 2. **Multispectral Imagery**
- Four spectral bands from the same acquisitions as RGB
- Organized into four band-specific folders:
- **Green (G)**: Green channel imagery
- **Red (R)**: Red channel imagery
- **Red Edge (RE)**: Red Edge channel for vegetation analysis
- **NIR**: Near-Infrared channel for vegetation health assessment
- Format: GeoTIFF (.TIF)
- Example paths:
- `train/sample_001/MS/Green/sample001_G_119.TIF`
- `train/sample_001/MS/Red/sample001_R_119.TIF`
- `train/sample_001/MS/Red_Edge/sample001_RE_119.TIF`
- `train/sample_001/MS/NIR/sample001_NIR_119.TIF`
#### 3. **3D Point Clouds**
- One NumPy file per scene containing the complete 3D reconstruction
- Format: `.npy` (NumPy binary format)
- Per-point attributes: `[x, y, z, r, g, b, label]`
- **x, y, z**: 3D spatial coordinates (meters)
- **r, g, b**: RGB color values (0-255)
- **label**: Semantic class label (integer)
- Example path: `train/sample_001/sample001.npy`
## Data Splits
The dataset is divided into fixed training and test sets:
- **Training Set**: Used for model training and development
- **Test Set**: Reserved for model evaluation and benchmarking
Each split contains a consistent collection of scenes to ensure reliable evaluation.
## Usage
### Loading 3D Point Clouds
```python
import numpy as np
# Load point cloud with semantic labels
point_cloud = np.load('train/sample_001/sample001.npy')
# Extract coordinates
xyz = point_cloud[:, :3]
# Extract colors
rgb = point_cloud[:, 3:6]
# Extract semantic labels
labels = point_cloud[:, 6]
```
### Loading Multispectral Imagery
```python
from PIL import Image
import numpy as np
# Load a single band
green_band = np.array(Image.open('train/sample_001/MS/Green/sample001_G_001.TIF'))
# Load all four bands for a given image
green = np.array(Image.open('train/sample_001/MS/Green/sample001_G_001.TIF'))
red = np.array(Image.open('train/sample_001/MS/Red/sample001_R_001.TIF'))
red_edge = np.array(Image.open('train/sample_001/MS/Red_Edge/sample001_RE_001.TIF'))
nir = np.array(Image.open('train/sample_001/MS/NIR/sample001_NIR_001.TIF'))
# Stack into multiband image
multispectral = np.stack([green, red, red_edge, nir], axis=-1)
```
### Using with Hugging Face Datasets Library
```python
from datasets import load_dataset
# Load the dataset
dataset = load_dataset('NEST3D/dataset')
```
## Downloading the Dataset
### Option 1: Using Hugging Face Hub
```bash
pip install huggingface_hub
huggingface-cli download NEST3D/dataset --repo-type dataset --local-dir ./NEST3D
```
## Dataset Information
- **Total Size**:
- **Number of Scenes**: 104 samples. Split into train/test as 83/21
- **Modalities**: RGB, Multispectral (4 bands), 3D Point Clouds
- **Image Format**: JPEG (RGB), GeoTIFF (Multispectral)
- **Point Cloud Format**: NumPy arrays
- **Annotation Type**: Per-point semantic labels
## Acknowledgments
This work was funded by:
- **European Union's Horizon Europe** research and innovation programme through the Marie Skłodowska-Curie project **"WildDrone – Autonomous Drones for Nature Conservation"** (grant agreement no. 101071224)
- **EPSRC-funded** "Autonomous Drones for Nature Conservation Missions" grant (EP/X029077/1)
- **Swiss State Secretariat for Education, Research and Innovation (SERI)** under contract number 22.00280
We extend our gratitude to our collaborators and field partners in Namibia for their invaluable support during data collection.
## Contact & Support
For questions, issues, or contributions, please visit the [dataset discussion forum](https://huggingface.co/datasets/NEST3D/dataset/discussions).
**Last Updated**: February 2026
**Dataset Version**: 1.0