File size: 11,357 Bytes
ac32ead ddc02a1 ac32ead 93d7341 1ecbab6 bbae260 1ecbab6 6298310 1ecbab6 6298310 1ecbab6 47fcf87 1ecbab6 6298310 1ecbab6 6298310 1ecbab6 6298310 1ecbab6 1b368d6 1ecbab6 6298310 1ecbab6 1b368d6 6298310 1ecbab6 6298310 1ecbab6 6298310 1ecbab6 6298310 1b368d6 6298310 1b368d6 6298310 1b368d6 6298310 1b368d6 6298310 1b368d6 6298310 1b368d6 6298310 1ecbab6 7d95c17 f913011 1ecbab6 2c12356 1ecbab6 6298310 1ecbab6 9ef39a5 1ecbab6 9ef39a5 402cf62 9ef39a5 1ecbab6 402cf62 1ecbab6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 |
---
license: cc0-1.0
task_categories:
- video-classification
- object-detection
tags:
- baboons
- video
- animal behavior
- behavior recognition
- annotation
- annotated video
- conservation
- drone
- UAV
- imbalanced
- Kenya
- Mpala Research Centre
pretty_name: >-
BaboonLand Dataset: Tracking Primates in the Wild and Automating Behaviour
Recognition from Drone Videos
description: >-
BaboonLand is an aerial drone video dataset of wild olive baboons in Laikipia,
Kenya, collected over 21 consecutive days across three troops. It contains 20+
hours of footage with dense multi-individual scenes (up to ~70 baboons per
frame) and annotations enabling detection, multi-object tracking, and behavior
recognition.
size_categories:
- 1M<n<10M
language:
- en
configs:
- config_name: default
data_files:
- split: train
path: "Dataset Viewer/train.csv"
- split: validation
path: "Dataset Viewer/val.csv"
---
# Dataset Card for BaboonLand Dataset: Tracking Primates in the Wild and Automating Behaviour Recognition from Drone Videos
## Dataset Description
- **Homepage:** [BaboonLand Site](https://baboonland.xyz/)
- **Repository:** https://huggingface.co/datasets/imageomics/BaboonLand
- **Paper:** [https://link.springer.com/article/10.1007/s11263-025-02493-5](https://link.springer.com/article/10.1007/s11263-025-02493-5)
- **arXiv:** [https://arxiv.org/pdf/2405.17698](https://arxiv.org/pdf/2405.17698)
### Dataset Summary
BaboonLand is an aerial drone video dataset of wild olive baboons (*Papio anubis*) collected over 21 consecutive days in Laikipia (Mpala Research Centre), Kenya, following three troops during morning and evening movements to and from sleeping sites. The dataset contains UAV footage across diverse environments (e.g., sleeping tree, river, rock, open savannah, cliff), with up to ~70 individuals per frame, yielding dense multi-object scenes from an overhead viewpoint.
The dataset supports three core subtasks: detection, multi-object tracking, and behavior recognition. It includes (1) a detection dataset derived from ~5.3K-resolution frames via multi-scale tiling (≈30K images), (2) ~0.5 hours of dense tracking annotations, and (3) ~20 hours of behavior “mini-scenes” annotated into 12 behavior classes and additional category for occlusions.
### Download & Reconstruct
BaboonLand is stored as split ZIP archives (`*.zip.part.*`) tracked with **Git LFS**. You can either download everything at once, or pull only a specific subset (Charades / Dataset / Tracking), then **concatenate parts**.
**Integrity check:** Compare the printed `md5sum` values with the reference hashes in `BaboonLand/manifest.json`.
---
### Option A — Download everything (all parts)
```bash
git clone https://huggingface.co/datasets/imageomics/BaboonLand
cd BaboonLand
cat BaboonLand/charades.zip.part* > BaboonLand/charades.zip
cat BaboonLand/dataset.zip.part* > BaboonLand/dataset.zip
cat BaboonLand/tracking.zip.part* > BaboonLand/tracking.zip
md5sum BaboonLand/charades.zip
md5sum BaboonLand/dataset.zip
md5sum BaboonLand/tracking.zip
rm -rf BaboonLand/charades.zip.part*
rm -rf BaboonLand/dataset.zip.part*
rm -rf BaboonLand/tracking.zip.part*
```
---
### Option B — Download only Charades part
```bash
GIT_LFS_SKIP_SMUDGE=1 git clone https://huggingface.co/datasets/imageomics/BaboonLand
cd BaboonLand
git lfs pull --include="BaboonLand/charades.zip.part*"
cat BaboonLand/charades.zip.part* > BaboonLand/charades.zip
md5sum BaboonLand/charades.zip
rm -rf BaboonLand/charades.zip.part*
```
---
### Option C — Download only Dataset part
```bash
GIT_LFS_SKIP_SMUDGE=1 git clone https://huggingface.co/datasets/imageomics/BaboonLand
cd BaboonLand
git lfs pull --include="BaboonLand/dataset.zip.part*"
cat BaboonLand/dataset.zip.part* > BaboonLand/dataset.zip
md5sum BaboonLand/dataset.zip
rm -rf BaboonLand/dataset.zip.part*
```
---
### Option D — Download only Tracking part
```bash
GIT_LFS_SKIP_SMUDGE=1 git clone https://huggingface.co/datasets/imageomics/BaboonLand
cd BaboonLand
git lfs pull --include="BaboonLand/tracking.zip.part*"
cat BaboonLand/tracking.zip.part* > BaboonLand/tracking.zip
md5sum BaboonLand/tracking.zip
rm -rf BaboonLand/tracking.zip.part*
```
---
### Option E — Download only CVAT templates
```bash
GIT_LFS_SKIP_SMUDGE=1 git clone https://huggingface.co/datasets/imageomics/BaboonLand
cd BaboonLand
git lfs pull --include="BaboonLand/cvat_templates"
```
---
### Directory Layout
```text
BaboonLand
/charades -> The dataset converted to Charades format to train and evaluate behavior
recognition models. You can download the generated dataset from our webpage
or you can generate it yourself.
...
/cvat_templates -> Templates to backup projects in CVAT and explore/adjust annotations.
/behavior.zip
/tracking.zip
/dataset
/video_1
/actions
/0.xml
/1.xml -> Behavior annotations for individual with ID=1
...
/n.xml
/mini-scenes -> Mini-scenes generated from video.mp4 and tracks.xml
/0.mp4
/1.mp4
...
/n.mp4
/timeline.jpg
/tracks.xml -> Tracks + bounding boxes (CVAT for video 1.1). Each track has a unique ID.
/video.mp4 -> Original drone video
/video_2
...
/video_n
...
/scripts
/requirements.txt
/tracks2mini-scenes.py
/dataset2charades.py
/charades2video.py
/charades2visual.py
/dataset2tracking.py
/tracking2ultralytics.py
/ultralytics2pyramid.py
/tracking -> Tracking split + (optionally) Ultralytics-format detection data.
...
/README.md
```
### Supported Tasks and Leaderboards
#### Detection
We evaluate YOLOv8-X model with input resolution of 768x768 on our dataset and report mAP@50, Precision, and Recall:
| Model | mAP@50 | Precision | Recall |
| --- | ---: | ---: | ---: |
| YOLOv8-X | 92.62 | 93.70 | 87.60 |
#### Tracking
We evaluate SORT, DeepSORT, StrongSORT, ByteTrack, and BotSort tracking algorithms on our dataset and report MOTA, MOTP, IDF1, Precision, and Recall:
| Tracker | MOTA | MOTP | IDF1 | Precision | Recall |
| --- | ---: | ---: | ---: | ---: | ---: |
| SORT | 84.76 | 50.15 | 77.43 | 90.83 | 91.19 |
| DeepSORT | 84.40 | 87.22 | 81.38 | 90.26 | 91.57 |
| StrongSORT | 82.48 | 85.37 | 84.98 | 88.00 | 90.10 |
| ByteTrack | 63.55 | 34.10 | 77.01 | 96.32 | 64.90 |
| BotSort | 63.81 | 34.31 | 78.24 | 97.21 | 66.16 |
#### Behavior Classes:
- Walking/Running
- Sitting/Standing
- Fighting/Playing
- Self-Grooming
- Being Groomed
- Grooming Somebody
- Mutual Grooming
- Infant-Carrying
- Foraging
- Drinking
- Mounting
- Sleeping
- Occluded
#### Behavior Recognition
We evaluate I3D, SlowFast, and X3D models on our dataset and report Micro-Average (Per Instance) and Macro-Average (Per Class) accuracy.
| Method | WI | Micro Top-1 | Micro Top-3 | Micro Top-5 | Macro Top-1 | Macro Top-3 | Macro Top-5 |
| --- | --- | ---: | ---: | ---: | ---: | ---: | ---: |
| I3D | Random | 61.29 | 89.38 | 92.34 | 26.53 | 54.51 | 65.47 |
| SlowFast | Random | 61.71 | 90.35 | 93.11 | 27.08 | 56.73 | 67.61 |
| X3D | Random | 63.97 | 91.34 | 95.17 | 30.04 | 60.58 | 72.13 |
| X3D | K-400 | 64.89 | 92.54 | 96.66 | 31.41 | 62.04 | 74.01 |
### Languages
English
## Dataset Structure
BaboonLand provides original videos, CVAT-formatted annotations, derived mini-scenes, and scripts to generate task-specific training formats (e.g., Ultralytics/YOLO and Charades for SlowFast).
### Data Instances
Each `dataset/video_k/` directory contains:
- `video.mp4`: original UAV video
- `tracks.xml`: per-frame tracks (IDs + bounding boxes)
- `actions/*.xml`: per-track behavior labels (filename matches track ID)
- `mini-scenes/*.mp4`: cropped clips centered on each tracked individual (filename matches track ID)
### Data Fields
BaboonLand supports three derived tasks:
- **Detection:** bounding boxes for baboons (also convertible to Ultralytics/YOLO format via provided scripts).
- **Tracking:** per-frame tracks with persistent IDs and bounding boxes (stored in simplified CVAT for video 1.1).
- **Behavior recognition:** per-individual **mini-scenes** (cropped clips centered on each tracked individual) labeled into **12 behavior classes + Occluded**.
### Data Splits
BaboonLand includes task-specific evaluation sets:
- **Tracking:** 75% of each video for training, 25% for testing.
- **Detection (YOLO-formatted):** 80% training, 7% validation, 13% testing.
- **Behavior recognition (Charades format):** 75% training, 25% testing.
#### Data Collection and Procedures
- **Species:** Olive baboons (*Papio anubis*)
- **Location:** Mpala Research Centre, Laikipia County, Kenya
- **Capture:** DJI Air 2S, videos recorded in **5.3K**
- **Procedure:** all flights were conducted above 20 meters from a closes animal.
## Personal and Sensitive Information
- No humans can be distinguished in the videos.
- Data collection followed research licensing and animal care protocols (see Acknowledgments).
### Authors
- Isla Duporge
- Maksim Kholiavchenko
- Roi Harel
- Scott Wolf
- Daniel Rubenstein
- Meg Crofoot
- Tanya Berger-Wolf
- Stephen Lee
- Julie Barreau
- Jenna Kline
- Michelle Ramirez
- Charles Stewart
### Citation Information
#### Dataset
####
```
@misc{hdr_imageomics_institute_2026,
author = { Isla Duporge and Maksim Kholiavchenko and Roi Harel and Scott Wolf and Daniel Rubenstein and Tanya Berger-Wolf and Margaret Crofoot and Stephen Lee and Julie Barreau and Jenna Kline and Michelle Ramirez and Charles Stewart },
title = { BaboonLand Dataset: Tracking Primates in the Wild and Automating Behaviour Recognition from Drone Videos },
year = 2026,
url = { https://huggingface.co/datasets/imageomics/BaboonLand },
doi = { 10.57967/hf/7470 },
publisher = { Hugging Face }
}
```
#### Paper
```
@article{duporge2025baboonland,
title={BaboonLand Dataset: Tracking Primates in the Wild and Automating Behaviour Recognition from Drone Videos},
author={Duporge, Isla and Kholiavchenko, Maksim and Harel, Roi and Wolf, Scott and Rubenstein, Daniel I and Crofoot, Margaret C and Berger-Wolf, Tanya and Lee, Stephen J and Barreau, Julie and Kline, Jenna and others},
journal={International Journal of Computer Vision},
pages={1--12},
year={2025},
publisher={Springer}
}
```
### Contributions / Acknowledgments
This material is based upon work supported by the National Science Foundation under Award No. 2118240 and Award No. 2112606. ID was supported by the National Academy of Sciences Research Associate Program and the United States Army Research Laboratory while conducting this study. ID collected all the UAV data on a Civil Aviation Authority Drone License CAA NQE Approval Number: 0216/1365 in conjunction with authorization from a KCAA operator under a Remote Pilot License. The data was gathered at the Mpala Research Centre in Kenya, in accordance with Research License No. NACOSTI/P/22/18214. The data collection protocol adhered strictly to the guidelines set forth by the Institutional Animal Care and Use Committee under permission No. IACUC 1835F. |