File size: 2,963 Bytes
4ad3244 1204d31 4ad3244 1204d31 4ad3244 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 |
---
license: cc-by-nc-sa-4.0
language:
- en
- de
- vi
- gsw
tags:
- video
- audio
- speech
- image
- 6dof
size_categories:
- n>1T
pretty_name: 'The CASTLE 2024 Dataset: Advancing the Art of Multimodal Understanding'
---
# What is CASTLE?

The CASTLE dataset is a large-scale, multimodal dataset designed for advancing research in lifelogging, human activity recognition, and multimodal retrieval. It provides a rich collection of time-aligned sensor and video data for analysis and benchmarking. See the [Paper](https://doi.org/10.1145/3746027.3758199) (or its [arXiv pre-print](https://arxiv.org/abs/2503.17116)) for more details.
You can check our [website](https://castle-dataset.github.io/) for more details.
## Characteristics
* Captured over **four days** in a controlled environment
* **10 participants** engaged in natural activities
* **15 video streams** (10 egocentric, 5 static perspectives)
* Over **600 hours** of UHD 50fps video with audio
* Includes **6DoF IMU, GPS, and biometric data**
* **8.22TB** total size
## License
The CASTLE dataset is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
## Terms of Use
By downloading the dataset, you agree to the following terms:
* The dataset is provided for research purposes only.
* You will not use the dataset for any commercial purposes.
* You will not distribute the dataset or any derivative works to others.
* You will provide appropriate credit to the dataset authors in your publications.
If you are using the dataset in your research, please consider citing the following paper:
```bibtex
@inproceedings{10.1145/3746027.3758199,
author = {Rossetto, Luca and Bailer, Werner and Dang-Nguyen, Duc-Tien and Healy, Graham and J\'{o}nsson, Bj\"{o}rn \TH{}\'{o}r and Kongmeesub, Onanong and Le, Hoang-Bao and Rudinac, Stevan and Sch\"{o}ffmann, Klaus and Spiess, Florian and Tran, Allie and Tran, Minh-Triet and Tran, Quang-Linh and Gurrin, Cathal},
title = {The CASTLE 2024 Dataset: Advancing the Art of Multimodal Understanding},
year = {2025},
isbn = {9798400720352},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3746027.3758199},
doi = {10.1145/3746027.3758199},
booktitle = {Proceedings of the 33rd ACM International Conference on Multimedia},
pages = {12629–12635},
numpages = {7},
keywords = {dataset, egocentric vision, lifelogging, multi-perspective video, multimodal understanding},
location = {Dublin, Ireland},
series = {MM '25}
}
```
# Challenges
We are organizing a series of challenges to encourage the research community to explore and utilize the CASTLE dataset.
To see the list of challenges and their details, please visit the [Challenges](https://castle-dataset.github.io/challenges/) page on the website. |