--- task_categories: - robotics tags: - human-robot-interaction - multi-agent - embodied-ai - reinforcement-learning dataset_info: features: - name: map_name dtype: string - name: trajectory_id dtype: int64 - name: steps_data dtype: string - name: num_steps dtype: int64 splits: - name: train num_bytes: 562424179 num_examples: 912 - name: evaluation num_bytes: 54684865 num_examples: 96 - name: all_data num_bytes: 617109044 num_examples: 1008 download_size: 261118576 dataset_size: 1234218088 configs: - config_name: default data_files: - split: train path: data/train-* - split: evaluation path: data/evaluation-* - split: all_data path: data/all_data-* --- # Moving Out: Physically-grounded Human-AI Collaboration Dataset This repository contains the human-human interaction dataset for the paper [Moving Out: Physically-grounded Human-AI Collaboration](https://huggingface.co/papers/2507.18623). The **Moving Out** benchmark focuses on human-AI collaboration in physical environments, specifically for embodied agents (e.g., robots). It introduces a new benchmark that resembles a wide range of collaboration modes affected by physical attributes and constraints, such as moving heavy items together and maintaining consistent actions to move a big item around a corner. The dataset includes human-human interaction data collected for two tasks, designed to evaluate models' abilities to adapt to diverse human behaviors and unseen physical attributes in continuous state-action spaces with constrained dynamics. **Project Page:** [https://live-robotics-uva.github.io/movingout_ai/](https://live-robotics-uva.github.io/movingout_ai/) **Code:** [https://github.com/live-robotics-uva/moving_out_ai](https://github.com/live-robotics-uva/moving_out_ai) ## Sample Usage You can visualize trajectories from this dataset using the `dataset_to_video.py` script provided in the associated GitHub repository. **Save trajectory as MP4:** ```bash python dataset_to_video.py -f ShuaKang/movingout_task2 -m HandOff -t 4 -v video ``` **Show trajectory in popup window:** ```bash python dataset_to_video.py -f ShuaKang/movingout_task2 -m HandOff -t 4 -v human ``` Use `-m` for map name and `-t` for trajectory ID. ## Citation If you use this dataset in your research, please cite the original paper: ```bibtex @article{, title={Moving Out: Physically-grounded Human-AI Collaboration}, author={Shua Kang, Wenxin Xia, Ziqi Li, Mingxing Yuan, Xufan Wu, Min Bai, Boyi Liu, David Abel, Wenhao Zhang}, journal={arXiv preprint arXiv:2507.18623}, year={2025}, url={https://huggingface.co/papers/2507.18623} } ```