| Information |
| =========== |
|
|
| This is the YCB-Video dataset [1] for 6D object pose estimation. |
| It provides accurate 6D poses of 21 objects from the YCB dataset [2] |
| observed in 92 videos with 133,827 frames. |
|
|
| This project was funded in part by Siemens and by NSF |
| STTR grant 63-5197 with Lula Robotics. |
|
|
|
|
| How to Cite |
| =========== |
|
|
| @article{xiang2017posecnn, |
| author = {Xiang, Yu and Schmidt, Tanner and Narayanan, Venkatraman and Fox, Dieter}, |
| title = {PoseCNN: A Convolutional Neural Network for 6D Object Pose Estimation in Cluttered Scenes}, |
| journal = {arXiv preprint arXiv:1711.00199}, |
| year = {2017} |
| } |
|
|
|
|
| Basic Usage |
| =========== |
|
|
| The description of the directories in this package: |
|
|
| 1. cameras. The camera parameters used to capture the videos. asus-uw.json for video 0000 ~ 0059, asus-cmu.json for video 0060 ~ 0091. |
| 2. data. The 92 videos in the dataset. |
| 3. data_syn. 80,000 synthetic images of the 21 YCB objects. |
| 4. image_sets. Separation of the videos into training set (train.txt) and the testing set (val.txt, keyframe.txt). |
| 5. keyframes. Keyframe indexes of the 12 testing videos. |
| 6. models. 3D models of the 21 YCB objects. |
| 7. pairs. Stereo pair indexes of the 12 testing videos. |
| 8. poses. All the 6D poses of the 21 YCB objects in the dataset (quaternion + translation). |
|
|
| References |
| ========== |
|
|
| [1] Y. Xiang, T. Schmidt, V. Narayanan and D. Fox. |
| PoseCNN: A convolutional neural network for 6D object pose estimation in cluttered scenes. In arXiv:1711.00199, 2017. |
|
|
| [2] B. Calli, A. Singh, A. Walsman, S. Srinivasa, P. Abbeel, and A. M. Dollar. The YCB object and model set: Towards common benchmarks |
| for manipulation research. In International Conference on Advanced Robotics (ICAR), pp. 510–517, 2015. |
|
|