Skip to content

A demo for understanding the structure and pre-processing of YCB-Video Dataset.

Notifications You must be signed in to change notification settings

ChengChen2020/YCB_demo

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Demo for YCB-Video Dataset Usage

Adapted from PoseCNN-PyTorch

Running the demo

  • run the following script

    ./experiments/scripts/ycb_video_test.sh
  • on hpc

    sbatch hpc.sh

Information

This is the YCB-Video dataset [1] for 6D object pose estimation. It provides accurate 6D poses of 21 objects from the YCB dataset [2] observed in 92 videos with 133,827 frames.

This project was funded in part by Siemens and by NSF STTR grant 63-5197 with Lula Robotics.

How to Cite

@article{xiang2017posecnn, author = {Xiang, Yu and Schmidt, Tanner and Narayanan, Venkatraman and Fox, Dieter}, title = {PoseCNN: A Convolutional Neural Network for 6D Object Pose Estimation in Cluttered Scenes}, journal = {arXiv preprint arXiv:1711.00199}, year = {2017} }

Basic Usage

The description of the directories in this package:

  1. cameras. The camera parameters used to capture the videos. asus-uw.json for video 0000 ~ 0059, asus-cmu.json for video 0060 ~ 0091.
  2. data. The 92 videos in the dataset.
  3. data_syn. 80,000 synthetic images of the 21 YCB objects.
  4. image_sets. Separation of the videos into training set (train.txt) and the testing set (val.txt, keyframe.txt).
  5. keyframes. Keyframe indexes of the 12 testing videos.
  6. models. 3D models of the 21 YCB objects.
  7. pairs. Stereo pair indexes of the 12 testing videos.
  8. poses. All the 6D poses of the 21 YCB objects in the dataset (quaternion + translation).

References

[1] Y. Xiang, T. Schmidt, V. Narayanan and D. Fox. PoseCNN: A convolutional neural network for 6D object pose estimation in cluttered scenes. In arXiv:1711.00199, 2017.

[2] B. Calli, A. Singh, A. Walsman, S. Srinivasa, P. Abbeel, and A. M. Dollar. The YCB object and model set: Towards common benchmarks for manipulation research. In International Conference on Advanced Robotics (ICAR), pp. 510–517, 2015.

About

A demo for understanding the structure and pre-processing of YCB-Video Dataset.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published