RobotP: A Benchmark Dataset for 6D Object Pose Estimation
- PMID: 33670325
- PMCID: PMC7917891
- DOI: 10.3390/s21041299
RobotP: A Benchmark Dataset for 6D Object Pose Estimation
Abstract
Deep learning has achieved great success on robotic vision tasks. However, when compared with other vision-based tasks, it is difficult to collect a representative and sufficiently large training set for six-dimensional (6D) object pose estimation, due to the inherent difficulty of data collection. In this paper, we propose the RobotP dataset consisting of commonly used objects for benchmarking in 6D object pose estimation. To create the dataset, we apply a 3D reconstruction pipeline to produce high-quality depth images, ground truth poses, and 3D models for well-selected objects. Subsequently, based on the generated data, we produce object segmentation masks and two-dimensional (2D) bounding boxes automatically. To further enrich the data, we synthesize a large number of photo-realistic color-and-depth image pairs with ground truth 6D poses. Our dataset is freely distributed to research groups by the Shape Retrieval Challenge benchmark on 6D pose estimation. Based on our benchmark, different learning-based approaches are trained and tested by the unified dataset. The evaluation results indicate that there is considerable room for improvement in 6D object pose estimation, particularly for objects with dark colors, and photo-realistic images are helpful in increasing the performance of pose estimation algorithms.
Keywords: 3D reconstruction; 6D pose estimation; benchmark dataset; sensors.
Conflict of interest statement
The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.
Figures
References
-
- Wang C., Xu D., Zhu Y., Martín-Martín R., Lu C., Fei-Fei L., Savarese S. Densefusion: 6D object pose estimation by iterative dense fusion; Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Long Beach, CA, USA. 15–21 June 2019; pp. 3343–3352.
-
- Chen W., Duan J., Basevi H., Chang H.J., Leonardis A. Ponitposenet: Point pose network for robust 6D object pose estimation; Proceedings of the IEEE Winter Conference on Applications of Computer Vision; Snowmass Village, CO, USA. 1–5 March 2020; pp. 2824–2833.
-
- Tekin B., Sinha S.N., Fua P. Real-time seamless single shot 6D object pose prediction; Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Salt Lake City, UT, USA. 18–23 June 2018; pp. 292–301.
-
- Garcia-Garcia A., Martinez-Gonzalez P., Oprea S., Castro-Vargas J.A., Orts-Escolano S., Garcia-Rodriguez J., Jover-Alvarez A. The robotrix: An extremely photorealistic and very-large-scale indoor dataset of sequences with robot trajectories and interactions; Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS); Madrid, Spain. 1–5 October 2018; pp. 6790–6797.
-
- Lepetit V., Moreno-Noguer F., Fua P. Epnp: An accurate O(n) solution to the pnp problem. Int. J. Comput. Vis. 2009;81:155. doi: 10.1007/s11263-008-0152-6. - DOI
LinkOut - more resources
Full Text Sources
Other Literature Sources
