Abstract
Robot grasping and human-robot handover (HRH) tasks can significantly facilitate people’s production and life. In these tasks, robots need to obtain the real-time 3D position of the object, and the distance from the object to the camera plane is the critical information to get the object position. Currently, depth camera-based distance detection methods always need additional equipment, which results in more complexity and cost. In contrast, RGB camera-based methods often assume that the object’s size is known or the object is at a fixed height. To make distance detection more adaptive and with low cost, a novel and efficient distance detection method based on monocular RGB images is proposed in this paper. With a simple marker, the method can estimate the object’s distance in real-time from the pixel information obtained by a general, lightweight target detector. Experiments on the Baxter robot platform show the effectiveness of the proposed method, where the success rate of the grasping test reaches 87.5%, and the success rate of the HRH test goes 84.7%.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Andrew, A.M.: Multiple view geometry in computer vision. Kybernetes 30(9/10), 1333–1341 (2001)
Bauer, A., Wollherr, D., Buss, M.: Human-robot collaboration: a survey. Int. J. Hum. Robot. 5(1), 47–66 (2008)
Chiang, Y.-M., Hsu, N.-Z., Lin, K.-L.: Driver assistance system based on monocular vision. In: Nguyen, N.T., Borzemski, L., Grzech, A., Ali, M. (eds.) IEA/AIE 2008. LNCS (LNAI), vol. 5027, pp. 1–10. Springer, Heidelberg (2008). https://doi.org/10.1007/978-3-540-69052-8_1
Glenn, J., Alex, S., Jirka, B.: Ultralytics/YOLOv5: v5.0 - YOLOv5-P6 1280 models, AWS, Supervise.ly and YouTube integrations (2021). https://doi.org/10.5281/zenodo.4679653
Chiang, Y.-M., Hsu, N.-Z., Lin, K.-L.: Driver assistance system based on monocular vision. In: Nguyen, N.T., Borzemski, L., Grzech, A., Ali, M. (eds.) IEA/AIE 2008. LNCS (LNAI), vol. 5027, pp. 1–10. Springer, Heidelberg (2008). https://doi.org/10.1007/978-3-540-69052-8_1
Haseeb, M., Guan, J., Ristic-Durrant, D.: Disnet: a novel method for distance estimation from monocular camera. In: 10th Planning. Perception and Navigation for Intelligent Vehicles (PPNIV18), IROS (2018)
Jamzad, M., Foroughnassiraei, A., Chiniforooshan, E.: Middle sized soccer robots: Arvand. In: Robot Soccer World Cup (RoboCup) 1999. vol. 1856, pp. 61–73. Springer (1999). https://doi.org/10.1007/3-540-45327-X_4
Krishnan, J.V.G., Manoharan, N., Rani, B.S.: Estimation of distance to texture surface using complex log mapping. J. Comput. Appl. 3(3), 16 (2010)
Mousavian, A., Eppner, C., Fox, D.: 6-DoF GraspNet: variational grasp generation for object manipulation. In: International Conference on Computer Vision (ICCV) 2019, pp. 2901–2910. IEEE/CVF (2019). https://doi.org/10.1109/ICCV.2019.00299
Murali, A., Mousavian, A., Eppner, C.: 6-DOF grasping for target-driven object manipulation in clutter. In: International Conference on Robotics and Automation (ICRA) 2020, pp. 6232–6238. IEEE (2020). https://doi.org/10.1109/ICRA40945.2020.9197318
Ortenzi, V., Cosgun, A., Pardi, T.: Object handovers: a review for robotics. IEEE Trans. Robot. 37, 1855–1873 (2021)
Pathi, S.K., Kiselev, A., Kristoffersson, A.: A novel method for estimating distances from a robot to humans using egocentric RGB camera. Sensors 19(14), 3142 (2019)
Ricardo, S.M., Konstantinos, C., Apostolos, M.: Benchmark for human-to-robot handovers of unseen containers with unknown filling. IEEE Robot. Autom. Lett. 5(2), 1642–1649 (2020)
Rosenberger, P., Cosgun, A., Newbury, R.: Object-independent human-to-robot handovers using real time robotic vision. IEEE Robot. Autom. Lett. 6(1), 17–23 (2021)
Shan, D., Geng, J., Shu, M.: Understanding human hands in contact at internet scale. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2020). https://doi.org/10.1109/CVPR42600.2020.00989
Vohra, M., Prakash, R., Behera, L.: Real-time grasp pose estimation for novel objects in densely cluttered environment. In: International Conference on Robot and Human Interactive Communication (RO-MAN) 2019, pp. 1–6. IEEE (2019). https://doi.org/10.1109/RO-MAN46459.2019.8956438
Vyas, D.R., Markana, A., Padhiyar, N.: Robotic grasp synthesis using deep learning approaches: a survey. In: Sahni, M., Merigó, J.M., Jha, B.K., Verma, R. (eds.) Mathematical Modeling, Computational Intelligence Techniques and Renewable Energy. AISC, vol. 1287, pp. 117–130. Springer, Singapore (2021). https://doi.org/10.1007/978-981-15-9953-8_11
Wang, C., Xu, D., Zhu, Y.: Densefusion: 6d object pose estimation by iterative dense fusion. In: Conference on Computer Vision and Pattern Recognition (CVPR) 2019, pp. 3338–3347. IEEE/CVF (2019). https://doi.org/10.1109/CVPR.2019.00346
Yang, W., Paxton, C., Arsalan, M.: Reactive human-to-robot handovers of arbitrary objects (2020)
Yang, W., Paxton, C., Cakmak, M.: Human grasp classification for reactive human-to-robot handovers. In: International Conference on Intelligent Robots and Systems (IROS) 2020, pp. 11123–11130. IEEE/RSJ (2020). https://doi.org/10.1109/IROS45743.2020.9341004
Zhang, H., Lan, X., Bai, S.: Roi-based robotic grasp detection for object overlapping scenes. In: International Conference on Intelligent Robots and Systems (IROS) 2019. pp. 4768–4775. IEEE/RSJ (2019). https://doi.org/10.1109/IROS40897.2019.8967869
Zhu, J., Fang, Y.: Learning object-specific distance from a monocular image. In: International Conference on Computer Vision (ICCV) 2019, pp. 3838–3847. IEEE/CVF (2019). https://doi.org/10.1109/ICCV.2019.00394
Acknowledgement
This work was supported in part by the Key Program of NSFC (Grant No. U1908214), Special Project of Central Government Guiding Local Science and Technology Development (Grant No. 2021JH6/10500140), Program for the Liaoning Distinguished Professor, Program for Innovative Research Team in University of Liaoning Province, Dalian and Dalian University, the Scientific Research fund of Liaoning Provincial Education Department (No. L2019606), Dalian University Scientific Research Platform Project (No. 202101YB03), and in part by the Science and Technology Innovation Fund of Dalian (Grant No. 2020JJ25CY001).
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering
About this paper
Cite this paper
Liu, D. et al. (2021). A Novel and Efficient Distance Detection Based on Monocular Images for Grasp and Handover. In: Gao, H., Wang, X. (eds) Collaborative Computing: Networking, Applications and Worksharing. CollaborateCom 2021. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 406. Springer, Cham. https://doi.org/10.1007/978-3-030-92635-9_37
Download citation
DOI: https://doi.org/10.1007/978-3-030-92635-9_37
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-92634-2
Online ISBN: 978-3-030-92635-9
eBook Packages: Computer ScienceComputer Science (R0)