Hostname: page-component-699b5d5946-nm5pm Total loading time: 0 Render date: 2026-02-27T00:40:00.363Z Has data issue: false hasContentIssue false

Research on robotic grasp detection using improved generative convolution neural network with Gaussian representation

Published online by Cambridge University Press:  14 November 2025

Zhanglai Chen
Affiliation:
School of Mechatronic Engineering and Automation, Shanghai University, Shanghai, China
Xu Zhang
Affiliation:
School of Mechatronic Engineering and Automation, Shanghai University, Shanghai, China
Dawei Tu*
Affiliation:
School of Mechatronic Engineering and Automation, Shanghai University, Shanghai, China
*
Corresponding author: Dawei Tu; Email: tdw@shu.edu.cn

Abstract

Grasp detection is a significant research direction in the field of robotics. Traditional analysis methods typically require prior knowledge of the object parameters, limiting grasp detection to structured environments and resulting in suboptimal performance. In recent years, the generative convolutional neural network (GCNN) has gained increasing attention, but they suffer from issues such as insufficient feature extraction capabilities and redundant noise. Therefore, we proposed an improved method for the GCNN, aimed at enabling fast and accurate grasp detection. First, a two-dimensional (2D) Gaussian kernel was introduced to re-encode grasp quality to address the issue of false positives in grasp rectangular metrics, emphasizing high-quality grasp poses near the central point. Additionally, to address the insufficient feature extraction capabilities of the shallow network, a receptive field module was added at the neck to enhance the network’s ability to extract distinctive features. Furthermore, the rich feature information in the decoding phase often contains redundant noise. To address this, we introduced a global-local feature fusion module to suppress noise and enhance features, enabling the model to focus more on target information. Finally, relevant evaluation experiments were conducted on public grasping datasets, including Cornell, Jacquard, and GraspNet-1 Billion, as well as in real-world robotic grasping scenarios. All results showed that the proposed method performs excellently in both prediction accuracy and inference speed and is practically feasible for robotic grasping.

Information

Type
Research Article
Copyright
© The Author(s), 2025. Published by Cambridge University Press

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

Article purchase

Temporarily unavailable

References

Dong, M. and Zhang, J., “A review of robotic grasp detection technology,” Robotica 41(12), 38463885 (2023).CrossRefGoogle Scholar
Yun, J., Jiang, D., Huang, L., Tao, B., Liao, S., Liu, Y., Liu, X., Li, G., Chen, D. and Chen, B., “Grasping detection of dual manipulators based on Markov decision process with neural network,” Neural Netw. 169, 778792 (2024).CrossRefGoogle Scholar
Huang, X., Halwani, M., Muthusamy, R., Ayyad, A., Swart, D., Seneviratne, L., Gan, D. and Zweiri, Y., “Real-time grasping strategies using event camera,” J. Intell. Manuf. 33(2), 593615 (2022).CrossRefGoogle Scholar
Marwan, Q. M., Chua, S. C. and Kwek, L. C., “Comprehensive review on reaching and grasping of objects in robotics,” Robotica 39(10), 18491882 (2021).CrossRefGoogle Scholar
Du, G., Wang, K., Lian, S. and Zhao, K., “Vision-based robotic grasping from object localization, object pose estimation to grasp estimation for parallel grippers: A review,” Artif. Intell. Rev. 54(3), 16771734 (2021).CrossRefGoogle Scholar
Chen, L., Niu, M., Yang, J., Qian, Y., Li, Z., Wang, K., Yan, T. and Huang, P., “Robotic grasp detection using structure prior attention and multiscale features,” IEEE Trans. Syst. Man Cybern. -Syst. 54(11), 70397053 (2024).CrossRefGoogle Scholar
Zhu, M., Derpanis, K. G., Yang, Y., Brahmbhatt, S., Zhang, M., Phillips, C., Lecce, M. and Daniilidis, K., "Single Image 3D Object Detection and Pose Estimation for Grasping," In: 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China (2014).Google Scholar
Ten Pas, A. and Platt, R., “Using geometry to detect grasp poses in 3d point clouds,” Robotics Research. 2, 307324 (2018).CrossRefGoogle Scholar
Newbury, R., Gu, M., Chumbley, L., Mousavian, A., Eppner, C., Leitner, J. C., Bohg, J., Morales, A., Asfour, T., Kragic, D., Fox, D. and Cosgun, A., “Deep learning approaches to grasp synthesis: A review,” IEEE Trans. Robot. 39(5), 39944015 (2023).CrossRefGoogle Scholar
Kushwaha, V., Shukla, P. and Nandi, G. C., “Vision-based intelligent robot grasping using sparse neural network,” Int J Intell Robot Appl. 9, 12141227 (2025).CrossRefGoogle Scholar
Kim, D., Li, A. and Lee, J., “Stable robotic grasping of multiple objects using deep neural networks,” Robotica 39(4), 735748 (2021).CrossRefGoogle Scholar
Yun, J., Moseson, S. and Saxena, A., "Efficient Grasping from RGBD Images: Learning Using a New Rectangle Representation," In: 2011 IEEE International Conference on Robotics and Automation, Shanghai, China (2011).Google Scholar
Lenz, I., Lee, H. and Saxena, A., “Deep learning for detecting robotic grasps,” Int. J. Robot. Res. 34(4-5), 705724 (2015).CrossRefGoogle Scholar
Redmon, J. and Angelova, A., "Real-time Grasp Detection Using Convolutional Neural Networks," In: 2015 IEEE International Conference on Robotics and Automation (ICRA), Washington, USA (2015).Google Scholar
Chu, F.-J., Xu, R. and Vela, P. A., “Deep grasp: Detection and localization of grasps with deep neural networks,” ArXiv. abs/1802.00520, (2018).Google Scholar
Zhou, X., Lan, X., Zhang, H., Tian, Z., Zhang, Y. and Zheng, N., "Fully Convolutional Grasp Detection Network with Oriented Anchor Box," In: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain (2018).Google Scholar
Morrison, D., Corke, P. and Leitner, J., “Learning robust, real-time, reactive robotic grasping,” Int. J. Robot. Res. 39(2-3), 183201 (2020).CrossRefGoogle Scholar
Kumra, S., Joshi, S. and Sahin, F., "Antipodal Robotic Grasping Using Generative Residual Convolutional Neural Network," In: 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, Nevada, USA (2020).Google Scholar
Kumra, S., Joshi, S. and Sahin, F., “GR-ConvNet v2: A real-time multi-grasp detection network for robotic grasping,” Sensors 22(16), 6208 (2022).CrossRefGoogle Scholar
Prew, W., Breckon, T., Bordewich, M. and Beierholm, U., "Evaluating Gaussian Grasp Maps for Generative Grasping Models," In: 2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy (2022).Google Scholar
Cao, H., Chen, G., Li, Z., Feng, Q., Lin, J. and Knoll, A., “Efficient grasp detection network with Gaussian-based grasp representation for robotic manipulation,” IEEE/ASME Trans. Mechatron. 28(3), 13841394 (2022).CrossRefGoogle Scholar
Liu, S., Huang, D. and Wang, Y., “Receptive Field Block Net for Accurate and Fast Object Detection,” In: Computer Vision – ECCV 2018, Cham, (2018).Google Scholar
Hu, J., Shen, L. and Sun, G., "Squeeze-and-Excitation Networks," In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, Utah (2018).Google Scholar
Depierre, A., Dellandréa, E. and Chen, L., "Jacquard: A Large Scale Dataset for Robotic Grasp Detection," In: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain (2018).Google Scholar
Fang, H. S., Wang, C., Gou, M. and Lu, C., "GraspNet-1Billion: A Large-Scale Benchmark for General Object Grasping," In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, USA (2020).Google Scholar
Kumra, S. and Kanan, C., "Robotic Grasp Detection Using Deep Convolutional Neural Networks," In: 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada (2017).Google Scholar
Karaoguz, H. and Jensfelt, P., "Object Detection Approach for Robot Grasp Detection," In: 2019 International Conference on Robotics and Automation (ICRA), Montreal, Canada (2019).Google Scholar
Chen, L., Huang, P., Li, Y. and Meng, Z., “Edge-dependent efficient grasp rectangle search in robotic grasp detection,” IEEE/ASME Trans. Mechatron. 26(6), 29222931 (2021).CrossRefGoogle Scholar
Yu, S., Zhai, D. H., Xia, Y., Wu, H. and Liao, J., “SE-ResUNet: A novel robotic grasp detection method,” IEEE Rob. Autom. Lett. 7(2), 52385245 (2022).CrossRefGoogle Scholar
Zhai, D. H., Yu, S. and Xia, Y., “FANet: Fast and accurate robotic grasp detection based on keypoints,” IEEE Trans. Autom. Sci. Eng. 21(3), 29742986 (2024).CrossRefGoogle Scholar
Wang, S., Zhou, Z. and Kan, Z., “When transformer meets robotic grasping: Exploits context for efficient grasp detection,” IEEE Rob. Autom. Lett. 7(3), 81708177 (2022).CrossRefGoogle Scholar
Xi, H., Li, S. and Liu, X., “A pixel-level grasp detection method based on efficient grasp aware network,” Robotica 42(9), 31903210 (2024).CrossRefGoogle Scholar
Zhou, Z., Zhang, X., Ran, L., Han, Y. and Chu, H., "DSC-GraspNet: A Lightweight Convolutional Neural Network for Robotic Grasp Detection," In: 2023 9th International Conference on Virtual Reality (ICVR), Xianyang, China (2023).Google Scholar
Johns, E., Leutenegger, S. and Davison, A. J., "Deep learning a grasp function for grasping under gripper pose uncertainty," In: 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Korea (2016).Google Scholar
Guo, C., Zhu, C., Liu, Y., Huang, R., Cao, B., Zhu, Q., Zhang, R. and Zhang, B., “End-to-end lightweight transformer-based neural network for grasp detection towards fruit robotic handling,” Comput. Electron. Agric. 221, 109014 (2024).CrossRefGoogle Scholar
Zhu, Z., Huang, S., Xie, J., Meng, Y., Wang, C. and Zhou, F., “A refined robotic grasp detection network based on coarse-to-fine feature and residual attention,” Robotica 43(2), 415432 (2025).CrossRefGoogle Scholar
Yan, Y., Tong, L., Song, K., Tian, H., Man, Y. and Yang, W., “SISG-Net: Simultaneous instance segmentation and grasp detection for robot grasp in clutter,” Adv. Eng. Inform. 58, 102189 (2023).CrossRefGoogle Scholar