AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
Article Link
Collect
Submit Manuscript
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Regular Paper

Accurate Robotic Grasp Detection with Angular Label Smoothing

School of Control and Computer Engineering, North China Electric Power University, Beijing 102206, China
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
Show Author Information

Abstract

Grasp detection is a visual recognition task where the robot makes use of its sensors to detect graspable objects in its environment. Despite the steady progress in robotic grasping, it is still difficult to achieve both real-time and high accuracy grasping detection. In this paper, we propose a real-time robotic grasp detection method, which can accurately predict potential grasp for parallel-plate robotic grippers using RGB images. Our work employs an end-to-end convolutional neural network which consists of a feature descriptor and a grasp detector. And for the first time, we add an attention mechanism to the grasp detection task, which enables the network to focus on grasp regions rather than background. Specifically, we present an angular label smoothing strategy in our grasp detection method to enhance the fault tolerance of the network. We quantitatively and qualitatively evaluate our grasp detection method from different aspects on the public Cornell dataset and Jacquard dataset. Extensive experiments demonstrate that our grasp detection method achieves superior performance to the state-of-the-art methods. In particular, our grasp detection method ranked first on both the Cornell dataset and the Jacquard dataset, giving rise to the accuracy of 98.9% and 95.6%, respectively at real-time calculation speed.

Electronic Supplementary Material

Download File(s)
JCST-2103-11458-Highlights.pdf (139.8 KB)

References

[1]
Kumra S, Kanan C. Robotic grasp detection using deep convolutional neural networks. In Proc. the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems, Sept. 2017, pp.769–776. DOI: 10.1109/IROS.2017.8202237.
[2]
Aoki Y, Goforth H, Srivatsan R A, Lucey S. PointNetLK: Robust & efficient point cloud registration using PointNet. In Proc. the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Jun. 2019, pp.7163–7172. DOI: 10.1109/CVPR.2019.00733.
[3]
Choy C, Dong W, Koltun V. Deep global registration. In Proc. the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Jun. 2020, pp.2511–2520. DOI: 10.1109/CVPR42600.2020.00259.
[4]
Wang Y, Solomon J. Deep closest point: Learning representations for point cloud registration. In Proc. the 2019 IEEE/CVF International Conference on Computer Vision, Oct. 27–Nov. 2, 2019, pp.3522–3531. DOI: 10.1109/ICCV.2019.00362.
[5]
Zhou Q Y, Park J, Koltun V. Fast global registration. In Proc. the 14th European Conference on Computer Vision, Oct. 2016, pp.766–782. DOI: 10.1007/978-3-319-46475-6_47.
[6]

Lenz I, Lee H, Saxena A. Deep learning for detecting robotic grasps. The International Journal of Robotics Research , 2015, 34(4/5): 705–724. DOI: 10.1177/027836491454 9607.

[7]

Krizhevsky A, Sutskever I, Hinton G E. ImageNet classification with deep convolutional neural networks. Communications of the ACM , 2017, 60(6): 84–90. DOI: 10.1145/3065386.

[8]
Redmon J, Angelova A. Real-time grasp detection using convolutional neural networks. In Proc. the 2015 IEEE International Conference on Robotics and Automation, May 2015, pp.1316–1322. DOI: 10.1109/ICRA.2015.7139 361.
[9]
Zhou X W, Lan X G, Zhang H B, Tian Z Q, Zhang Y, Zheng N. Fully convolutional grasp detection network with oriented anchor box. In Proc. the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems, Oct. 2018, pp.7223–7230. DOI: 10.1109/IROS.2018.8594116.
[10]
He K M, Zhang X Y, Ren S Q, Sun J. Deep residual learning for image recognition. In Proc. the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Jun. 2016, pp.770–778. DOI: 10.1109/CVPR.2016.90.
[11]
Park D, Chun S Y. Classification based grasp detection using spatial transformer network. arXiv: 1803.01356, 2018. https://arxiv.org/abs/1803.01356, Oct. 2023.
[12]
Asif U, Tang J B, Harrer S. GraspNet: An efficient convolutional neural network for real-time grasp detection for low-powered devices. In Proc. the 27th International Joint Conference on Artificial Intelligence, July 2018, pp.4875–4882. DOI: 10.24963/ijcai.2018/677.
[13]
Kumra S, Joshi S, Sahin F. Antipodal robotic grasping using generative residual convolutional neural network. In Proc. the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems, Oct. 2020, pp.9626–9633. DOI: 10.1109/IROS45743.2020.9340777.
[14]
Karaoguz H, Jensfelt P. Object detection approach for robot grasp detection. In Proc. the 2019 International Conference on Robotics and Automation, May 2019, pp.4953–4959. DOI: 10.1109/ICRA.2019.8793751.
[15]

Chu F J, Xu R N, Vela P A. Real-world multiobject, multigrasp detection. IEEE Robotics and Automation Letters , 2018, 3(4): 3355–3362. DOI: 10.1109/LRA.2018.2852777.

[16]
Wang C Y, Liao H Y M, Wu Y H, Chen P Y, Hsieh J W, Yeh I H. CSPNet: A new backbone that can enhance learning capability of CNN. In Proc. the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Jun. 2020, pp.390–391. DOI: 10.1109/CVPRW50498.2020.00203.
[17]
Pavlakos G, Zhou X W, Chan A, Derpanis K G, Daniilidis K. 6-DoF object pose from semantic keypoints. In Proc. the 2017 IEEE International Conference on Robotics and Automation, May 29–June 3, 2017, pp.2011–2018. DOI: 10.1109/ICRA.2017.7989233.
[18]

Satish V, Mahler J, Goldberg K. On-policy dataset synthesis for learning robot grasping policies using fully convolutional deep networks. IEEE Robotics and Automation Letters , 2019, 4(2): 1357–1364. DOI: 10.1109/LRA.2019.2895878.

[19]
Tekin B, Sinha S N, Fua P. Real-time seamless single shot 6D object pose prediction. In Proc. the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Jun. 2018, pp.292–301. DOI: 10.1109/CVPR.2018.00038.
[20]
Peng S D, Liu Y, Huang Q X, Zhou X W, Bao H J. PVNet: Pixel-wise voting network for 6DoF pose estimation. In Proc. the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Jun. 2019, pp.4556–4565. DOI: 10.1109/CVPR.2019.00469.
[21]
He Y S, Sun W, Huang H B, Liu J R, Fan H Q, Sun J. PVN3D: A deep point-wise 3D keypoints voting network for 6DoF pose estimation. In Proc. the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Jun. 2020, pp.11632–11641. DOI: 10.1109/CVPR42600.2020.01165.
[22]
Wang C, Xu D F, Zhu Y K, Martín-Martín R, Lu C W, Li F F, Savarese S. DenseFusion: 6D object pose estimation by iterative dense fusion. In Proc. the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Jun. 2019, pp.3338–3347. DOI: 10.1109/CVPR.2019.00346.
[23]

Choi C, Schwarting W, DelPreto J, Rus D. Learning object grasping for soft robot hands. IEEE Robotics and Automation Letters , 2018, 3(3): 2370–2377. DOI: 10.1109/LRA.2018.2810544.

[24]
Liang H Z, Ma X J, Li S, Görner M, Tang S, Fang B, Sun F C, Zhang J W. PointNetGPD: Detecting grasp configurations from point sets. In Proc. the 2019 International Conference on Robotics and Automation, May 2019, pp.3629–3635. DOI: 10.1109/ICRA.2019.8794435.
[25]

Pas A, Gualtieri M, Saenko K, Platt R. Grasp pose detection in point clouds. The International Journal of Robotics Research , 2017, 36(13/14): 1455–1473. DOI: 10.1177/0278364917735594.

[26]
Mahler J, Liang J, Niyaz S, Laskey M, Doan R, Liu X Y, Ojea J A, Goldberg K. Dex-Net 2.0: Deep learning to plan robust grasps with synthetic point clouds and analytic grasp metrics. arXiv: 1703.09312, 2017. https://arxiv.org/abs/1703.09312, Oct. 2023.
[27]
Mousavian A, Eppner C, Fox D. 6-DOF GraspNet: Variational grasp generation for object manipulation. In Proc. the 2019 IEEE/CVF International Conference on Computer Vision, Oct. 27–Nov. 2, 2019, pp.2901–2910. DOI: 10.1109/ICCV.2019.00299.
[28]
Jiang Y, Moseson S, Saxena A. Efficient grasping from RGBD images: Learning using a new rectangle representation. In Proc. the 2011 IEEE International Conference on Robotics and Automation, May 2011, pp.3304–3311. DOI: 10.1109/ICRA.2011.5980145.
[29]
Guo D, Sun F C, Liu H P, Kong T, Fang B, Xi N. A hybrid deep architecture for robotic grasp detection. In Proc. the 2017 IEEE International Conference on Robotics and Automation, May 29–June 3, 2017, pp.1609–1614. DOI: 10.1109/ ICRA.2017.7989191.
[30]
Ren S Q, He K M, Girshick R, Sun J. Faster R-CNN: Towards real-time object detection with region proposal networks. In Proc. the 29th Annual Conference on Neural Information Processing Systems, Dec. 2015, pp.91–99.
[31]
Depierre A, Dellandréa E, Chen L M. Jacquard: A large scale dataset for robotic grasp detection. In Proc. the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems, Oct. 2018, pp.3511–3516. DOI: 10.1109/IROS.2018.8593950.
[32]
Redmon J, Farhadi A. YOLOV3: An incremental improvement. arXiv: 1804.02767, 2018. https://arxiv.org/abs/1804.02767, Oct. 2023.
[33]
Hu J, Shen L, Sun G. Squeeze-and-excitation networks. In Proc. the IEEE Conference on Computer Vision and Pattern Recognition, Aug. 2020, pp.7132–7141. DOI: 10.1109/TPAMI.2019.2913372.
[34]

Morrison D, Corke P, Leitner J. Learning robust, real-time, reactive robotic grasping. The International Journal of Robotics Research , 2020, 39(2/3): 183–201. DOI: 10.1177/0278364919859066.

[35]
Kingma D P, Ba J. Adam: A method for stochastic optimization. arXiv: 1412.6980, 2014. https://arxiv.org/abs/1412.6980, Oct. 2023.
Journal of Computer Science and Technology
Pages 1149-1161
Cite this article:
Shi M, Lu H, Li Z-X, et al. Accurate Robotic Grasp Detection with Angular Label Smoothing. Journal of Computer Science and Technology, 2023, 38(5): 1149-1161. https://doi.org/10.1007/s11390-022-1458-5

358

Views

1

Crossref

1

Web of Science

1

Scopus

0

CSCD

Altmetrics

Received: 22 March 2021
Accepted: 14 August 2022
Published: 30 September 2023
© Institute of Computing Technology, Chinese Academy of Sciences 2023
Return