A Grasp Pose Detection Network Based on the DeepLabv3+ Semantic Segmentation Model

  • Conference paper
  • First Online:
Intelligent Robotics and Applications (ICIRA 2022)

Abstract

Gras** is an important and fundamental action for the interaction between robots and the environment. However, because gras** is a complex system engineering, there is still much room for development. At present, many studies use regression to solve the problem of grasp detection or use the unstable 3D point cloud as input, which may cause poor results to a certain extent. In this paper, we propose to use semantic segmentation of pixel-level classification to solve the problem of grasp pose detection. We adopt a grasp detection method based on the DeepLabv3+ model, which includes semantic segmentation and post-processing. In the semantic segmentation part, the classification mask of the objects is predicted through the input RGB image, and then the predicted objects of different classifications are fitted with the minimum bounding directed rectangle to obtain the two-dimensional grasp pose, and the final three-dimensional gras** pose is calculated through the conversion of the input depth image. On the validation dataset, we use the indicator of semantic segmentation to evaluate the proposed network and achieve a great result. In addition, the simulation robot experiment further verifies the effectiveness of the network.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (Canada)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 99.00
Price excludes VAT (Canada)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 129.99
Price excludes VAT (Canada)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free ship** worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Bicchi, A., Kumar, V.: Robotic gras** and contact: a review. In: Proceedings 2000 ICRA. Millennium Conference. IEEE International Conference on Robotics and Automation, vol. 1, pp. 348–353. IEEE, San Francisco (2000)

    Google Scholar 

  2. Bohg, J., Morales, A., Asfour, T., Kragic, D.: Data-driven grasp synthesis—a survey. IEEE Trans. Rob. 30(2), 289–309 (2014)

    Article  Google Scholar 

  3. Dang, H., Allen, P.K.: Semantic gras**: planning robotic grasps functionally suitable for an object manipulation task. In: 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 1311–1317. IEEE, Vilamoura-Algarve (2012)

    Google Scholar 

  4. Jiang, Y., Moseson, S., Saxena, A.: Efficient gras** from RGBD images: learning using a new rectangle representation. In: 2011 IEEE International Conference on Robotics and Automation, pp. 3304–3311. IEEE, Shanghai (2011)

    Google Scholar 

  5. Lenz, I., Lee, H., Saxena, A.: Deep learning for detecting robotic grasps. Int. J. Robot. Res. 34(4–5), 705–724 (2015)

    Article  Google Scholar 

  6. Pinto, L., Gupta, A.: Supersizing self-supervision: learning to grasp from 50K tries and 700 robot hours. In: 2016 IEEE International Conference on Robotics and Automation, pp. 3406–3413. IEEE, Stockholm (2016)

    Google Scholar 

  7. Depierre, A., Dellandréa, E., Chen, L.: Jacquard: a large scale dataset for robotic grasp detection. In: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 3511–3516. IEEE, Madrid (2018)

    Google Scholar 

  8. Kumra, S., Kanan, C.: Robotic grasp detection using deep convolutional neural networks. In: 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 769–776. IEEE, Vancouver (2017)

    Google Scholar 

  9. Asir, U., Tang, J.B., Harrer, S.: GraspNet: an efficient convolutional neural network for real-time grasp detection for low-powered devices. In: 27th International Joint Conference on Artificial Intelligence, Stockholm, pp. 4875–4882 (2018)

    Google Scholar 

  10. Qi, C.R., Su, H., Mo, K.C., Guibas, L.J.: PointNet: deep learning on point sets for 3d classification and segmentation. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, pp. 652–660. IEEE, Honolulu (2017)

    Google Scholar 

  11. Qi, C.R., Su, H., Mo, K.C., Guibas, L.J.: PointNet plus plus: deep hierarchical feature learning on point sets in a metric space. In: 31st Annual Conference on Neural Information Processing Systems, Long Beach (2017)

    Google Scholar 

  12. Ten Pas, A., Gualtieri, M., Saenko, K., Platt, R.: Grasp pose detection in point clouds. Int. J. Robot. Res. 36(13–14), 1455–1473 (2017)

    Google Scholar 

  13. Liang, H.Z., et al.: PointNetGPD: detecting grasp configurations from point sets. In: 2019 International Conference on Robotics and Automation, pp. 3629–3635. IEEE, Montreal (2019)

    Google Scholar 

  14. Fang, H.S., Wang, C., Gou, M., Lu, C.: GraspNet-1Billion: a large-scale benchmark for general object gras**. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11441–11450. IEEE, Seattle (2020)

    Google Scholar 

  15. Gou, M., Fang, H.S., Zhu, Z., Xu, S., Wang, C., Lu, C.: RGB matters: learning 7-DoF grasp poses on monocular RGBD images. In: 2021 IEEE International Conference on Robotics and Automation, pp. 13459–13466. IEEE, **’an (2021)

    Google Scholar 

  16. Li, Y., Kong, T., Chu, R., Li, Y., Wang, P., Li, L.: Simultaneous semantic and collision learning for 6-DoF grasp pose estimation. In: 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 3571–3578. IEEE, Prague (2021)

    Google Scholar 

  17. Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., Adam, H.: Encoder-decoder with atrous separable convolution for semantic image segmentation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 833–851. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01234-2_49

    Chapter  Google Scholar 

  18. Zhao, H.S., Shi, J.P., Qi, X.J., Wang X.G., Jia J.Y.: Pyramid scene parsing network. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, pp. 6230–6239. IEEE, Honolulu (2017)

    Google Scholar 

Download references

Acknowledgments

This work was supported by the National Natural Science Foundation of China (Grant No. 62003048) and the National Key Research and Development Program of China (Grant No. 2019YFB1309802).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Haiyuan Li .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhang, Q., Zhang, X., Li, H. (2022). A Grasp Pose Detection Network Based on the DeepLabv3+ Semantic Segmentation Model. In: Liu, H., et al. Intelligent Robotics and Applications. ICIRA 2022. Lecture Notes in Computer Science(), vol 13458. Springer, Cham. https://doi.org/10.1007/978-3-031-13841-6_67

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-13841-6_67

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-13840-9

  • Online ISBN: 978-3-031-13841-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics

Navigation