Action Transition Recognition Using Principal Component Analysis for Agricultural Robot Following

  • Conference paper
  • First Online:
Intelligent Autonomous Systems 18 (IAS 2023)

Part of the book series: Lecture Notes in Networks and Systems ((LNNS,volume 795))

Included in the following conference series:

  • 67 Accesses

Abstract

During harvesting using a robot, a farmer performs the actions of harvesting the crop and placing it on the robot. Meanwhile, the robot follows the farmer, and the farmer is comfortable if the distance at which the robot follows changes dependent on the farmer’s actions. In this study, we proposed a method of recognizing the action transition using the result of the principal component analysis of the skeletal information. It was confirmed that the proposed method recognized the transition to the placing action at the start of the action by the values of the 1st–3rd principal components. In addition, the proposed method recognizes the action transition even when the employed data is not used to calculate eigenvectors for principal component analysis. These results confirm that the proposed method is sufficient to follow dependent on the actions.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now
Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free ship** worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Islam, M.M., Lam, A., Fukuda, H., Kobayashi, Y., Kuno, Y.: A person-following shop** support robot based on human pose skeleton data and lidar sensor. In: Intelligent Computing Methodologies: 15th International Conference, ICIC 2019, pp. 9–19. Nanchang, China, August 3–6, 2019, Proceedings, Part III 15, Springer (2019)

    Google Scholar 

  2. Kosuke, I., Miki, A., Masayuki, K., Kiminori, S., Mutsumi, W.: Person tracking system using a two-wheeled robot with rgb-d camera. ROBOMECH2014 2014 (2014) P1–T06 (In Japanese)

    Google Scholar 

  3. Yuta, K., Takehiko, S., Yoshimitsu, A.: Human action recognition with pose feature extraction and action transition using CNN. Instit. Electr. Inf. Commun. Eng. D 100(7), 681–691 (2017). (In Japanese)

    Google Scholar 

  4. Donahue, J., Anne Hendricks, L., Guadarrama, S., Rohrbach, M., Venugopalan, S., Saenko, K., Darrell, T.: Long-term recurrent convolutional networks for visual recognition and description. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2625–2634 (2015)

    Google Scholar 

  5. Wu, Z., Wang, X., Jiang, Y.G., Ye, H., Xue, X.: Modeling spatial-temporal clues in a hybrid deep learning framework for video classification. In: Proceedings of the 23rd ACM international conference on Multimedia. (2015) 461–470

    Google Scholar 

  6. Quintero, R., Parra, I., Llorca, D.F., Sotelo, M.: Pedestrian intention and pose prediction through dynamical models and behaviour classification. In: 2015 IEEE 18th International Conference on Intelligent Transportation Systems, pp. 83–88. IEEE (2015)

    Google Scholar 

  7. Lucia, A., Ayanori, Y., Akihisa, O., Takashi, T.: Human recognition for agricultural robots to follow worker in a narrow furrow -recognition of center position of the body using rgb-d camera and posenet-. In: 2021 JSME Conference on Robotics and Mechatronics (2021) (In Japanese)

    Google Scholar 

  8. Yorozu, A., Ishigami, G., Takahashi, M.: Ridge-tracking for strawberry harvesting support robot according to farmer’s behavior. In: Field and Service Robotics: Results of the 12th International Conference, pp. 235–245. Springer (2021)

    Google Scholar 

  9. Wang, H., Kläser, A., Schmid, C., Liu, C.L.: Dense trajectories and motion boundary descriptors for action recognition. Int. J. Comput. Vision 103(1), 60–79 (2013)

    Article  MathSciNet  Google Scholar 

  10. Tran, D., Bourdev, L., Fergus, R., Torresani, L., Paluri, M.: Learning spatiotemporal features with 3d convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 489–4497 (2015)

    Google Scholar 

  11. Ji, S., Xu, W., Yang, M., Yu, K.: 3d convolutional neural networks for human action recognition. IEEE Trans. Patt. Anal. Mach. Intell. 35(1), 221–231 (2012)

    Article  Google Scholar 

  12. Akutsu, T., Chihiba, S.: Development of an algorithm for estimating farm work using wearable sensors. Multimedia, Distrib., Cooperat., Mob. Sympos. 2017(2017), 213–216 (2017). (In Japanese)

    Google Scholar 

  13. Koppula, H.S., Saxena, A.: Anticipating human activities using object affordances for reactive robotic response. IEEE Trans. Patt. Anal. Mach. Intell. 38(1), 14–29 (2015)

    Article  Google Scholar 

  14. Ryoo, M.S., Fuchs, T.J., **a, L., Aggarwal, J.K., Matthies, L.: Robot-centric activity prediction from first-person videos: What will they do to me? In: Proceedings of the Tenth Annual ACM/IEEE International Conference on Human-Robot Interaction, pp. 295–302 (2015)

    Google Scholar 

  15. Yao, A., Gall, J., Gool, L.V., Urtasun, R.: Learning probabilistic non-linear latent variable models for tracking complex activities. In Shawe-Taylor, J., Zemel, R., Bartlett, P., Pereira, F., Weinberger, K.Q., (eds.) Advances in Neural Information Processing Systems, vol. 24. Curran Associates, Inc. (2011)

    Google Scholar 

  16. Google: Tensorflow - poseestimation. https://www.tensorflow.org/lite/examples/pose_estimation/overview

    Google Scholar 

Download references

Acknowledgements

The robot and experimental fields were provided by the DONKEY Corporation.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Chihiro Ooka .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ooka, C., Ohya, A., Yorozu, A. (2024). Action Transition Recognition Using Principal Component Analysis for Agricultural Robot Following. In: Lee, SG., An, J., Chong, N.Y., Strand, M., Kim, J.H. (eds) Intelligent Autonomous Systems 18. IAS 2023. Lecture Notes in Networks and Systems, vol 795. Springer, Cham. https://doi.org/10.1007/978-3-031-44851-5_14

Download citation

Publish with us

Policies and ethics

Navigation