Abstract
The focus of this paper is on investigating the use of visual communication between two humanoid robots, in order to enhance the coordination of tasks between them. The problem continues to be an interesting and fruitful area of research from the days of using multiple robot manipulator arms for manufacturing as well as space robotics to current research in medical robotics. The approach here is to employ several off-the-shelf algorithms, software and hardware such as the NAO robot and support software, including Choregraphe, OpenCV to capture and process images, the SVM to classify objects in images, and the Python programming environment. Five robotic actions were studied and three modes. The experiments used one robot as the “viewer” and the second robot as the “subject” being analyzed. Results show that the visual communication system has an accuracy of 90% in correctly identifying the five movements. This research has shown an original solution, as a model that can enable robots to run in the complex service tasks consisting of multiple connected actions in a dynamic environment. This methodology can also let the intelligent operation of the robots serve in different scenes according to their actual requirements. This research focuses on enhancing the prototype robot vision function and development of additional value for consolidation manageable platform that increases service robots in the home environment of intelligent control capability.
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-023-17989-w/MediaObjects/11042_2023_17989_Fig1_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-023-17989-w/MediaObjects/11042_2023_17989_Fig2_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-023-17989-w/MediaObjects/11042_2023_17989_Fig3_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-023-17989-w/MediaObjects/11042_2023_17989_Fig4_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-023-17989-w/MediaObjects/11042_2023_17989_Fig5_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-023-17989-w/MediaObjects/11042_2023_17989_Fig6_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-023-17989-w/MediaObjects/11042_2023_17989_Fig7_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-023-17989-w/MediaObjects/11042_2023_17989_Fig8_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-023-17989-w/MediaObjects/11042_2023_17989_Fig9_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-023-17989-w/MediaObjects/11042_2023_17989_Fig10_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-023-17989-w/MediaObjects/11042_2023_17989_Fig11_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11042-023-17989-w/MediaObjects/11042_2023_17989_Fig12_HTML.png)
Similar content being viewed by others
Data availability
Data sharing not applicable to this article as no datasets were generated or analyzed during the current study.
References
Aldana-Murillo NG, Hayet JB, Becerra H (2018) Comparison of Local Descriptors for Humanoid Robots Localization Using a Visual Bag of Words Approach. Intell Autom Soft Comput 24(3):471–481
Yan Lu, Song D (2017) Visual Navigation Using Heterogeneous Landmarks and Unsupervised Geometric Constraints.". IEEE Trans Robotics 31(3):736–749
Porzi L, Buló SR, Penate-Sanchez A, Ricci E (2017) Learning Depth-Aware Deep Representations for Robotic Perception. IEEE Robotics Autom Lett 2(2):468–475
Becerra H (2014) Fuzzy Visual Control for Memory-Based Navigation Using the Trifocal Tensor. Intell Autom Soft Comput 20(2):245–262
Chen MY (2018) The SLAM Algorithm for Multiple Robots Based on Parameter Estimation. Intell Autom Soft Comput 24(3):593–607
Chou YC, Nakajima M (2016) Particle Filter Planar Target Tracking with a Monocular Camera for Mobile Robots. Intell Autom Soft Comput 24(3):117–125
Mukai T, Onishi M, Odashima T, Hirano S, Luo Z (Apr.2008) Development of the tactile sensor system of a human-interactive robot ‘RI-MAN. IEEE Trans Robot 24(2):505–512
Takeda T, Hirata Y, Kosuge K (2007) Dance partner robot cooperative motion generation with adjustable length of dance step stride based on physical interaction, in Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst., pp. 3258–3263
NAO-technical overview. http://doc.aldebaran.com/2-1/family/robots/dimensions_robot.html/22/10/2018. Accessed 22/10/2018
Mohaimin Sultan Mohammad, Saha Sajib Kumar, Khan AlveMahamud, Arif Abu Shamim Mohammad, Kanagasingam Yogesan (2018) Automated method for the detection and segmentation of drusen in colour fundus image for the diagnosis of age-related macular degeneration. IET Image Process 12(6):919–927
Jenkins M, Kantor G (2017) Online detection of occluded plant stalks for manipulation. 5162 – 5167
Le P-H, Košecka J (2017) Dense piecewise planar RGB-D SLAM for indoor environments. 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS),4944 – 4949
Kim W, Johnson-Roberson M (2017) A probabilistic framework for intrinsic image decomposition from RGB-D streams. 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS),6498 – 6505
Dixit Rahul, Naskar Ruchira, Mishra Swati (2017) Blur-invariant copy-move forgery detection technique with improved detection accuracy utilising SWT-SVD. IET Image Process 11(5):301–309
Zhou P, Han X, Morariu VI, Davis LS (2018) Learning Rich Features for Image Manipulation Detection. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 1053 – 1061.
Rehman Yawar, Khan Jameel Ahmed, Shin Hyunchul (2018) Efficient coarser-to-fine holistic traffic sign detection for occlusion handling. IET Image Process 12(12):2229–2237
Spexard T, Hanheide M, Sagerer G (Oct.2007) Human-oriented interaction with an anthropomorphic robot. IEEE Trans Robot 23(5):852–862
Nigam A, Penate-Sanchez A, Agapito L (2018) Detect Globally, Label Locally: Learning Accurate 6-DOF Object Pose Estimation by Joint Segmentation and Coordinate Regression. volume: 3, Issue: 4, 3960 – 3967
George L, and Mazel A (2013) Humanoid robot indoor navigation based on 2D bar codes: application to the NAO robot. Ieee-Ras International Conference on Humanoid Robots IEEE, 329–335
Yoder JD and Seelinger M (2006) Visual Coordination of Heterogeneous Mobile Manipulators, Springer Tracts in Advanced Robotics: Proc of the 9th International Symposium on Experimental Robotics, vol. 21, no. 1, pp. 387–398
Fleurmond R, Cadenat V (2014) Multi-Cameras Visual Servoing to Perform a Coordinated Task Using a Dual Arm Robot, Proc. of the 11th International Conference on Informatics in Control. Autom Robotics 2(2):37–44
Chen ML, Kume S, Rizzi AA, Hollis RL (2000) Visually Guided Coordination for Distributed Precision Assembly. Proc IEEE Int Conf Robotics Autom 2:1651–1656
Kumar S, Kumar DK, Sharma A, McLachlan N (2004) “Classification of Visual Hand Movements Using Multiresolution Wavelet Images. Proc Int Conf Intell Sens Inform Process:373–378
Rababaah H, Haroun, Shirkhodaie, A (2008) Human-Posture Classification for Intelligent Visual Surveillance Systems”, Proc. of SPIE - The International Society for Optical Engineering: Special Sessions on Food Safety, Visual Analytics, Resource-Restricted Embedded and Sensor Networks, and 3D Imaging and Display, vol. 6983, Article 69830F
Amine Elforaici MEl, Chaaraoui I, Bouachir Y, Ouakrim W, Mezghani N (2018) Posture Recognition using an RGB-D Camera: Exploring 3D Body Modeling and Deep Learning Approaches, Proc. of the IEEE Life Sciences Conference, pp. 69–72
Shahverdi P, Masouleh MT (2016) A Simple and Fast Geometric Kinematic Solution for Imitation of Human Arms by a NAO Humanoid Robot, Proceedings of the 4th International Conference on Robotics and Mechatronics, 26–28
Shamsuddin S, Yussof H et al (2012) Humanoid Robot NAO Interacting with Autistic Children of Moderately Impaired Intelligence to Augment Communication Skills. Procedia Eng 41:1533–1538
Yu YH, Ting YS, Mayer NM, et al. (2017) A new paradigm of ubiquitous home care robot using Nao and Choregraphe, International Conference on Advanced Robotics and Intelligent Systems. IEEE
Becerra HM, Sagues C (2014) Visual Control of Wheeled Mobiles Robots, Springer Tracts in Advanced Robotics 103
Kragic D, Miller AT, Allen PK (2001) Realtime tracking meets online grasp planning, IEEE International Conference on Robotics & Automation. IEEE
Adachi J, Sato J (2010) Uncalibrated Visual Servoing from Projective Reconstruction of Control Values [J]. Electron Commun Japan 89(10):47–56
Sato T, Sato J (2015) Visual servoing from uncalibrated cameras for uncalibrated robots. Syst Comput Japan 31(14):11–19
Kumar M, Kumar **dal M, Kumar M (2021) A Novel Attack on Monochrome and Greyscale Devanagari CAPTCHAs, ACM Transactions on Asian and Low-Resource Language Information Processing vo. 20, Issue 4Article no.: 56,pp. 1–30
Kumar M, **dal MK, Kumar Munish (2023) An efficient technique for breaking of coloured Hindi CAPTCHA. Soft Comput 27:11661–11686
Acknowledgements
The author deeply acknowledges Mr. Zhang,Yi-**ang initial test support at first rough model.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Below is the link to the electronic supplementary material.
Supplementary file1 (MP4 19007 KB)
Supplementary file2 (MP4 25297 KB)
Supplementary file3 (MP4 23235 KB)
Supplementary file4 (MP4 17289 KB)
Supplementary file5 (MP4 14975 KB)
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Juang, LH. Intelligent communication of two humanoid robots based on computer vison. Multimed Tools Appl 83, 63459–63477 (2024). https://doi.org/10.1007/s11042-023-17989-w
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-023-17989-w