Abstract
Human-Robot Interaction challenges the field of research on Artificial Intelligence in many ways, especially regarding the complexity of the physical world. While physical interactions require Artificial Intelligence techniques to handle dynamic, nondeterministic, and partially unknown environments, the communication with humans requires socially acceptable responses and common-sense knowledge to handle a broad variety of situations with complex semantics to interpret and understand. In the context of emotional design, different Artificial Intelligence techniques are necessary to allow robots to express, understand, and induce emotions as part of the interaction process. This chapter explores Human-Robot Interaction from the Artificial Intelligence point of view, presenting the main challenges, techniques, and our particular vision for future developments in this research area.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
In a nutshell, given a set of training points (xi, yi), a system that learns by example tries to find a function f that maps a given x to its corresponding y (within a certain error tolerance). In a neural network, this function is represented by numerical weights associated with each node. During training, these numbers are continually adjusted until training data with the same labels consistently yielding similar outputs.
- 2.
Apart from being “creative”, this is totally aligned with early psychological theories, such as the one by Edward Thorndike in the very ending of the nineteenth century (Thorndike 1911), which are one of the first references on learning mentioned by researchers of artificial neural networks (also known as connectionists) (Knight 2017; Ertugrul and Tagluk 2017).
- 3.
A fun but complete and accurate history of AI until the early 1990s can be found in Crevier (1993).
- 4.
GPUs are designed for rendering graphics by having a large number of simple process units for massively parallel calculation. However, we can use GPUs to perform any sort of computation (e.g., deep learning computation). We can use multiple GPUs to increase processing power.
- 5.
Cognitive robots, different from industrial robots, are robots that reason, remember, learn, anticipate, plan, and communicate with humans and with each other.
- 6.
Such as aerial vehicles, where a single failure is catastrophic.
References
Aggarwal C, Reddy C (2013) Data clustering: algorithms and applications. Chapman and Hall/CRC, London
Altman NS (2012) An introduction to kernel and nearest-neighbor nonparametric regression. Am Stat 46(3):175–185. https://doi.org/10.2307/2685209
Amodei D, Olah C, Steinhardt J, Christiano P, Schulman J, Mané D (2016) Concrete problems in AI safety. ar**v preprint: 1606.06565
Bab-Hadiashar A, Suter D (1998) Robust optic flow computation. Int J Comput Vis 29(1):59–77. https://doi.org/10.1023/A:1008090730467
Barros P, Weber C, Wermter S (2015). Emotional expression recognition with a cross-channel convolutional neural network for human-robot interaction. In: Proceedings of the 2015 IEEE-RAS 15th international conference on humanoid robots (humanoids). https://doi.org/10.1109/HUMANOIDS.2015.7363421
Berger J (1985) Statistical decision theory and Bayesian analysis. Springer, New York
Bhatti M, Wang Y, Guan L (2004) A neural network approach for human emotion recognition in speech. In: 2004 IEEE international symposium on circuits and systems. https://doi.org/10.1109/ISCAS.2004.1329238
Bojarski M et al (2016) End to end learning for self-driving cars. ar**v preprint: 1604.07316v1
Britannica Encyclopedia (2003) Encyclopedia Britannica, 15th edn. Encyclopedia Britannica Inc., Chicago
Brox T, Bruhn A, Papenberg N, Weickert J (2004) High accuracy optical flow using a theory for war**. In: Pajdla T, Matas J (eds) Computer vision—ECCV 2004, vol 3024, pp 25–36. https://doi.org/10.1007/978-3-540-24673-2_3
Carter S, Nielsen M (2017) Using artificial intelligence to augment human intelligence. Distill J. https://doi.org/10.23915/distill.00009
Castleman K (1996) Digital image processing. Prentice Hall, New York
Cianciolo AT, Sternberg RJ (2008) Intelligence: a brief history. Wiley-Blackwell Publishing, Malden
Crevier D (1993) AI: the tumultuous history of the search for artificial intelligence. Basic Books, New York
Darrell T, Gordon G, Harville M, Woodfill J (2000) Integrated person tracking using stereo, color, and pattern detection. Int J Comput Vis 37(2):175–185. https://doi.org/10.1023/A:1008103604354
Dehaene S, Lau H, Kouider S (2017) What is consciousness, and could machines have it? Science 358:486–492. https://doi.org/10.1126/science.aan8871
Dellaert F, Fox D, Burgard W, Thrun S (1999) Monte Carlo localization for mobile robots. In: Proceedings of the 1999 IEEE international conference on robotics and automation. https://doi.org/10.1109/ROBOT.1999.772544
Ertugrul O, Tagluk ME (2017) A novel machine learning method based on generalized behavioral learning theory. Neural Comput Appl 28(12):3921–3939. https://doi.org/10.1007/s00521-016-2314-8
Ge S, Li Y, He H (2011) Neural-network-based human intention estimation for physical human-robot interaction. In: Proceedings of the 8th international conference on ubiquitous robots and ambient intelligence (URAI). https://doi.org/10.1109/URAI.2011.6145849
Horn B, Schunck B (1981) Determining optical flow. Artif Intell 17(1–3):185–203. https://doi.org/10.1016/0004-3702(81)90024-2
Hu M (1962) Visual problem recognition by moment invariants. IRE Trans Inf Theory 8:179–187
Ito M, Noda K, Hoshino Y, Tani J (2006) Dynamic and interactive generation of object handling behaviors by a small humanoid robot using a dynamic neural network model. Neural Netw 19(3):323–337. https://doi.org/10.1016/j.neunet.2006.02.007
Jetto L, Longhi S, Venturini G (1999) Development and experimental validation of an adaptive extended Kalman filter for the localization of mobile robots. IEEE Trans Robot Autom 15(2):219–229. https://doi.org/10.1109/70.760343
Kass M, Witkin A, Terzopoulos D (1988) Snakes: active contour models. Int J Comput Vis 1(4):321–331. https://doi.org/10.1007/BF00133570
Khan F, Anwer R, Weijer J, Bagdanov A, Vanrell M, Lopez A (2012) Color attributes for object detection. In: Proceedings of the 2012 IEEE conference on computer vision and pattern recognition (CVPR), pp 3306–3313. https://doi.org/10.1109/CVPR.2012.6248068
Knight W (2017) Reinforcement learning. MIT Technology Review. https://www.technologyreview.com/s/603501/10-breakthrough-technologies-2017-reinforcement-learning/. Accessed 12 June 2018
Knight W (2018) Intel’s new chips are more brain-like than ever. MIT Technology Review. https://www.technologyreview.com/s/609909/intels-new-chips-are-more-brain-like-than-ever/. Accessed 12 June 2018
Lawrence S, Giles C, Tsoi A, Back C (1997) Face recognition: a convolutional neural-network approach. IEEE Trans Neural Netw 8(1):98–113. https://doi.org/10.1109/72.554195
Le QV, Ranzato M, Monga R, Devin M, Chen K, Corrado GS, Dean J, Ng AY (2012) Building high-level features using large scale unsupervised learning. ar**v preprint: 1112.6209v5 [cs.LG]
LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436–444. https://doi.org/10.1038/nature14539
Lemaignan S, Warnier M, Sisbot E, Clodic A, Alami R (2017) Artificial cognition for social human–robot interaction: an implementation. Artif Intell 247:45–69. https://doi.org/10.1016/j.artint.2016.07.002
Lipton ZC (2017) The mythos of model interpretability. ar**v preprint: 1606.03490v3
Loehlin JC (1998) Latent variable models: an introduction to factor, path, and structural analysis, 3rd edn. Lawrence Erlbaum Associates Publishers, Mahwah
Lowe D (1999) Object recognition from local scale-invariant features. In: Proceedings of the Seventh IEEE International Conference on Computer Vision, pp 1150–1157. https://doi.org/10.1109/ICCV.1999.790410
Maturana D, Scherer S (2015) VoxNet: a 3D convolutional neural network for real-time object recognition. In: Proceedings of the 2015 IEEE/RSJ international conference on intelligent robots and systems (IROS). https://doi.org/10.1109/IROS.2015.7353481
Merriam-Webster (2016). The Merriam-Webster dictionary, New Edition. Merriam-Webster, Massachusetts
Metz C (2017) AI is about to learn more like humans—with a little uncertainty. Wired, Business. https://www.wired.com/2017/02/ai-learn-like-humans-little-uncertainty/. Accessed 12 June 2018
Mitchell T (1997) Machine learning. McGraw-Hill, New York
Ojala T, Pietikäinen M, Harwood D (1996) A comparative study of texture measures with classification based on featured distributions. Pattern Recognit 29(1):51–59. https://doi.org/10.1016/0031-3203(95)00067-4
Olah C, Satyanarayan A, Johnson I, Carter S, Schubert L, Ye K, Mordvintsev A (2018) The building blocks of interpretability. Distill J. https://doi.org/10.23915/distill.00010
Pierson H, Gashler M (2017) Deep learning in robotics: a review of recent research. Adv Robot 31(16):821–835. https://doi.org/10.1080/01691864.2017.1365009
Pratt GA (2015) Is a cambrian explosion coming for robotics? J Econ Perspect 29(3):51–60. https://doi.org/10.1257/jep.29.3.51
Priddy K, Keller P (2005) Artificial neural networks: an introduction. SPIE Publications, Washington
Rokach L, Maimon O (2014) Data mining with decision trees: theory and applications, 2nd edn. World Scientific Publishing Company, Singapore
Russell S, Norvig P (2009) Artificial intelligence: a modern approach, 3rd edn. Pearson, London
Santoro A, Bartunov S, Botvinick M, Wierstra D, Lillicrap T (2016) One-shot learning with memory-augmented neural networks. ar**v preprint: 1605.06065v1 [cs.LG]
Seemann E, Nickel K, Stiefelhagen R (2004) Head pose estimation using stereo vision for human-robot interaction. In: Proceedings of the sixth IEEE international conference on automatic face and gesture recognition. https://doi.org/10.1109/AFGR.2004.1301603
Silver D et al (2016) Mastering the game of Go with deep neural networks and tree search. Nature 529:484–489. https://doi.org/10.1038/nature16961
Steinwart I, Christmann A (2008) Support vector machines. Springer, Berlin
Stevenson A (2010) Oxford dictionary of English, 3rd edn. Oxford University Press, Oxford
Szepesvári C (2010) Algorithms for reinforcement learning. Morgan and Claypool Publishers, California
Thorndike EL (1911) Animal intelligence. Macmillan, New York
Thrun S, Leonard JJ (2008) Simultaneous localization and map**. In: Siciliano B, Khatib O (eds) Springer handbook of robotics. Springer, Berlin, Heidelberg
Thrun S, Burgard W, Fox D (2005) Probabilistic robotics. MIT Press, Massachusetts
Wang L, He D (1990) Texture classification using texture spectrum. Pattern Recognit 23(8):905–910. https://doi.org/10.1016/0031-3203(90)90135-8
Wang X, Xu H, Wang H, Li H (2008). Robust real-time face detection with skin color detection and the modified census transform. In: IEEE international conference on information and automation, pp 590–595. https://doi.org/10.1109/ICINFA.2008.4608068
Watkins C, Dayan P (1992) Q-learning. Mach Learn 8(3–4):279–292. https://doi.org/10.1007/BF00992698
Wu Y et al (2016) Google’s neural machine translation system: bridging the gap between human and machine translation. ar**v preprint: 1609.08144v2
Yan H, Ang MH, Poo A (2014) A survey on perception methods for human-robot interaction in social robots. Int J Soc Robot 6(1):85–119. https://doi.org/10.1007/s12369-013-0199-6
Yin X, **e M (2007) Finger identification and hand posture recognition for human–robot interaction. Image Vis Comput 25(8):1291–1300. https://doi.org/10.1016/j.imavis.2006.08.003
Yosinski J, Clune J, Bengio Y, Lipson H (2014) How transferable are features in deep neural networks? In: Proceedings of the 27th international conference on neural information processing systems (NIPS’14), pp 3320–3328
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this chapter
Cite this chapter
de Lima, E.S., Feijó, B. (2019). Artificial Intelligence in Human-Robot Interaction. In: Ayanoğlu, H., Duarte, E. (eds) Emotional Design in Human-Robot Interaction. Human–Computer Interaction Series. Springer, Cham. https://doi.org/10.1007/978-3-319-96722-6_11
Download citation
DOI: https://doi.org/10.1007/978-3-319-96722-6_11
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-96721-9
Online ISBN: 978-3-319-96722-6
eBook Packages: Computer ScienceComputer Science (R0)