Abstract
This work examines one possibility of using a deep neural network to control a virtual dance partner behavior. A neural-network-based system is designed and used for classification, evaluation, prediction, and generation of socially emotional behavior of a virtual actor. The network is trained using deep learning on the data generated with an algorithm implementing the eBICA cognitive architecture. Results show that, in the selected virtual dance paradigm, (1) the functionality of the cognitive model can be efficiently transferred to the neural network using deep learning, allowing the network to generate socially emotional behavior of a dance partner similar to a human participant behavior or the behavior generated algorithmically based on eBICA, and (2) the trained neural network can correctly identify the character types of virtual dance partners based on their behavior. When considered together with related studies, our findings lead to more general implications extending beyond the selected paradigm.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Trafton, J.G., Hiatt, L.M., Brumback, B., McCurry, J.M.: Using cognitive models to train big data models with small data. In: An, B., Yorke-Smith, N., El Fallah Seghrouchni, A., Sukthankar, G. (eds.). Proceedings of the 19th International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2020), pp. 1413–1421. International Foundation for Autonomous Agents and Multiagent Systems, Richland, SC (2020). ISBN 978-1-4503-7518-4
Karabelnikova, Y., Samsonovich, A.V.: Virtual partner dance as a paradigm for empirical study of cognitive models of emotional intelligence. Procedia Comput. Sci. 190, 414–433 (2021). https://doi.org/10.1016/j.procs.2021.06.05
Samsonovich, A.V.: Emotional biologically inspired cognitive architecture. Biol. Inspired Cogn. Archit. 6, 109–125 (2013)
Samsonovich, A.V.: Socially emotional brain-inspired cognitive architecture framework for artificial intelligence. Cogn. Syst. Res. 60, 57–76 (2020). https://doi.org/10.1016/j.cogsys.2019.12.002
Marsella, S., Gratch, J., Petta, P.: Computational models of emotion. In: Scherer, K.R., Bänziger, T., Roesch, E. (eds.) A Blueprint for Affective Computing: A Sourcebook and Manual. Oxford University Press, Oxford (2010)
Goleman, D.: Emotional Intelligence. Bantam Books, New York (1995)
Goodfellow, I., Bengio, Y., Courville, A.: Deep learning. MIT Press, Cambridge (2016)
Khaikin, S.: Neural Networks: A Complete Course, 2nd edn. Williams Publishing House, Moscow (2006). 1104 p.
Scholle, F.: Deep Learning in Python. Publishing House “Peter”, St. Petersburg (2018). (in Russian)
Isakov, S.: Recurrent neural networks: types, training, examples and applications (Electronic resource). https://neurohive.io/ru/osnovy-data-science/rekurrentnye-nejronnye-seti. (in Russian)
Gafarov, A.F.: G12 Artificial Neural Networks and Applications. Kazan Publishing House, Kazan (2018). (in Russian). 121 p.
Tsaregorodtsev, V.G.: Optimization of preprocessed data. Neurocomput. Dev. Appl. 7, 3–8 (2003)
Chollet, F.: Deep Learning with Python, 2nd edn. Manning Pub. Co., Shelter Island, New York (2021). ISBN-13: 978-1617296864
Nikolenko, S., Kadurin, A., Arkhangelskaya, E.: Deep learning. Piter, St. Petersburg (2018). (in Russian)
Osgood, C.E., Suci, G., Tannenbaum, P.: The Measurement of Meaning. University of Illinois Press, Urbana (1957)
Russell, J.A.: A circumplex model of affect. J. Pers. Soc. Psychol. 39(6), 1161–1178 (1980)
Plutchik, R.: A psychoevolutionary theory of emotions. Soc. Sci. Inf. 21, 529–553 (1982)
Lövheim, H.: A new three-dimensional model for emotions and monoamine neurotransmitters. Med. Hypotheses 78(2), 341–348 (2012)
Harris, D., Harris, S.: Digital design and computer architecture, 2nd edn. Morgan Kaufmann, San Francisco (2012). ISBN 978-0-12-394424-5
Samsonovich, A.V.: A virtual actor behavior model based on emotional biologically inspired cognitive architecture. In: Goertzel, B., Iklé, M., Potapov, A. (eds.) AGI 2021. LNCS, vol. 13154, pp. 221–227. Springer, Cham (2022)
Samsonovich, A., Dodonov, A., Klychkov, M., Budanitsky, A., Grishin, I., Anisimova, A.: A virtual clown behavior model based on emotional biologically inspired cognitive architecture. In: Kryzhanovsky, B., Dunin-Barkowski, W., Redko, V., Tiumentsev, Y., Klimov, V.V. (eds.) NEUROINFORMATICS 2021. SCI, vol. 1008, pp. 99–108. Springer, Cham (2022). https://doi.org/10.1007/978-3-030-91581-0_14
Berman, A., James, V.: Kinetic imaginations: exploring the possibilities of combining AI and dance. In: Proceedings of the 24th International Joint Conference on Artificial Intelligence, pp. 2431–2437 (2015)
Deng, L.Q., Leung, H., Gu, N.J., Yang, Y.: Real-time mocap dance recognition for an interactive dancing game. Comput. Animation Virtual Worlds 22(2–3), 229–237 (2011). https://doi.org/10.1002/cav.397
Ho, E.S.L., Chan, J.C.P., Komura, T., Leung, H.: Interactive partner control in close interactions for real-time applications. ACM Trans. Multimed. Comput. Commun. Appl. 9(3), 19 (2013). https://doi.org/10.1145/2487268.2487274
Holldampf, J., Peer, A., Buss, M.: Virtual partner for a haptic interaction task. In: Ritter, H., Sagerer, G., Dillmann, R., Buss, M. (eds.) Human Centered Robot Systems. COSMOS, vol. 6, pp. 183–191. Springer, Heidelberg (2009). ISBN: 978-3-642-10402-2. https://doi.org/10.1007/978-3-642-10403-9_19
Kirakosian, S., Maravelakis, E., Mania, K., and IEEE: Immersive simulation and training of person-to-3D character dance in real-time, pp. 170–173 (2019). ISBN: 978-1-7281-4540-2
Mousas, C.: Performance-driven dance motion control of a virtual partner character, pp. 57–64. IEEE (2018). ISBN: 978-1-5386-3365-6
Senecal, S., Nijdam, N.A., Aristidou, A., Magnenat-Thalmann, N.: Salsa dance learning evaluation and motion analysis in gamified virtual reality environment. Multimedia Tools Appl. 79(33–34), 24621–24643 (2020). https://doi.org/10.1007/s11042-020-09192-y
Tamborini, R., et al.: The effect of behavioral synchrony with black or white virtual agents on outgroup trust. Comput. Hum. Behav. 83, 176–183 (2018). https://doi.org/10.1016/j.chb.2018.01.037
Tsampounaris, G., El Raheb, K., Katifori, V., Ioannidis, Y.: Exploring visualizations in real-time motion capture for dance education. Association for Computing Machinery (2016). https://doi.org/10.1145/3003733.3003811
Yokoyama, R., Sugiura, M., Yamamoto, Y., Nejad, K.K., Kawashima, R.: Neural bases of the adaptive mechanisms associated with reciprocal partner choice. Neuroimage 145, 74–81 (2017). https://doi.org/10.1016/j.neuroimage.2016.09.052
Samsonovich, A.V., Ascoli, G.A.: Augmenting weak semantic cognitive maps with an “abstractness” dimension. Comput. Intell. Neurosci. 2013, 308176 (2013). https://doi.org/10.1155/2013/308176
Samsonovich, A.V., Ascoli, G.A.: The conscious self: ontology, epistemology and the mirror quest. Cortex 41(5), 621–636 (2005). https://doi.org/10.1016/S0010-9452(08)70280-6
Acknowledgments
This work was supported by the Ministry of Science and Higher Education of the Russian Federation, state assignment project No. 0723-2020-0036.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Kuzmin, A.I., Semyonov, D.A., Samsonovich, A.V. (2022). Classification and Generation of Virtual Dancer Social Behaviors Based on Deep Learning in a Simple Virtual Environment Paradigm. In: Klimov, V.V., Kelley, D.J. (eds) Biologically Inspired Cognitive Architectures 2021. BICA 2021. Studies in Computational Intelligence, vol 1032. Springer, Cham. https://doi.org/10.1007/978-3-030-96993-6_23
Download citation
DOI: https://doi.org/10.1007/978-3-030-96993-6_23
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-96992-9
Online ISBN: 978-3-030-96993-6
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)