Abstract
In this work, we present the preliminary results from an exploratory experiment in which we studied the influence of exposing students to a personal emotional score on their classroom learning. This score derived from deep learning algorithms is presented to students on a computer screen while they are learning in a classroom. With the advent of deep learning models for the classification of emotions, it is now possible to gauge emotions in real-time. Most of these algorithms depend on facial expressions to identify human emotional and attentional levels. Such algorithms are even being integrated within video conferencing platforms such as Zoom, Teams, etc. to quantify if the participants in the meeting are attentive. The assumption here is that our facial expressions reflect our emotions.
The hypotheses tested by this study is a) if such algorithms accurately reflect human emotions, and b) if the very act of showing such a metric to learners (students in a classroom) will influence their cognitive learning capabilities. We conducted the experiment in a freshman-year undergraduate class where students were asked to compile a bibliography, read the research papers and annotate them, and then write a summary. We used a state-of-the-art deep neural network trained on thousands of face images in multiple orientations to detect seven basic human emotions (neutral, happiness, sadness, anger, surprise, disgust, and fear) from the students’ facial expressions. Students’ facial images were captured in real-time through their laptop’s camera. The experimental group of students were presented with their emotional scores and labels while they were working in the class. The control group of students were not exposed to this information. After each class, we asked all students to do a self-assessment of scores and asked for the instructor’s assessment of the students’ work as well. We investigated if the self-perception of their own emotional states helped students to perform better. The key outcomes of this study are threefold. First, the AI models have not reached a stage yet where they could accurately reflect human emotions using facial expressions. Second, the study shows that the self-perception of emotions provide a boost to the perception of productivity in students. Finally, this study also reveals that a higher self-perception of productivity due to being exposed to the emotion recognition AI model do not necessarily contribute to actual higher productivity.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Adolphs, R.: Recognizing emotion from facial expressions: psychological and neurological mechanisms. Behav. Cogn. Neurosci. Rev. 1(1), 21–62 (2002). https://doi.org/10.1177/1534582302001001003
Aoki, H., Ohnishi, A., Isoyama, N., Terada, T., Tsukamoto, M.: FaceRecGlasses: a wearable system for recognizing self facial expressions using compact wearable cameras. In: 2021 Proceedings of the Augmented Humans International Conference, pp. 55–65 (2021). https://doi.org/10.1145/3458709.3458983
Ark, W.S., Dryer, D.C., Lu, D.J.: The Emotion Mouse. In: HCI, no. 1, pp. 818–823 (1999)
Barrett, L.F., Adolphs, R., Marsella, S., Martinez, A.M., Pollak, S.D.: Emotional expressions reconsidered: challenges to inferring emotion from human facial movements. Psychol. Sci. Public Interest 20(1), 1–68 (2019)
Buolamwini, J., Gebru, T.: Gender shades: intersectional accuracy disparities in commercial gender classification. In: Conference on Fairness, Accountability and Transparency, pp. 77–91. PMLR (2018)
Cordaro, D.T., Sun, R., Keltner, D., Kamble, S., Huddar, N., McNeil, G.: Universals and cultural variations in 22 emotional expressions across five cultures. Emotion 18(1), 75 (2018)
Cowie, R., et al.: Emotion recognition in human-computer interaction. IEEE Signal Process. Mag. 18(1), 32–80 (2001)
Craig, B.M., Lipp, O.V.: The influence of multiple social categories on emotion perception. J. Exp. Soc. Psychol. 75, 27–35 (2018)
De Silva, T.R.S., Dayananda, K.Y., Galagama Arachchi, R.C., Amerasekara, M.K.S.B., Silva, S., Gamage, N.: Solution to measure employee productivity with employee emotion detection. In: 2022 4th International Conference on Advancements in Computing (ICAC), pp. 210–215 (2022). https://doi.org/10.1109/ICAC57685.2022.10025132
D’Mello, S., Kappas, A., Gratch, J.: The affective computing approach to affect measurement. Emot. Rev. 10(2), 174–183 (2018)
Dryer, D.C.: Getting personal with computers: how to design personalities for agents. Appl. Artif. Intell. 13(3), 273–295 (1999)
Ekman, P.: Methods for measuring facial action. Handb. Methods Nonverbal Behav. Res. 45–90 (1982)
Ekman, P., et al.: Universals and cultural differences in the judgments of facial expressions of emotion. J. Pers. Soc. Psychol. 53(4), 712 (1987)
Emotional Expressions Reconsidered: Challenges to Inferring Emotion From Human Facial Movements. Association for Psychological Science - APS (n.d.). https://www.psychologicalscience.org/publications/emotional-expressions-reconsidered-challenges-to-inferring-emotion-from-human-facial-movements.html. Accessed 29 Nov 2023
Gadigi, V., Veerabhadrappa, H.: International research journal of engineering and technology (IRJET) a literature review on technique of facial variance to detect the state of emotion and map productivity. Int. Res. J. Eng. Sci. Technol. Innov. 7, 814–817 (2023)
Harley, J.M., Azevedo, R.: Toward a feature-driven understanding of students’ emotions during interactions with agent-based learning environments: a selective review. Int. J. Gaming Comput.-Mediat. Simul. (IJGCMS) 6(3), 17–34 (2014)
Jarrell, A., Harley, J.M., Lajoie, S., Naismith, L.: Success, failure and emotions: examining the relationship between performance feedback and emotions in diagnostic reasoning. Educ. Tech. Res. Dev. 65(5), 1263–1284 (2017)
Miranda, C.R., et al.: Facial expressions tracking and recognition: database protocols for systems validation and evaluation. ar**v preprint ar**v:1506.00925 (2015)
Moridis, C.N., Economides, A.A.: Mood recognition during online self-assessment tests. IEEE Trans. Learn. Technol. 2(1), 50–61 (2009)
Picard, R.W.: Affective computing for HCI. In: HCI, no. 1, pp. pp. 829–833 (1999)
Postareff, L., Mattsson, M., Lindblom-Ylänne, S., Hailikari, T.: The complex relationship between emotions, approaches to learning, study success and study progress during the transition to university. High. Educ. 73, 441–457 (2017)
Shen, L., Wang, M., Shen, R.: Affective e-learning: using “emotional” data to improve learning in pervasive learning environment. J. Educ. Technol. Soc. 12(2), 176–189 (2009)
Sundström, P.: Exploring the affective loop. Doctoral dissertation (2005)
**a, Y.: Upper, middle and lower region learning for facial action unit detection. ar**v preprint ar**v:2002.04023 (2020)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Prince, B., Siddharth, Joshi, V., Keshav, R. (2024). Investigating the Effect of Personal Emotional Score Display on Classroom Learning. In: Auer, M.E., Langmann, R., May, D., Roos, K. (eds) Smart Technologies for a Sustainable Future. STE 2024. Lecture Notes in Networks and Systems, vol 1028. Springer, Cham. https://doi.org/10.1007/978-3-031-61905-2_5
Download citation
DOI: https://doi.org/10.1007/978-3-031-61905-2_5
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-61904-5
Online ISBN: 978-3-031-61905-2
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)