Abstract
Understanding the behavioral intentions and decision-making mechanisms of automated vehicles (AVs) is one of the reasons why people trust them. This study aims to explore the main effect of AV transparency on trust, situation awareness and mental workload, while considering whether different driving experiences and scenario types have a moderating effect on the above results. This study conducted a driving simulation experiment. The experiment adopted a three-factor mixed design with AV transparency (No Explanation, What-only, Why-only, What+Why) and driving experience (Not Rich, Rich) as between-subject variables, and scenario type (Expected, Unexpected) as within-subject variables. To balance the impact of the scenario presentation order in the experiment, half of the participants were exposed to the scenarios in order from expected to unexpected, and the other half were exposed to the scenarios in order from unexpected to expected. Froty-eight participants (24 females and 24 males) participated in the experiment. The research found that increased transparency information help people better perceive and understand AVs while having less workload. In addition, the research found that drivers with rich driving experience are less dependent on the information provided by the AVs. The study also found that drivers have higher trust and lower anxiety in AVs in expected scenarios.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Expected events referred to events in which the drivers can understand the information provided by AV based on the current perception of the surrounding environment; unexpected events referred to events in which the drivers cannot understand the information provided by AV based on the current perception of the surrounding environment; takeover events referred to events in which control of the vehicle had to be handed over to the driver due to the malfunction in AV.
- 2.
In SA1 and SA2, the correct answer was underlined.
- 3.
***p < .001, **p < .01, *p < .05.
References
Bhaskara, A., et al.: Effect of automation transparency in the management of multiple unmanned vehicles. Appl. Ergon. 90, 103243 (2021). https://doi.org/10.1016/j.apergo.2020.103243
Bhaskara, A., Skinner, M., Loft, S.: Agent transparency: a review of current theory and evidence. IEEE Trans. Hum.-Mach. Syst. 50(3), 215–224 (2020). https://doi.org/10.1109/THMS.2020.2965529
Committee, O.-R.A.D.: Taxonomy and definitions for terms related to driving automation systems for on-road motor vehicles. In: SAE International (2021)
Faber, K., van Lierop, D.: How will older adults use automated vehicles? Assessing the role of AVs in overcoming perceived mobility barriers. Transp. Res. Part A Pol. Pract. 133, 353–363 (2020). https://doi.org/10.1016/j.tra.2020.01.022
Hart, S.G., Staveland, L.E.: Development of NASA-TLX (Task Load Index): results of empirical and theoretical research. In: Hancock, P.A., Meshkati, N. (eds.) Advances in Psychology, vol. 52, pp. 139–183. North-Holland (1988). https://doi.org/10.1016/S0166-4115(08)62386-9
Hoffman, R.R., Mueller, S.T., Klein, G., Litman, J.: Metrics for explainable AI: challenges and prospects. ar**v preprint ar**v:1812.04608 (2018)
Holländer, K., Wintersberger, P., Butz, A.: Overtrust in external cues of automated vehicles: an experimental investigation. In: Proceedings of the 11th International Conference on Automotive User Interfaces and Interactive Vehicular Applications, Utrecht, Netherlands (2019). https://doi.org/10.1145/3342197.3344528
Koo, J., Kwac, J., Ju, W., Steinert, M., Leifer, L., Nass, C.: Why did my car just do that? Explaining semi-autonomous driving actions to improve driver understanding, trust, and performance. Int. J. Interact. Des. Manuf. (IJIDeM) 9(4), 269–275 (2015). https://doi.org/10.1007/s12008-014-0227-2
Körber, M.: Theoretical considerations and development of a questionnaire to measure trust in automation. In: Bagnara, S., Tartaglia, R., Albolino, S., Alexander, T., Fujita, Y. (eds.) IEA 2018. AISC, vol. 823, pp. 13–30. Springer, Cham (2019). https://doi.org/10.1007/978-3-319-96074-6_2
Lee, J.D., See, K.A.: Trust in automation: designing for appropriate reliance. Hum. Factors 46(1), 50–80 (2004). https://doi.org/10.1518/hfes.46.1.50_30392
Schaefer, K.E., Chen, J.Y.C., Szalma, J.L., Hancock, P.A.: A meta-analysis of factors influencing the development of trust in automation: implications for understanding autonomy in future systems. Hum. Factors 58(3), 377–400 (2016). https://doi.org/10.1177/0018720816634228
Seong, Y., Bisantz, A.M.: The impact of cognitive feedback on judgment performance and trust with decision aids. Int. J. Ind. Ergon. 38(7), 608–625 (2008). https://doi.org/10.1016/j.ergon.2008.01.007
Shin, D.: The effects of explainability and causability on perception, trust, and acceptance: implications for explainable AI. Int. J. Hum.-Comput. Stud. 146 (2021). https://doi.org/10.1016/j.ijhcs.2020.102551
Stephenson, A.C., et al.: Effects of an unexpected and expected event on older adults’ autonomic arousal and eye fixations during autonomous driving [original research]. Front. Psychol. 11 (2020). https://doi.org/10.3389/fpsyg.2020.571961
Tatasciore, M., Loft, S.: Can increased automation transparency mitigate the effects of time pressure on automation use? Appl. Ergon. 114, 104142 (2024). https://doi.org/10.1016/j.apergo.2023.104142
van de Merwe, K., Mallam, S., Nazir, S.: Agent transparency, situation awareness, mental workload, and operator performance: a systematic literature review. Hum. Fact. 00187208221077804 (2022). https://doi.org/10.1177/00187208221077804
Yan, Z., Kantola, R., Zhang, P.: a research model for human-computer trust interaction. In: 2011IEEE 10th International Conference on Trust, Security and Privacy in Computing and Communications (2011)
Zhang, T., Tao, D., Qu, X., Zhang, X., Lin, R., Zhang, W.: The roles of initial trust and perceived risk in public’s acceptance of automated vehicles. Transp. Res. Part C: Emer. Technol. 98, 207–220 (2019)
Acknowledgments
This work was supported by the Guangdong Basic and Applied Basic Research Foundation (grant number 2021A1515011610) and the Foundation of Shenzhen Science and Technology Innovation Committee (grant number JCYJ20210324100014040).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Huang, W., Chen, M., Li, W., Zhang, T. (2024). Effects of Automated Vehicles’ Transparency on Trust, Situation Awareness, and Mental Workload. In: Krömker, H. (eds) HCI in Mobility, Transport, and Automotive Systems. HCII 2024. Lecture Notes in Computer Science, vol 14732. Springer, Cham. https://doi.org/10.1007/978-3-031-60477-5_9
Download citation
DOI: https://doi.org/10.1007/978-3-031-60477-5_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-60476-8
Online ISBN: 978-3-031-60477-5
eBook Packages: Computer ScienceComputer Science (R0)