Abstract
Manufacturing tools like 3D printers have become accessible to the wider society, making the promise of digital fabrication for everyone seemingly reachable. While the actual manufacturing process is largely automated today, users still require knowledge of complex design applications to produce ready-designed objects and adapt them to their needs or design new objects from scratch. To lower the barrier to the design and customization of personalized 3D models, we explored novice mental models in voice-based 3D modeling by conducting a high-fidelity Wizard of Oz study with 22 participants. We performed a thematic analysis of the collected data to understand how the mental model of novices translates into voice-based 3D modeling. We conclude with design implications for voice assistants. For example, they have to: deal with vague, incomplete and wrong commands; provide a set of straightforward commands to shape simple and composite objects; and offer different strategies to select 3D objects.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Billinghurst, M., Baldis, S., Matheson, L., Philips, M.: 3d palette: a virtual reality content creation tool. In: Proceedings of the ACM Symposium on Virtual Reality Software and Technology, VRST 1997, pp. 155–156. Association for Computing Machinery, Lausanne, Switzerland (1997). https://doi.org/10.1145/261135.261163
Bonner, S.E.: A model of the effects of audit task complexity. Acc. Organ. Soc. 19(3), 213–234 (1994). https://doi.org/10.1016/0361-3682(94)90033-7
Braun, M., Mainz, A., Chadowitz, R., Pfleging, B., Alt, F.: At your service: designing voice assistant personalities to improve automotive user interfaces. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, CHI 2019, pp. 1–11. Association for Computing Machinery, New York (2019). https://doi.org/10.1145/3290605.3300270
Braun, V., Clarke, V.: Using thematic analysis in psychology. Qual. Res. Psychol. 3(2), 77–101 (2006). https://doi.org/10.1191/1478088706qp063oa
Cuadra, A., Goedicke, D., Zamfirescu-Pereira, J.: Democratizing design and fabrication using speech: Exploring co-design with a voice assistant. In: CUI 2021–3rd Conference on Conversational User Interfaces, CUI 2021, Association for Computing Machinery, Bilbao (online), Spain (2021). https://doi.org/10.1145/3469595.3469624
Cuadra, A., Li, S., Lee, H., Cho, J., Ju, W.: My bad! repairing intelligent voice assistant errors improves interaction. Proc. ACM Hum.-Comput. Interact. 5(CSCW1) (2021). https://doi.org/10.1145/3449101
Davis, K.H., Biddulph, R., Balashek, S.: Automatic recognition of spoken digits. J. Acoustical Soc. Am. 24(6), 637–642 (1952). https://doi.org/10.1121/1.1906946
Etikan, I., Musa, S.A., Alkassim, R.S.: Comparison of convenience sampling and purposive sampling. Am. J. Theoret. Appli. Stat. 5(1), 1–4 (2016). https://doi.org/10.11648/j.ajtas.20160501.11
Fast, E., Chen, B., Mendelsohn, J., Bassen, J., Bernstein, M.S.: Iris: a conversational agent for complex tasks. CoRR abs/ ar**v: 1707.05015 (2017)
Feeman, S.M., Wright, L.B., Salmon, J.L.: Exploration and evaluation of cad modeling in virtual reality. Comput.-Aided Design Appli. 15(6), 892–904 (2018). https://doi.org/10.1080/16864360.2018.1462570
Fialho, P., Coheur, L.: Chatwoz: chatting through a wizard of oz. In: Proceedings of the 17th International ACM SIGACCESS Conference on Computers & Accessibility, ASSETS 2015, pp. 423–424. , Association for Computing Machinery, Lisbon, Portugal (2015). https://doi.org/10.1145/2700648.2811334
Friedrich, M., Langer, S., Frey, F.: Combining gesture and voice control for mid-air manipulation of cad models in vr environments. In: Proceedings of the 16th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, vol. 1: HUCAPP, pp. 119–127. INSTICC, SciTePress, Online Streaming (2021). https://doi.org/10.5220/0010170501190127
Gershenfeld, N.: How to make almost anything: the digital fabrication revolution. Foreign Aff. 91, 43 (2012)
Grigor, S., Nandra, C., Gorgan, D.: Voice-controlled 3d modelling with an intelligent personal assistant. Int. J. User-Syst. Interaction 13, 73–88 (2020). https://doi.org/10.37789/ijusi.2020.13.2.2
Grimes, G.M., Schuetzler, R.M., Giboney, J.S.: Mental models and expectation violations in conversational AI interactions. Decis. Support Syst. 144, 113515 (2021). https://doi.org/10.1016/j.dss.2021.113515
van Gumster, J.: Blender For Dummies, 2nd edn. For Dummies, USA (2011)
Huang, Y.-C., Chen, K.-L.: Brain-computer interfaces (BCI) based 3D computer-aided design (CAD): to improve the efficiency of 3D modeling for new users. In: Schmorrow, D.D., Fidopiastis, C.M. (eds.) AC 2017. LNCS (LNAI), vol. 10285, pp. 333–344. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-58625-0_24
Jain, M., Kumar, P., Kota, R., Patel, S.N.: Evaluating and informing the design of chatbots. In: Proceedings of the 2018 Designing Interactive Systems Conference, DIS 2018, pp. 895–906. Association for Computing Machinery, New York (2018). https://doi.org/10.1145/3196709.3196735
James, J., Watson, C.I., MacDonald, B.: Artificial empathy in social robots: an analysis of emotions in speech. In: Proceedings of the 27th IEEE International Symposium on Robot and Human Interactive Communication, RO-MAN 2018, pp. 632–637. IEEE Press, Nan**g, China (2018). https://doi.org/10.1109/ROMAN.2018.8525652
Jowers, I., Prats, M., McKay, A., Garner, S.: Evaluating an eye tracking interface for a two-dimensional sketch editor. Comput. Aided Des. 45(5), 923–936 (2013). https://doi.org/10.1016/j.cad.2013.01.006
Khan, S., Tunçer, B.: Gesture and speech elicitation for 3d cad modeling in conceptual design. Autom. Constr. 106, 102847 (2019). https://doi.org/10.1016/j.autcon.2019.102847
Khan, S., Tuncer, B., Subramanian, R., Blessing, L.: 3d cad modeling using gestures and speech: investigating cad legacy and non-legacy procedures. In: Lee, J.H. (ed.) Proceedings of the 18th International Conference on Computer Aided Architectural Design Futures, CAAD Futures 2019, pp. 347–366. CUMINCAD, Daejeon, Republic of Korea (2019). http://papers.cumincad.org/cgi-bin/works/paper/cf2019_042
Kou, X.Y., Tan, S.T.: Design by talking with computers. Comput.-Aided Design Appli. 5(1–4), 266–277 (2008). https://doi.org/10.3722/cadaps.2008.266-277
Kou, X., Xue, S., Tan, S.: Knowledge-guided inference for voice-enabled cad. Comput. Aided Des. 42(6), 545–557 (2010). https://doi.org/10.1016/j.cad.2010.02.002
Langevin, R., Lordon, R.J., Avrahami, T., Cowan, B.R., Hirsch, T., Hsieh, G.: Heuristic evaluation of conversational agents. In: Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, CHI 2021. Association for Computing Machinery, New York (2021). https://doi.org/10.1145/3411764.3445312
Lee, M., Billinghurst, M.: A wizard of oz study for an ar multimodal interface. In: Proceedings of the 10th International Conference on Multimodal Interfaces, ICMI 2008, pp. 249–256. Association for Computing Machinery, Chania, Crete, Greece (2008). https://doi.org/10.1145/1452392.1452444
Mayer, S., Laput, G., Harrison, C.: Enhancing mobile voice assistants with worldgaze. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, CHI 2020, pp. 1–10. Association for Computing Machinery, Honolulu, HI, USA (2020). https://doi.org/10.1145/3313831.3376479
Medhi Thies, I., Menon, N., Magapu, S., Subramony, M., O’Neill, J.: How do you want your Chatbot? an exploratory Wizard-of-Oz study with young, Urban Indians. In: Bernhaupt, R., Dalvi, G., Joshi, A., Balkrishan, D.K., O’Neill, J., Winckler, M. (eds.) INTERACT 2017. LNCS, vol. 10513, pp. 441–459. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-67744-6_28
Miller, G.A.: Wordnet: a lexical database for English. Commun. ACM 38(11), 39–41 (1995). https://doi.org/10.1145/219717.219748
Moore, R.J., Arar, R.: Conversational UX Design: A Practitioner’s Guide to the Natural Conversation Framework. Association for Computing Machinery, New York (2019)
Murad, C., Munteanu, C., Cowan, B.R., Clark, L.: Revolution or evolution? speech interaction and HCI design guidelines. IEEE Pervasive Comput. 18(2), 33–45 (2019). https://doi.org/10.1109/MPRV.2019.2906991
Murad, C., Munteanu, C., Cowan, B.R., Clark, L.: Finding a new voice: transitioning designers from gui to vui design. In: CUI 2021–3rd Conference on Conversational User Interfaces, CUI 2021. Association for Computing Machinery, New York (2021). https://doi.org/10.1145/3469595.3469617
Nanjundaswamy, V.G., et al.: Intuitive 3D computer-aided design (CAD) system with multimodal interfaces. In: Proceedings of the ASME 2013 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, vol. 2A: 33rd Computers and Information in Engineering Conference. ASME, Portland, Oregon, USA (08 2013). https://doi.org/10.1115/DETC2013-12277, v02AT02A037
Nielsen, J.: 10 usability heuristics for user interface design (1994). https://www.nngroup.com/articles/ten-usability-heuristics/
Nielsen, J.: Enhancing the explanatory power of usability heuristics. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI 1994, pp. 152–158. Association for Computing Machinery, Boston, Massachusetts, USA (1994). https://doi.org/10.1145/191666.191729
Niu, H., Van Leeuwen, C., Hao, J., Wang, G., Lachmann, T.: Multimodal natural human-computer interfaces for computer-aided design: A review paper. Appl. Sci. 12(13), 6510 (2022)
Nowacki, C., Gordeeva, A., Lizé, A.H.: Improving the usability of voice user interfaces: a new set of ergonomic criteria. In: Marcus, A., Rosenzweig, E. (eds.) Design, User Experience, and Usability. Design for Contemporary Interactive Environments, pp. 117–133. Springer International Publishing, Cham (2020)
Plumed, R., González-Lluch, C., Pérez-López, D., Contero, M., Camba, J.D.: A voice-based annotation system for collaborative computer-aided design. J. Comput. Design Eng. 8(2), 536–546 (2021). https://doi.org/10.1093/jcde/qwaa092
Sabharwal, N., Agrawal, A.: Introduction to Google Dialogflow, pp. 13–54. Apress, Berkeley, CA (2020). https://doi.org/10.1007/978-1-4842-5741-8_2
Samad, T., Director, S.W.: Towards a natural language interface for cad. In: Proceedings of the 22nd ACM/IEEE Design Automation Conference, pp. 2–8. DAC ’85, IEEE Press, Las Vegas, Nevada, USA (1985)
Seaborn, K., Miyake, N.P., Pennefather, P., Otake-Matsuura, M.: Voice in human-agent interaction: a survey. ACM Comput. Surv. (CSUR) 54(4), 1–43 (2021)
Sharp, H., Rogers, Y., Preece, J.: Interaction Design: Beyond Human Computer Interaction. John Wiley & Sons Inc., Hoboken, NJ, USA (2007)
Shum, H.Y., Han, M., Szeliski, R.: Interactive construction of 3d models from panoramic mosaics. In: Proceedings. 1998 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No.98CB36231), pp. 427–433. IEEE, Santa Barbara, California (1998). https://doi.org/10.1109/CVPR.1998.698641
Sree Shankar, S., Rai, R.: Human factors study on the usage of BCI headset for 3d cad modeling. Comput.-Aided Design 54, 51–55 (2014). https://doi.org/10.1016/j.cad.2014.01.006, application of brain-computer interfaces in cad/e systems
Stark, R., Israel, J., Wöhler, T.: Towards hybrid modelling environments–merging desktop-cad and virtual reality-technologies. CIRP Ann. 59(1), 179–182 (2010). https://doi.org/10.1016/j.cirp.2010.03.102
Thakur, A., Rai, R.: User study of hand gestures for gesture based 3D CAD modeling. In: Proceedings of the ASME 2015 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, vol. 1B: 35th Computers and Information in Engineering Conference. ASME, Boston, Massachusetts, USA (Aug 2015). https://doi.org/10.1115/DETC2015-46086, v01BT02A017
Voice2CAD: Voice2CAD (2022). https://voice2cad.com/
Völkel, S.T., Buschek, D., Eiband, M., Cowan, B.R., Hussmann, H.: Eliciting and analysing users’ envisioned dialogues with perfect voice assistants. In: Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, CHI 2021. Association for Computing Machinery, Yokohama, Japan (2021). https://doi.org/10.1145/3411764.3445536
Vtyurina, A., Fourney, A.: Exploring the role of conversational cues in guided task support with virtual assistants. In: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, CHI 2018, pp. 1–7. Association for Computing Machinery, Montreal QC, Canada (2018). https://doi.org/10.1145/3173574.3173782
Vuletic, T., et al.: A novel user-based gesture vocabulary for conceptual design. Int. J. Hum Comput Stud. 150, 102609 (2021). https://doi.org/10.1016/j.ijhcs.2021.102609
Vyas, S., Chen, T.J., Mohanty, R., Krishnamurthy, V.: Making-a-scene: a preliminary case study on speech-based 3D shape exploration through scene modeling. J. Comput. Inform. Sci. Eng. 1–11 (2022). https://doi.org/10.1115/1.4055239
Xue, S., Kou, X., Tan, S.: Natural voice-enabled cad: modeling via natural discourse. Comput.-Aided Design Appli. 6(1), 125–136 (2009)
Xue, S., Kou, X., Tan, S.: Command search for CAD system. Comput.-Aided Design Appli. 7(6), 899–910 (2010)
Yoon, S.M., Graf, H.: Eye tracking based interaction with 3d reconstructed objects. In: Proceedings of the 16th ACM International Conference on Multimedia, MM 2008, pp. 841–844. Association for Computing Machinery, Vancouver, British Columbia, Canada (2008). https://doi.org/10.1145/1459359.1459501
Acknowledgements
This work has been funded by the European Union’s Horizon 2020 research and innovation program under grant agreement No. 952026 (https://www.humane-ai.eu/). The research of Andrea Esposito is funded by a Ph.D. fellowship within the framework of the Italian “D.M. n. 352, April 9, 2022’ - under the National Recovery and Resilience Plan, Mission 4, Component 2, Investment 3.3 - Ph.D. Project “Human-Centered Artificial Intelligence (HCAI) techniques for supporting end users interacting with AI systems”, co-supported by “Eusoft S.r.l.” (CUP H91I22000410007).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
1 Electronic supplementary material
Below is the link to the electronic supplementary material.
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Desolda, G., Esposito, A., Müller, F., Feger, S. (2023). Digital Modeling for Everyone: Exploring How Novices Approach Voice-Based 3D Modeling. In: Abdelnour Nocera, J., Kristín Lárusdóttir, M., Petrie, H., Piccinno, A., Winckler, M. (eds) Human-Computer Interaction – INTERACT 2023. INTERACT 2023. Lecture Notes in Computer Science, vol 14145. Springer, Cham. https://doi.org/10.1007/978-3-031-42293-5_11
Download citation
DOI: https://doi.org/10.1007/978-3-031-42293-5_11
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-42292-8
Online ISBN: 978-3-031-42293-5
eBook Packages: Computer ScienceComputer Science (R0)