277
Views
1
CrossRef citations to date
0
Altmetric
Research Articles

The Unit and Size of Information Supporting Auditory Feedback for Voice User Interface

ORCID Icon, ORCID Icon & ORCID Icon
Pages 3071-3080 | Received 29 Aug 2022, Accepted 07 Feb 2023, Published online: 19 Feb 2023

References

  • Amershi, S., Weld, D., Vorvoreanu, M., Fourney, A., Nushi, B., Collisson, P., Suh, J., Iqbal, S., Bennett, P. N., Inkpen, K., Teevan, J., Kikin-Gil, R., & Horvitz, E. (2019, May 2). Guidelines for human-AI interaction. In Conference on Human Factors in Computing Systems – Proceedings. https://doi.org/10.1145/3290605.3300233
  • Asabuki, T., Hiratani, N., & Fukai, T. (2018). Interactive reservoir computing for chunking information streams. PLOS Computational Biology, 14(10), e1006400. https://doi.org/10.1371/journal.pcbi.1006400
  • Aylett, M. P., Kristensson, P. O., Whittaker, S., & Vazquez-Alvarez, Y. (2014). None of a CHInd: Relationship counselling for HCI and speech technology. In CHI 2014 (pp. 749–760). https://doi.org/10.1145/2559206.2578868
  • Baddeley, A. D. (2000). Short-term and working memory. In The Oxford handbook of memory (Vol. 4, pp. 77–92). Oxford University Press.
  • Baddeley, A. D., & Hitch, G. (1974). Working memory. In Psychology of learning and motivation (Vol. 8, pp. 47–89). Elsevier. https://doi.org/10.1016/S0079-7421(08)60452-1
  • Baddeley, A. D., Thomson, N., & Buchanan, M. (1975). Word length and the structure of short-term memory. Journal of Verbal Learning and Verbal Behavior, 14(6), 575–589. https://doi.org/10.1016/S0022-5371(75)80045-4
  • Bickmore, T. W., Pfeifer, L. M., & Jack, B. W. (2009). Taking the time to care: Empowering low health literacy hospital patients with virtual nurse agents. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (pp. 1265–1274). https://doi.org/10.1007/978-3-642-12770-0_8
  • Cohen, M. H., Cohen, M. H., Giangola, J. P., & Balogh, J. (2004). Voice user interface design. Addison-Wesley Professional.
  • Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24(1), 87–114. https://doi.org/10.1017/S0140525X01003922
  • Dirlam, D. K. (1972). Most efficient chunk sizes. Cognitive Psychology, 3(2), 355–359. https://doi.org/10.1016/0010-0285(72)90012-6
  • Gilchrist, A. L., Cowan, N., & Naveh-Benjamin, M. (2009). Investigating the childhood development of working memory using sentences: New evidence for the growth of chunk capacity. Journal of Experimental Child Psychology, 104(2), 252–265. https://doi.org/10.1016/j.jecp.2009.05.006
  • Glanzer, M., & Razel, M. (1974). The size of the unit in short-term storage. Journal of Verbal Learning and Verbal Behavior, 13(1), 114–131. https://doi.org/10.1016/S0022-5371(74)80036-8
  • Gobet, F., Lane, P. C. R., Croker, S., Cheng, P. C. H., Jones, G., Oliver, I., & Pine, J. M. (2001). Chunking mechanisms in human learning. Trends in Cognitive Sciences, 5(6), 236–243. https://doi.org/10.1016/S1364-6613(00)01662-4
  • Hauptmann, A. G., & Rudnicky, A. I. (1990). A comparison of speech and typed input. In HLT '90: Proceedings of the Workshop on Speech and Natural Language (pp. 219–224). https://doi.org/10.3115/116580.116652
  • Howell, M., Love, S., & Turner, M. (2006). Visualisation improves the usability of voice-operated mobile phone services. International Journal of Human–Computer Studies, 64(8), 754–769. https://doi.org/10.1016/j.ijhcs.2006.03.002
  • Hua, Z., & Ng, W. L. (2010). Speech recognition interface design for in-vehicle system. In Proceedings of the 2nd International Conference on Automotive User Interfaces and Interactive Vehicular Applications – AutomotiveUI ’10, AutomotiveUI (p. 29). https://doi.org/10.1145/1969773.1969780
  • Jung, J., Lee, S., Hong, J., Youn, E., & Lee, G. (2020). Voice + Tactile: Augmenting in-vehicle voice user interface with tactile touchpad interaction. In Conference on Human Factors in Computing Systems – Proceedings (pp. 1–12). https://doi.org/10.1145/3313831.3376863
  • Kim, H. C., Cha, M. C., & Ji, Y. G. (2021). The impact of an agent’s voice in psychological counseling: Session evaluation and counselor rating. Applied Sciences, 11(7), 2893. https://doi.org/10.3390/app11072893
  • Kim, J., Jeong, M., & Lee, S. C. (2019). “Why did this voice agent not understand me?”: Error recovery strategy for in-vehicle voice user interface. In Adjunct Proceedings – 11th International ACM Conference on Automotive User Interfaces and Interactive Vehicular Applications, AutomotiveUI 2019 (pp. 146–150). https://doi.org/10.1145/3349263.3351513
  • Kwon, O. W., & Park, J. (2003). Korean large vocabulary continuous speech recognition with morpheme-based recognition units. Speech Communication, 39(3–4), 287–300. https://doi.org/10.1016/S0167-6393(02)00031-6
  • Lee, S. C., & Ji, Y. G. (2019). Complexity of in-vehicle controllers and their effect on task performance. International Journal of Human–Computer Interaction, 35(1), 65–74. https://doi.org/10.1080/10447318.2018.1428263
  • Lo, V. E. W., & Green, P. A. (2013). Development and evaluation of automotive speech interfaces: Useful information from the human factors and the related literature. International Journal of Vehicular Technology, 2013, 1–13. https://doi.org/10.1155/2013/924170
  • López, G., Quesada, L., & Guerrero, L. A. (2018). Alexa vs. Siri vs. Cortana vs. Google Assistant: A comparison of speech-based natural user interfaces. Advances in Intelligent Systems and Computing, 592, 241–250. https://doi.org/10.1007/978-3-319-60366-7_23
  • Mane, A., Boyce, S., Karis, D., & Yankelovich, N. (1996). Designing the user interface for speech recognition applications. In CHI '96: Conference Companion on Human Factors in Computing Systems (Vol. 28, p. 431). https://doi.org/10.1145/257089.257431
  • Miller, G. (1956). Human memory and the storage of information. IEEE Transactions on Information Theory, 2(3), 129–137. https://doi.org/10.1109/TIT.1956.1056815
  • Montgomery, J. W. (2004). Sentence comprehension in children with specific language impairment: Effects of input rate and phonological working memory. International Journal of Language & Communication Disorders, 39(1), 115–133. https://doi.org/10.1080/13682820310001616985
  • Park, J., Choi, H., & Jung, Y. (2021). Users’ cognitive and affective response to the risk to privacy from a smart speaker. International Journal of Human–Computer Interaction, 37(8), 759–771. https://doi.org/10.1080/10447318.2020.1841422
  • Pearl, C. (2016). Designing voice user interfaces: Principles of conversational experiences. O’Reilly Media, Inc. https://books.google.co.kr/books?hl=ko&lr=&id=MmnEDQAAQBAJ&oi=fnd&pg=PR11&dq=VUI+advantage+multitasking&ots=HNb-0ubBhd&sig=GkFodoix6KBzwMh-UC_yDYZaBhU#v=onepage&q=VUIadvantagemultitasking&f=false
  • Peissner, M., & Doebler, V. (2011). Can voice interaction help reducing the level of distraction and prevent accidents? Whitepaper, May 24.
  • Perlman, D., Samost, A., Domel, A. G., Mehler, B., Dobres, J., & Reimer, B. (2019). The relative impact of smartwatch and smartphone use while driving on workload, attention, and driving performance. Applied Ergonomics, 75(September 2018), 8–16. https://doi.org/10.1016/j.apergo.2018.09.001
  • Perugini, S., Anderson, T. J., & Moroney, W. F. (2007). A study of out-of-turn interaction in menu-based, IVR, voicemail systems. In Proceedings of ACM CHI 2007: Conference on Human Factors in Computing Systems (pp. 961–970).
  • Ranney, T. A., Harbluk, J. L., & Noy, Y. I. (2005). Effects of voice technology on test track driving performance: Implications for driver distraction. Human Factors, 47(2), 439–454. https://doi.org/10.1518/0018720054679515
  • Saariluomaand, P., & Jokinen, J. P. P. (2014). Emotional dimensions of user experience: A user psychological analysis. International Journal of Human–Computer Interaction, 30(4), 303–320. https://doi.org/10.1080/10447318.2013.858460
  • Sawan, P. B. D., Gopy, K., Hurry, G., & Gopaul, T. T. (2013). A study on smart home control system through speech. International Journal of Computer Applications, 69(19), 30–39. https://doi.org/10.5120/12080-8244
  • Shneiderman, B. (2000). The limits of speech recognition. Communications of the ACM, 43(9), 63–65. https://doi.org/10.1145/348941.348990
  • Strayer, D. L., Cooper, J. M., McCarty, M. M., Getty, D. J., Wheatley, C. L., Motzkus, C. J., Goethe, R. M., Biondi, F., & Horrey, W. J. (2019). Visual and cognitive demands of CarPlay, Android Auto, and Five Native Infotainment Systems. Human Factors, 61(8), 1371–1386. https://doi.org/10.1177/0018720819836575
  • Voicebot.ai. (2020a, January). In-car voice assistant consumer adoption report January 2020. Voicebot.ai. https://voicebot.ai/wp-content/uploads/2020/02/in_car_voice_assistant_consumer_adoption_report_2020_voicebot.pdf
  • Voicebot.ai. (2020b, April). Smart specker consumer adoption report executive summary April 2020. https://www.voice2shop.com/wp-content/uploads/2020/11/executive-summary-smart-speaker-consumer-adoption-report-2020.pdf
  • Watanabe, M., Okano, A., Asano, Y., & Ogawa, K. (2007). VoiceBlog: Universally designed voice browser. International Journal of Human–Computer Interaction, 23(1–2), 95–113. https://doi.org/10.1080/10447310701362975
  • Wickelgren, W. A. (1964). Size of rehearsal group and short-term memory. Journal of Experimental Psychology, 68(4), 413–419. https://doi.org/10.1037/h0043584
  • Wickens, C. D. (2002). Multiple resources and performance prediction. Theoretical Issues in Ergonomics Science, 3(2), 159–177. https://doi.org/10.1080/14639220210123806
  • Winsum, W. V., Martens, M., & Herland, L. (1999). The effects of speech versus tactile driver support messages on workload, driver behaviour and user acceptance. In TNO Human Factors. https://doi.org/10.13140/RG.2.1.1776.1041
  • Yager, C. (2013). An evaluation of the effectiveness of voice- to-text programs at reducing incidences of distracted driving (Report SWUTC/13/600451-00011-1). Texas A&M Transportation Institute. http://swutc.tamu.edu/publications/technicalreports/600451-00011-1.pdf
  • Yankelovich, N., Levow, G. A., & Marx, M. (1995). Designing speechActs: Issues in speech user interfaces. In Conference on Human Factors in Computing Systems – Proceedings (Vol. 1, pp. 369–376).

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.