skip to main content
10.1145/1027933.1027957acmconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
Article

When do we interact multimodally?: cognitive load and multimodal communication patterns

Published:13 October 2004Publication History

ABSTRACT

Mobile usage patterns often entail high and fluctuating levels of difficulty as well as dual tasking. One major theme explored in this research is whether a flexible multimodal interface supports users in managing cognitive load. Findings from this study reveal that multimodal interface users spontaneously respond to dynamic changes in their own cognitive load by shifting to multimodal communication as load increases with task difficulty and communicative complexity. Given a flexible multimodal interface, users' ratio of multimodal (versus unimodal) interaction increased substantially from 18.6% when referring to established dialogue context to 77.1% when required to establish a new context, a +315% relative increase. Likewise, the ratio of users' multimodal interaction increased significantly as the tasks became more difficult, from 59.2% during low difficulty tasks, to 65.5% at moderate difficulty, 68.2% at high and 75.0% at very high difficulty, an overall relative increase of +27%. Analysis of users' task-critical errors and response latencies across task difficulty levels increased systematically and significantly as well, corroborating the manipulation of cognitive processing load. The adaptations seen in this study reflect users' efforts to self-manage limitations on working memory when task complexity increases. This is accomplished by distributing communicative information across multiple modalities, which is compatible with a cognitive load theory of multimodal interaction. The long-term goal of this research is the development of an empirical foundation for proactively guiding flexible and adaptive multimodal system design.

References

  1. Almor, A., Noun-phrase anaphora and focus: The informational load hypothesis. Psychological Review, 1999. 106: 748--765.]]Google ScholarGoogle Scholar
  2. Baddeley, A., Working Memory. Science, 1992. 255: 556--559.]]Google ScholarGoogle Scholar
  3. Benoit, C., J.-C. Martin, C. Pelachaud, L. Schomaker, & B. Suhm, Audio-visual and multimodal speech-based systems, Handbook of Multimodal and Spoken Dialogue Systems: Resources, Terminology and Product Evaluation, R. Moore, ed. 2000, Kluwer Academic Publishers: Boston, MA. 102--203.]]Google ScholarGoogle Scholar
  4. Calvert, G., C. Spence, & B.E. Stein, eds. The handbook of multisensory processing. 2004, MIT Press: Cambridge, MA.]]Google ScholarGoogle Scholar
  5. Chandler, P. & J. Sweller, Cognitive load theory and the format of instruction. Cognition and Instruction, 1991. 8: 293--332.]]Google ScholarGoogle Scholar
  6. Grant, K.W. & S. Greenberg. Speech intelligibility derived from asynchronous processing of auditory-visual information. Workshop on Audio-Visual Speech Processing (AVSP-2001). 2001. Scheelsminde, Denmark]]Google ScholarGoogle Scholar
  7. Grice, H.P., Logic and conversation, Syntax and Semantics: Speech Acts, J.L. Morgan, ed. 1975, Acad Press: NY. 41--58.]]Google ScholarGoogle Scholar
  8. Hinckley, K., Pierce J., E. Horvitz, & M. Sinclair, Foreground and background interaction with sensor-enhanced mobile devices. ACM Transactions on Computer Human Interaction, in press (Special Issue on Sensor-Based Interaction).]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Jacko, J., L. Barnard, T. Kongnakorn, K. Moloney, P. Edwards, V. Emery, & F. Sainfort. Isolating the effects of visual impairment: Exploring the effect of AMD on the utility of multimodal feedback. Conf. on Human Factors in Comp. Systems: CHI '04. 2004. NY, NY: ACM Press]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Jameson, A., Adaptive interfaces and agents, The human-computer interaction handbook, A. Sears, ed. 2003, Lawrence Erlbaum Associates: Mahwah NJ. 305--330.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Mayer, R.E. & R. Moreno, A split-attention effect in multimedia learning: evidence for dual processing systems in working memory. Journal of Educational Psychology, 1998. 90(2): 312--320.]]Google ScholarGoogle Scholar
  12. Mousavi, S.Y., R. Low, & J. Sweller, Reducing cognitive load by mixing auditory and visual presentation modes. Journal of Educational Psychology, 1995. 87(2): 319--334.]]Google ScholarGoogle Scholar
  13. Müller, C., B. Großmann-Hutter, A. Jameson, R. Rummer, & F. Wittig. Recognizing time pressure and cognitive load on the basis of speech: an experimental study. User Modeling. 2001: Springer]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Oviatt, S.L., Predicting spoken disfluencies during human-computer interaction. Computer Speech and Language, 1995. 9: 19--35.]]Google ScholarGoogle Scholar
  15. Oviatt, S.L., Multimodal interactive maps: Designing for human performance. Human Computer Interaction, 1997. 12(1-2): 93--129.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Oviatt, S.L., Ten myths of multimodal interaction. Communications of the ACM, 1999. 42(11): 74--81.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Oviatt, S.L. Mutual disambiguation of recognition errors in a multimodal architecture. ACM SIGCHI Conf. on Human Factors in Comp. Sys. (CHI'99). 1999. Pittsburgh, PA: ACM Press: 576--583.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Oviatt, S.L. Multimodal system processing in mobile environments. 13th ACM Symp. on User Interface Software Tech. (UIST'2000). 2000. New York: ACM Press: 21--30.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Oviatt, S.L., P.R. Cohen, L. Wu, J. Vergo, L. Duncan, B. Suhm, J. Bers, T.G. Holzman, T. Winograd, J. Landay, J. Larson, & D. Ferro, Designing the user interface for multimodal speech and gesture applications: State-of-the-art systems and research directions. Human Computer Interaction, 2000. 15(4): 263--322.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Oviatt, S.L., R. Coulston, S. Tomko, B. Xiao, R. Lunsford, M. Wesson, & L. Carmichael. Toward a theory of organized multimodal integration patterns during human-computer interaction. Internat. Conf. on Multimodal Interfaces. 2003. Vancouver, B.C.: ACM Press: 44--51.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Oviatt, S.L., T. Darrell, & M. Flickner, Multimodal Interfaces that flex, adapt, and persist, Comm. of the ACM. 2004. 30--33]]Google ScholarGoogle Scholar
  22. Penney, C.G., Modality effects and the structure of short-term verbal memory. Memory and Cognition, 1989. 17: 398--422.]]Google ScholarGoogle Scholar
  23. Prince, E., Toward a taxonomy of given-new information, Radical Pragmatics, P. Cole, ed. 1986, Academic: NY. 223--255.]]Google ScholarGoogle Scholar
  24. Sweller, J., Cognitive load during problem solving: Effects on learning. Cognitive Science, 1988. 12: 257--285.]]Google ScholarGoogle Scholar
  25. Technology for adaptive aging: Reports and papers. 2003, Nat. Acad. of Sci. Workshop: Nat. Acad. Press. http://www.nap.edu/books/0309091160/html/]]Google ScholarGoogle Scholar
  26. Teder-Sälejärvi, W.A., J.J. McDonald, F. Di Russo, & S.A. Hillyard, An analysis of audio-visual crossmodal integration by means of event-related potential (ERP) recordings. Cognitive Brain Research, 2002. 14: 106--114.]]Google ScholarGoogle Scholar
  27. Tindall-Ford, S., P. Chandler, & J. Sweller, When two sensory modes are better than one. Journal of Experimental Psychology: Applied, 1997. 3(3): 257--287.]]Google ScholarGoogle Scholar
  28. Wickens, C., Sandry, D., and Vidulich, M., Compatibility and resource competition between modalities of input, central processing, and output. Human Factors, 1983. 25(2): 227--248.]]Google ScholarGoogle Scholar
  29. Xiao, B., R. Lunsford, R. Coulston, M. Wesson, & S.L. Oviatt. Modeling multimodal integration patterns and performance in seniors: Toward adaptive processing of individual differences. Internat. Conf. on Multimodal Interfaces. 2003. Vancouver, BC: ACM Press: 265--272.]] Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. When do we interact multimodally?: cognitive load and multimodal communication patterns

                    Recommendations

                    Comments

                    Login options

                    Check if you have access through your login credentials or your institution to get full access on this article.

                    Sign in
                    • Published in

                      cover image ACM Conferences
                      ICMI '04: Proceedings of the 6th international conference on Multimodal interfaces
                      October 2004
                      368 pages
                      ISBN:1581139950
                      DOI:10.1145/1027933

                      Copyright © 2004 ACM

                      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

                      Publisher

                      Association for Computing Machinery

                      New York, NY, United States

                      Publication History

                      • Published: 13 October 2004

                      Permissions

                      Request permissions about this article.

                      Request Permissions

                      Check for updates

                      Qualifiers

                      • Article

                      Acceptance Rates

                      Overall Acceptance Rate453of1,080submissions,42%

                    PDF Format

                    View or Download as a PDF file.

                    PDF

                    eReader

                    View online with eReader.

                    eReader