• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Individual differences in multimodal integration patterns: What are they and why do they exist?” (2005)

by S L Oviatt, R Lunsford, R Coulston
Add To MetaCart

Tools

Sorted by:
Results 1 - 10 of 10

Multimodal human computer interaction: A survey

by Alejandro Jaimes, Nicu Sebe , 2005
"... In this paper we review the major approaches to Multimodal Human Computer Interaction, giving an overview of the field from a computer vision perspective. In particular, we focus on body, gesture, gaze, and affective interaction (facial expression recognition and emotion in audio). We discuss user ..."
Abstract - Cited by 119 (3 self) - Add to MetaCart
In this paper we review the major approaches to Multimodal Human Computer Interaction, giving an overview of the field from a computer vision perspective. In particular, we focus on body, gesture, gaze, and affective interaction (facial expression recognition and emotion in audio). We discuss user and task modeling, and multimodal fusion, highlighting challenges, open issues, and emerging applications for Multimodal Human Computer Interaction (MMHCI) research.
(Show Context)

Citation Context

...h may be available within several channels. No investigator has yet explored this possibility or the possibility that different individuals may typically attend to different types of information (see =-=[122]-=- for a recent study on this topic). Another important issue is the affective aspect of communication that should be considered when designing an MMHCI system. Emotion modulates almost all modes of hum...

Multimodal interfaces: Challenges and perspectives

by Nicu Sebe - Journal of Ambient Intelligence and Smart Environments , 2009
"... Abstract. The development of interfaces has been a technology-driven process. However, the newly developed multimodal interfaces are using recognition-based technologies that must interpret human-speech, gesture, gaze, movement patterns, and other behavioral cues. As a result, the interface design ..."
Abstract - Cited by 11 (0 self) - Add to MetaCart
Abstract. The development of interfaces has been a technology-driven process. However, the newly developed multimodal interfaces are using recognition-based technologies that must interpret human-speech, gesture, gaze, movement patterns, and other behavioral cues. As a result, the interface design requires a human-centered approach. In this paper we review the major approaches to multimodal Human Computer Interaction, giving an overview the user and task modeling, and to the multimodal fusion. We highlight the challenges, open issues, and the future trends in multimodal interfaces research.

Speech Pen: Predictive handwriting based on ambient multimodal recognition

by Kazutaka Kurihara, Masataka Goto - Proceedings of the ACM Conference on Human Factors in Computing Systems (CHI '05 , 2005
"... It is tedious to handwrite long passages of text by hand. To make this process more efficient, we propose predictive handwriting that provides input predictions when the user writes by hand. A predictive handwriting system presents possible next words as a list and allows the user to select one to s ..."
Abstract - Cited by 8 (0 self) - Add to MetaCart
It is tedious to handwrite long passages of text by hand. To make this process more efficient, we propose predictive handwriting that provides input predictions when the user writes by hand. A predictive handwriting system presents possible next words as a list and allows the user to select one to skip manual writing. Since it is not clear if people are willing to use prediction, we first run a user study to compare handwriting and selecting from the list. The result shows that, in Japanese, people prefer to select, especially when the expected performance gain from using selection is large. Based on these observations, we designed a multimodal input system, called speech-pen, that assists digital writing during lectures or presentations with background speech and handwriting recognition. The system recognizes speech and handwriting in the background and provides the instructor with predictions for further writing. The speech-pen system also allows the sharing of context information for predictions among the instructor and the audience; the result of the instructor’s speech recognition is sent to the audience to support their own note-taking. Our preliminary study shows the effectiveness of this system and the implications for further improvements. Author Keywords Predictive handwriting, speech recognition, handwriting
(Show Context)

Citation Context

...of a prior user study. We also decided to always show the latest result of speech recognition at the bottom of the screen because we found that the instructor often writes what he is speaking. Oviatt =-=[28]-=- reported the existence of individual differences in the order of input modalities when multimodal interfaces are used. Some people tend to speak and write sequentially, and some people prefer to spea...

Multi-modal text entry and selection on a mobile device

by David Dearman, Amy Karlson, Brian Meyers, Ben Bederson - In Proc. GI 2010, ACM , 2010
"... Rich text tasks are increasingly common on mobile devices, requiring the user to interleave typing and selection to produce the text and formatting she desires. However, mobile devices are a rich input space where input does not need to be limited to a keyboard and touch. In this paper, we present t ..."
Abstract - Cited by 6 (1 self) - Add to MetaCart
Rich text tasks are increasingly common on mobile devices, requiring the user to interleave typing and selection to produce the text and formatting she desires. However, mobile devices are a rich input space where input does not need to be limited to a keyboard and touch. In this paper, we present two complementary studies evaluating four different input modalities to perform selection in support of text entry on a mobile device. The modalities are: screen touch (Touch), device tilt (Tilt), voice recognition (Speech), and foot tap (Foot). The results show that Tilt is the fastest method for making a selection, but that Touch allows for the highest overall text throughput. The Tilt and Foot methods—although fast—resulted in users performing and subsequently correcting a high number of text entry errors, whereas the number of errors for Touch is significantly lower. Users experienced significant difficulty when using Tilt and Foot in coordinating the format selections in parallel with the text entry. This difficulty resulted in more errors and therefore lower text throughput. Touching the screen to perform a selection is slower than tilting the device or tapping the foot, but the action of moving the fingers off the keyboard to make a selection ensured high precision when interleaving selection and text entry. Additionally, mobile devices offer a breadth of promising rich input methods that need to be careful studied in situ when deciding if each is appropriate to support a given task; it is not sufficient to study the modalities independent of a natural task.
(Show Context)

Citation Context

...he performance of these alternate input modes will provide a better understanding of how multimodal selection and text entry can be integrated [12] and how to best support user‘s integration patterns =-=[13]-=-. 3 STUDY We performed two studies in order to explore human capabilities while performing selections during text entry. Four input modes were compared: standard screen touching (Touch), device tiltin...

Analysis and semantic modeling of modality preferences in industrial human-robot interaction

by Stefan Profanter, Er Perzylo, Nikhil Somani, Markus Rickert, Alois Knoll - in IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS , 2015
"... Abstract — Intuitive programming of industrial robots is es-pecially important for small and medium-sized enterprises. We evaluated four different input modalities (touch, gesture, speech, 3D tracking device) regarding their preference, usability, and intuitiveness for robot programming. A Wizard-of ..."
Abstract - Cited by 3 (2 self) - Add to MetaCart
Abstract — Intuitive programming of industrial robots is es-pecially important for small and medium-sized enterprises. We evaluated four different input modalities (touch, gesture, speech, 3D tracking device) regarding their preference, usability, and intuitiveness for robot programming. A Wizard-of-Oz experiment was conducted with 30 partic-ipants and its results show that most users prefer touch and gesture input over 3D tracking device input, whereas speech input was the least preferred input modality. The results also indicate that there are gender specific differences for preferred input modalities. We show how the results of the user study can be formalized in a semantic description language in such a way that a cognitive robotic workcell can benefit from the additional knowledge of input and output modalities, task parameter types, and preferred combinations of the two. I.
(Show Context)

Citation Context

...t have shown that multimodal systems are preferred by users over their unimodal alternatives, resulting in higher flexibility and reliability while better meeting the needs of a variety of users [2], =-=[3]-=-. Such systems provide multiple advantages, including improved efficiency, alternating interaction techniques, and accommodation of individual differences, such as permanent or temporary handicaps [4]...

Towards Simulating Humans in Augmented Multi-party Interaction

by Anton Nijholt
"... Human-computer interaction requires modeling of the user. A user profile typically contains preferences, interests, characteristics, and interaction behavior. However, in its multimodal interaction with a smart environment the user displays characteristics that show how the user, not necessarily con ..."
Abstract - Add to MetaCart
Human-computer interaction requires modeling of the user. A user profile typically contains preferences, interests, characteristics, and interaction behavior. However, in its multimodal interaction with a smart environment the user displays characteristics that show how the user, not necessarily consciously, verbally and nonverbally provides the smart environment with useful input and feedback. Especially in ambient intelligence environments we encounter situations where the environment supports interaction between the environment, smart objects (e.g., mobile robots, smart furniture) and human participants in the environment. Therefore it is useful for the profile to contain a physical representation of the user obtained by multi-modal capturing techniques. We discuss the modeling and simulation of interacting participants in the European
(Show Context)

Citation Context

...hen the system the user interacts withsallows multimodality then moresinformation about the user can be extractedsin real-time. For example, the system mayslearn about interaction pattern preferencess=-=[1]-=- or detect the user’s emotional state andsadapt its interaction behavior, its interfacesand its feedback according to them. Thesbody and what the user is doing with his orsher body is becoming importa...

Masataka Goto

by Kazutaka Kurihara
"... m.goto ..."
Abstract - Add to MetaCart
Abstract not found

Getting rid of “OK Google”: Individual Multimodal Input Adaption in Real World Applications

by Frank Honold, Nikola Bubalo, Anke Huckauf, Michael Weber
"... Abstract. Multimodal Interaction has the potential to significantly increase the ease of use of human computer interaction (HCI). At the same time, due to error-prone recognition based inputs, it is merely used in real-world applications. While literature on multimodal input fusion describes modelin ..."
Abstract - Add to MetaCart
Abstract. Multimodal Interaction has the potential to significantly increase the ease of use of human computer interaction (HCI). At the same time, due to error-prone recognition based inputs, it is merely used in real-world applications. While literature on multimodal input fusion describes modeling of different user behaviors as a key for increased robustness, it still failed to prove it’s practical use. This article presents the design of a user study that applies previous theoretical work on individual user adaption in a smartwatch scenario. We describe the practical implementation of a process for error recognition and recovery based on the history of multimodal inputs and a concrete scenario suitable for evaluating its practical impact on the ease of use. This could prove the real-world use of individual multimodal input adaption and finally lead to multimodal systems less cumbersome than today.
(Show Context)

Citation Context

...vidual User Adaption of Input Fusion Research from Oviatt et al. and others repeatedly showed the existence of different user behaviors in terms of temporal relations between multimodal constructions =-=[12, 11, 13, 6, 7, 9, 3, 5]-=-. These are called integration patterns and describe the dominant use of modalities in simultaneous or sequential temporal manner (cf. [10]). Twomodalities are used sequentially, when their time inter...

Multimodal

by Alexander Perzylo Nikhil Somani
"... binding of parameters for task-based robot programming based on semantic descriptions of modalities and parameter types ..."
Abstract - Add to MetaCart
binding of parameters for task-based robot programming based on semantic descriptions of modalities and parameter types
(Show Context)

Citation Context

... be interpreted as a formal evaluation. II. RELATED WORK Using multimodal input technologies for interacting with a system has various advantages resulting in a more flexible and reliable system [1], =-=[2]-=-. An approach for programming industrial robots using markerless gesture recognition is evaluated in [3]. The authors conclude that using a multimodal system results in a significant reduction of requ...

MMWA-ae: boosting knowledge from Multimodal Interface Design, Reuse and Usability Evaluation

by Americo Talarico Neto, Renata Pontin, M. Fortes, Rafael Rossi, Solange Rezende
"... The technological progress designing new de-vices and the scientific growth in the field of Human-Computer Interaction are enabling new in-teraction modalities to move from research to com-mercial products. However, developing multimodal interfaces is still a difficult task due to the lack of tools ..."
Abstract - Add to MetaCart
The technological progress designing new de-vices and the scientific growth in the field of Human-Computer Interaction are enabling new in-teraction modalities to move from research to com-mercial products. However, developing multimodal interfaces is still a difficult task due to the lack of tools that consider not only code generation, but usability of such interfaces. In this paper, we present the MultiModal Web Approach and its au-thoring environment, whose main goal is boosting the dissemination of project knowledge for future developments benefited by the solutions to recur-ring problems in that multimodal context. 1.
(Show Context)

Citation Context

...ing availability of Web services to the worldwide population and due to the great expressive power, naturalness and portability that multimodal interfaces offer the users to perform their daily tasks =-=[9]-=-. However, the use of combined modalities such as voice, touch and graph, raises a number of usability and interaction issues that the project team is faced with, like syncronization and integration r...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University