direkt zum Inhalt springen

direkt zum Hauptnavigationsmenü

Sie sind hier

TU Berlin

Page Content

Reviewed Journal Papers

go back to overview

Modeling Input Modality Choice in Mobile Graphical and Speech Interfaces
Citation key schaffer2015a
Author Schaffer, Stefan and Schleicher, Robert and Möller, Sebastian
Pages 21–34
Year 2015
ISSN 1071-5819
DOI 10.1016/j.ijhcs.2014.11.004
Journal Int. Journal of Human-Computer Studies
Volume 73
Number 3
Month mar
Note print/online
Abstract In this paper, we review three experiments with a mobile application that integrates graphical input with a touch-screen and a speech interface and develop a model for input modality choice in multimodal interaction. The model aims to enable simulation of multimodal human–computer interaction for automatic usability evaluation. The experimental results indicate that modality efficiency and input performance are important moderators of modality choice. Accordingly, we establish a utility-driven model that provides probability estimations of modality usage, based on the parameters of modality efficiency and input performance. Four variants of the model that differ in training data are fitted by means of Sequential Least Squares Programming. The analysis reveals a considerable fit regarding averaged modality usage. When applied to individual modality usage profiles, the accuracy decreases significantly. In an application example it is shown how the modality choice mechanism can be deployed for simulating interaction in the field of automatic usability evaluation. Results and possible limitations are discussed.
Link to publication Link to original publication Download Bibtex entry

go back to overview

Zusatzinformationen / Extras

Quick Access:

Schnellnavigation zur Seite über Nummerneingabe

Auxiliary Functions