Authors
Michael Johnston, Philip R Cohen, David McGee, Sharon Oviatt, James A Pittman, Ira Smith
Publication date
1997/7
Conference
35th Annual Meeting of the Association for Computational Linguistics and 8th Conference of the European Chapter of the Association for Computational Linguistics
Pages
281-288
Description
Recent empirical research has shown conclusive advantages of multimodal interaction over speech-only interaction for mapbased tasks. This paper describes a multimodal language processing architecture which supports interfaces allowing simultaneous input from speech and gesture recognition. Integration of spoken and gestural input is driven by unification of typed feature structures representing the semantic contributions of the different modes. This integration method allows the component modalities to mutually compensate for each others' errors. It is implemented in Quick-Set, a multimodal (pen/voice) system that enables users to set up and control distributed interactive simulations.
Total citations
1997199819992000200120022003200420052006200720082009201020112012201320142015201620172018201920202021202220232024115182462916191822182323813137109594551211
Scholar articles
M Johnston, PR Cohen, D McGee, S Oviatt, JA Pittman… - 35th Annual Meeting of the Association for …, 1997