Authors
S. L. Oviatt
Publication date
1999
Conference
Proceedings of the Conference on Human Factors in Computing Systems
Pages
576-583
Publisher
ACM Press
Description
As a new generation of multimodal/media systems begins to define itself, researchers are attempting to learn how to combine different modes into strategically integrated whole systems. In theory, well designed multimodal systems should be able to integrate complementary modalities in a manner that supports mutual disambiguation (MD) of errors and leads to more robust performance. In this study, over 2,000 multimodal utterances by both native and accented speakers of English were processed by a multimodal system, and then logged and analyzed. The results confirmed that multimodal systems can indeed support significant levels of MD, and also higher levels of MD for the more challenging accented users. As a result, although speech recognition as a stand-alone performed far more poorly for accented speakers, their multimodal recognition rates did not differ from those of native speakers. Implications are …
Total citations
199920002001200220032004200520062007200820092010201120122013201420152016201720182019202020212022202320246172133213333372926241515101661058820107263
Scholar articles