Oznur Alkan, Massimilliano Mattetti, et al.
INFORMS 2020
Multimodal user interfaces allow users to interact with computers through multiple modalities, such as speech, gesture, and gaze. To be effective, multimodal user interfaces must correctly identify all objects which users refer to in their inputs. To systematically resolve different types of references, we have developed a probabilistic approach that uses a graph-matching algorithm. Our approach identifies the most probable referents by optimizing the satisfaction of semantic, temporal, and contextual constraints simultaneously. Our preliminary user study results indicate that our approach can successfully resolve a wide variety of referring expressions, ranging from simple to complex and from precise to ambiguous ones. Copyright 2004 ACM.
Oznur Alkan, Massimilliano Mattetti, et al.
INFORMS 2020
Casey Dugan, Werner Geyer, et al.
CHI 2010
Rajesh Balchandran, Leonid Rachevsky, et al.
INTERSPEECH 2009
Seyed Omid Sadjadi, Jason W. Pelecanos, et al.
INTERSPEECH 2014