Samenvatting
Against the background of developments in the area of speech-based and multimodal interfaces, we present research on determining the addressee of an utterance in the context of mixed human-human and multimodal human-computer interaction. Working with data that are taken from realistic scenarios, we explore several features with respect to their relevance to the question who is the addressee of an utterance: eye gaze both of speaker and listener, dialogue history and utterance length. With respect to eye gaze, we inspect the detailed timing of shifts in eye gaze between different communication partners (human or computer). We show that these features result in an improved classification of utterances in terms of addressee-hood relative to a simple classification algorithm that assumes that "the addressee is where the eye is", and compare our results to alternative approaches.
Originele taal-2 | Engels |
---|---|
Titel | Proceedings of the 7th international conference on Multimodal interfaces, October 4-6, 2005, Torento, Italy |
Plaats van productie | New York, USA |
Uitgeverij | Association for Computing Machinery, Inc |
Pagina's | 175-182 |
ISBN van geprinte versie | 1-59593-028-0 |
DOI's | |
Status | Gepubliceerd - 2005 |
Evenement | 7th International Conference on Multimodal Interfaces, ICMI 2005 - Torento, Italië Duur: 4 okt. 2005 → 6 okt. 2005 Congresnummer: 7 |
Congres
Congres | 7th International Conference on Multimodal Interfaces, ICMI 2005 |
---|---|
Land/Regio | Italië |
Stad | Torento |
Periode | 4/10/05 → 6/10/05 |
Ander | ICMI ’05, International conference on multimodal interfaces |