Abstract
In this paper, we present our work in building technologies for natural multimodal human-robot interaction. We present our systems for spontaneous speech recognition, multimodal dialogue processing, and visual perception of a user, which includes localization, tracking, and identification of the user, recognition of pointing gestures, as well as the recognition of a person's head orientation. Each of the components is described in the paper and experimental results are presented. We also present several experiments on multimodal human-robot interaction, such as interaction using speech and gestures, the automatic determination of the addressee during human-human-robot interaction, as well on interactive learning of dialogue strategies. The work and the components presented here constitute the core building blocks for audiovisual perception of humans and multimodal human-robot interaction used for the humanoid robot developed within the German research project (Sonderforschungsbereich) on humanoid cooperative robots.
Original language | English (US) |
---|---|
Pages (from-to) | 840-851 |
Number of pages | 12 |
Journal | IEEE Transactions on Robotics |
Volume | 23 |
Issue number | 5 |
DOIs | |
State | Published - Oct 1 2007 |
Keywords
- Audiovisual perception
- Human-centered robotics
- Human-robot interaction
- Multimodal interaction
ASJC Scopus subject areas
- Control and Systems Engineering
- Computer Science Applications
- Electrical and Electronic Engineering