Detecting robot-directed speech by situated understanding in object manipulation tasks

Xiang Zuo, Naoto Iwahashi, Ryo Taguchi, Kotaro Funakoshi, Mikio Nakano, Shigeki Matsuda, Komei Sugiura, Natsuki Oka

Research output: Chapter in Book/Report/Conference proceedingConference contribution

5 Citations (Scopus)

Abstract

In this paper, we propose a novel method for a robot to detect robot-directed speech, that is, to distinguish speech that users speak to a robot from speech that users speak to other people or to themselves. The originality of this work is the introduction of a multimodal semantic confidence (MSC) measure, which is used for domain classification of input speech based on the decision on whether the speech can be interpreted as a feasible action under the current physical situation in an object manipulation task. This measure is calculated by integrating speech, object, and motion confidence with weightings that are optimized by logistic regression. Then we integrate this measure with gaze tracking and conduct experiments under conditions of natural human-robot interaction. Experimental results show that the proposed method achieves a high performance of 94% and 96% in average recall and precision rates, respectively, for robot-directed speech detection.

Original languageEnglish
Title of host publication19th International Symposium in Robot and Human Interactive Communication, RO-MAN 2010
Pages608-613
Number of pages6
DOIs
Publication statusPublished - 2010 Dec 13
Externally publishedYes
Event19th IEEE International Conference on Robot and Human Interactive Communication, RO-MAN 2010 - Viareggio, Italy
Duration: 2010 Sep 122010 Sep 15

Publication series

NameProceedings - IEEE International Workshop on Robot and Human Interactive Communication

Other

Other19th IEEE International Conference on Robot and Human Interactive Communication, RO-MAN 2010
CountryItaly
CityViareggio
Period10/9/1210/9/15

ASJC Scopus subject areas

  • Software
  • Artificial Intelligence
  • Human-Computer Interaction

Fingerprint Dive into the research topics of 'Detecting robot-directed speech by situated understanding in object manipulation tasks'. Together they form a unique fingerprint.

  • Cite this

    Zuo, X., Iwahashi, N., Taguchi, R., Funakoshi, K., Nakano, M., Matsuda, S., Sugiura, K., & Oka, N. (2010). Detecting robot-directed speech by situated understanding in object manipulation tasks. In 19th International Symposium in Robot and Human Interactive Communication, RO-MAN 2010 (pp. 608-613). [5598729] (Proceedings - IEEE International Workshop on Robot and Human Interactive Communication). https://doi.org/10.1109/ROMAN.2010.5598729