Social interaction of humanoid robot based on audio-visual tracking

Hiroshi G. Okuno, Kazuhiro Nakadai, Hiroaki Kitano

Research output: Chapter in Book/Report/Conference proceedingConference contribution

24 Citations (Scopus)

Abstract

Social interaction is essential in improving robot human interface. Such behaviors for social interaction may include paying attention to a new sound source, moving toward it, or keeping face to face with a moving speaker. Some sound-centered behaviors may be difficult to attain, because the mixture of sounds is not well treated or auditory processing is too slow for real-time applications. Recently, Nakadai et al have developed real-time auditory and visual multiple-talker tracking technology by associating auditory and visual streams. The system is implemented on an upper-torso humanoid and the real-time talker tracking is attained with 200 msec of delay by distributed processing on four PCs connected by Gigabit Ethernet. Focus-of-attention is programmable and allows a variety of behaviors. The system demonstrates non-verbal social interaction by realizing a receptionist robot by focusing on an associated stream, while a companion robot on an auditory stream.

Original languageEnglish
Title of host publicationLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
PublisherSpringer Verlag
Pages725-735
Number of pages11
Volume2358
ISBN (Print)3540437819, 9783540437819
Publication statusPublished - 2002
Externally publishedYes
Event15th International Conference on Industrial and Engineering Applications of Artificial Intelligence and Expert Systems, IEA/AIE 2002 - Cairns, Australia
Duration: 2002 Jun 172002 Jun 20

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume2358
ISSN (Print)03029743
ISSN (Electronic)16113349

Other

Other15th International Conference on Industrial and Engineering Applications of Artificial Intelligence and Expert Systems, IEA/AIE 2002
CountryAustralia
CityCairns
Period02/6/1702/6/20

Fingerprint

Visual Tracking
Humanoid Robot
Social Interaction
Robot
Acoustic waves
Robots
Real-time
Human Interface
Distributed Processing
Ethernet
Processing
Demonstrate
Sound
Vision

ASJC Scopus subject areas

  • Computer Science(all)
  • Theoretical Computer Science

Cite this

Okuno, H. G., Nakadai, K., & Kitano, H. (2002). Social interaction of humanoid robot based on audio-visual tracking. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 2358, pp. 725-735). (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Vol. 2358). Springer Verlag.

Social interaction of humanoid robot based on audio-visual tracking. / Okuno, Hiroshi G.; Nakadai, Kazuhiro; Kitano, Hiroaki.

Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). Vol. 2358 Springer Verlag, 2002. p. 725-735 (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Vol. 2358).

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Okuno, HG, Nakadai, K & Kitano, H 2002, Social interaction of humanoid robot based on audio-visual tracking. in Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). vol. 2358, Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 2358, Springer Verlag, pp. 725-735, 15th International Conference on Industrial and Engineering Applications of Artificial Intelligence and Expert Systems, IEA/AIE 2002, Cairns, Australia, 02/6/17.
Okuno HG, Nakadai K, Kitano H. Social interaction of humanoid robot based on audio-visual tracking. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). Vol. 2358. Springer Verlag. 2002. p. 725-735. (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)).
Okuno, Hiroshi G. ; Nakadai, Kazuhiro ; Kitano, Hiroaki. / Social interaction of humanoid robot based on audio-visual tracking. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). Vol. 2358 Springer Verlag, 2002. pp. 725-735 (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)).
@inproceedings{a8973d1a6a25499d9ab829f148ec37c0,
title = "Social interaction of humanoid robot based on audio-visual tracking",
abstract = "Social interaction is essential in improving robot human interface. Such behaviors for social interaction may include paying attention to a new sound source, moving toward it, or keeping face to face with a moving speaker. Some sound-centered behaviors may be difficult to attain, because the mixture of sounds is not well treated or auditory processing is too slow for real-time applications. Recently, Nakadai et al have developed real-time auditory and visual multiple-talker tracking technology by associating auditory and visual streams. The system is implemented on an upper-torso humanoid and the real-time talker tracking is attained with 200 msec of delay by distributed processing on four PCs connected by Gigabit Ethernet. Focus-of-attention is programmable and allows a variety of behaviors. The system demonstrates non-verbal social interaction by realizing a receptionist robot by focusing on an associated stream, while a companion robot on an auditory stream.",
author = "Okuno, {Hiroshi G.} and Kazuhiro Nakadai and Hiroaki Kitano",
year = "2002",
language = "English",
isbn = "3540437819",
volume = "2358",
series = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
publisher = "Springer Verlag",
pages = "725--735",
booktitle = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",

}

TY - GEN

T1 - Social interaction of humanoid robot based on audio-visual tracking

AU - Okuno, Hiroshi G.

AU - Nakadai, Kazuhiro

AU - Kitano, Hiroaki

PY - 2002

Y1 - 2002

N2 - Social interaction is essential in improving robot human interface. Such behaviors for social interaction may include paying attention to a new sound source, moving toward it, or keeping face to face with a moving speaker. Some sound-centered behaviors may be difficult to attain, because the mixture of sounds is not well treated or auditory processing is too slow for real-time applications. Recently, Nakadai et al have developed real-time auditory and visual multiple-talker tracking technology by associating auditory and visual streams. The system is implemented on an upper-torso humanoid and the real-time talker tracking is attained with 200 msec of delay by distributed processing on four PCs connected by Gigabit Ethernet. Focus-of-attention is programmable and allows a variety of behaviors. The system demonstrates non-verbal social interaction by realizing a receptionist robot by focusing on an associated stream, while a companion robot on an auditory stream.

AB - Social interaction is essential in improving robot human interface. Such behaviors for social interaction may include paying attention to a new sound source, moving toward it, or keeping face to face with a moving speaker. Some sound-centered behaviors may be difficult to attain, because the mixture of sounds is not well treated or auditory processing is too slow for real-time applications. Recently, Nakadai et al have developed real-time auditory and visual multiple-talker tracking technology by associating auditory and visual streams. The system is implemented on an upper-torso humanoid and the real-time talker tracking is attained with 200 msec of delay by distributed processing on four PCs connected by Gigabit Ethernet. Focus-of-attention is programmable and allows a variety of behaviors. The system demonstrates non-verbal social interaction by realizing a receptionist robot by focusing on an associated stream, while a companion robot on an auditory stream.

UR - http://www.scopus.com/inward/record.url?scp=84923324505&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=84923324505&partnerID=8YFLogxK

M3 - Conference contribution

AN - SCOPUS:84923324505

SN - 3540437819

SN - 9783540437819

VL - 2358

T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)

SP - 725

EP - 735

BT - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)

PB - Springer Verlag

ER -