抄録
Robot audition in real-world should cope with motor and other noises caused by the robot's own movements in addition to environmental noises and reverberation. This paper reports how auditory processing is improved by audio-visual integration with active movements. The key idea resides in hierarchical integration of auditory and visual streams to disambiguate auditory or visual processing. The system runs in real-time by using distributed processing on 4 PCs connected by Gigabit Ethernet. The system implemented in a upper-torso humanoid tracks multiple talkers and extracts speech from a mixture of sounds. The performance of epipolar geometry based sound source localization and sound source separation by active and adaptive direction-pass filtering is also reported.
本文言語 | English |
---|---|
ホスト出版物のタイトル | Proceedings - IEEE International Conference on Robotics and Automation |
ページ | 1043-1049 |
ページ数 | 7 |
巻 | 1 |
出版ステータス | Published - 2002 |
外部発表 | はい |
イベント | 2002 IEEE International Conference on Robotics and Automation - Washington, DC, United States 継続期間: 2002 5月 11 → 2002 5月 15 |
Other
Other | 2002 IEEE International Conference on Robotics and Automation |
---|---|
国/地域 | United States |
City | Washington, DC |
Period | 02/5/11 → 02/5/15 |
ASJC Scopus subject areas
- ソフトウェア
- 制御およびシステム工学