TY - GEN
T1 - Robot auditory system using head-mounted square microphone array
AU - Hosoya, Kosuke
AU - Ogawa, Tetsuji
AU - Kobayashi, Tetsunori
PY - 2009/12/11
Y1 - 2009/12/11
N2 - A new noise reduction method suitable for autonomous mobile robots was proposed and applied to preprocessing of a hands-free spoken dialogue system. When a robot talks with a conversational partner in real environments, not only speech utterances by the partner but also various types of noise, such as directional noise, diffuse noise, and noise from the robot, are observed at microphones. We attempted to remove these types of noise simultaneously with small and light-weighted devices and low-computational-cost algorithms. We assumed that the conversational partner of the robot was in front of the robot. In this case, the aim of the proposed method is extracting speech signals coming from the frontal direction of the robot. The proposed noise reduction system was evaluated in the presence of various types of noise: the number of word errors was reduced by 69 % as compared to the conventional methods. The proposed robot auditory system can also cope with the case in which a conversational partner (i.e., a sound source) moves from the front of the robot: the sound source was localized by face detection and tracking using facial images obtained from a camera mounted on an eye of the robot. As a result, various types of noise could be reduced in real time, irrespective of the sound source positions, by combining speech information with image information.
AB - A new noise reduction method suitable for autonomous mobile robots was proposed and applied to preprocessing of a hands-free spoken dialogue system. When a robot talks with a conversational partner in real environments, not only speech utterances by the partner but also various types of noise, such as directional noise, diffuse noise, and noise from the robot, are observed at microphones. We attempted to remove these types of noise simultaneously with small and light-weighted devices and low-computational-cost algorithms. We assumed that the conversational partner of the robot was in front of the robot. In this case, the aim of the proposed method is extracting speech signals coming from the frontal direction of the robot. The proposed noise reduction system was evaluated in the presence of various types of noise: the number of word errors was reduced by 69 % as compared to the conventional methods. The proposed robot auditory system can also cope with the case in which a conversational partner (i.e., a sound source) moves from the front of the robot: the sound source was localized by face detection and tracking using facial images obtained from a camera mounted on an eye of the robot. As a result, various types of noise could be reduced in real time, irrespective of the sound source positions, by combining speech information with image information.
UR - http://www.scopus.com/inward/record.url?scp=76249103539&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=76249103539&partnerID=8YFLogxK
U2 - 10.1109/IROS.2009.5354090
DO - 10.1109/IROS.2009.5354090
M3 - Conference contribution
AN - SCOPUS:76249103539
SN - 9781424438044
T3 - 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2009
SP - 2736
EP - 2741
BT - 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2009
T2 - 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2009
Y2 - 11 October 2009 through 15 October 2009
ER -