Making a robot dance to diverse musical genre in noisy environments

João Lobato Oliveira, Keisuke Nakamura, Thibault Langlois, Fabien Gouyon, Kazuhiro Nakadai, Angelica Lim, Luis Paulo Reis, Hiroshi G. Okuno

Research output: Chapter in Book/Report/Conference proceedingConference contribution

1 Citation (Scopus)

Abstract

In this paper we address the problem of musical genre recognition for a dancing robot with embedded microphones capable of distinguishing the genre of a musical piece while moving in a real-world scenario. For this purpose, we assess and compare two state-of-the-art musical genre recognition systems, based on Support Vector Machines and Markov Models, in the context of different real-world acoustic environments. In addition, we compare different preprocessing robot audition variants (single channel and separated signal from multiple channels) and test different acoustic models, learned a priori, to tackle multiple noise conditions of increasing complexity in the presence of noises of different natures (e.g., robot motion, speech). The results with six different musical genres suggest improved results, in the order of 43.6pp for the most complex conditions, when recurring to Sound Source Separation and acoustic models trained in similar conditions to the testing scenarios. A robot dance demonstration session confirms the applicability of the proposed integration for genre-adaptive dancing robots in real-world noisy environments.

Original languageEnglish
Title of host publicationIEEE International Conference on Intelligent Robots and Systems
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages1896-1901
Number of pages6
ISBN (Print)9781479969340
DOIs
Publication statusPublished - 2014 Oct 31
Externally publishedYes
Event2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2014 - Chicago
Duration: 2014 Sep 142014 Sep 18

Other

Other2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2014
CityChicago
Period14/9/1414/9/18

Fingerprint

Robots
Acoustics
Intelligent robots
Source separation
Audition
Microphones
Support vector machines
Demonstrations
Acoustic waves
Testing

ASJC Scopus subject areas

  • Control and Systems Engineering
  • Software
  • Computer Vision and Pattern Recognition
  • Computer Science Applications

Cite this

Oliveira, J. L., Nakamura, K., Langlois, T., Gouyon, F., Nakadai, K., Lim, A., ... Okuno, H. G. (2014). Making a robot dance to diverse musical genre in noisy environments. In IEEE International Conference on Intelligent Robots and Systems (pp. 1896-1901). [6942812] Institute of Electrical and Electronics Engineers Inc.. https://doi.org/10.1109/IROS.2014.6942812

Making a robot dance to diverse musical genre in noisy environments. / Oliveira, João Lobato; Nakamura, Keisuke; Langlois, Thibault; Gouyon, Fabien; Nakadai, Kazuhiro; Lim, Angelica; Reis, Luis Paulo; Okuno, Hiroshi G.

IEEE International Conference on Intelligent Robots and Systems. Institute of Electrical and Electronics Engineers Inc., 2014. p. 1896-1901 6942812.

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Oliveira, JL, Nakamura, K, Langlois, T, Gouyon, F, Nakadai, K, Lim, A, Reis, LP & Okuno, HG 2014, Making a robot dance to diverse musical genre in noisy environments. in IEEE International Conference on Intelligent Robots and Systems., 6942812, Institute of Electrical and Electronics Engineers Inc., pp. 1896-1901, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2014, Chicago, 14/9/14. https://doi.org/10.1109/IROS.2014.6942812
Oliveira JL, Nakamura K, Langlois T, Gouyon F, Nakadai K, Lim A et al. Making a robot dance to diverse musical genre in noisy environments. In IEEE International Conference on Intelligent Robots and Systems. Institute of Electrical and Electronics Engineers Inc. 2014. p. 1896-1901. 6942812 https://doi.org/10.1109/IROS.2014.6942812
Oliveira, João Lobato ; Nakamura, Keisuke ; Langlois, Thibault ; Gouyon, Fabien ; Nakadai, Kazuhiro ; Lim, Angelica ; Reis, Luis Paulo ; Okuno, Hiroshi G. / Making a robot dance to diverse musical genre in noisy environments. IEEE International Conference on Intelligent Robots and Systems. Institute of Electrical and Electronics Engineers Inc., 2014. pp. 1896-1901
@inproceedings{46002e08ea8a4a32af9f22d87cc28e8b,
title = "Making a robot dance to diverse musical genre in noisy environments",
abstract = "In this paper we address the problem of musical genre recognition for a dancing robot with embedded microphones capable of distinguishing the genre of a musical piece while moving in a real-world scenario. For this purpose, we assess and compare two state-of-the-art musical genre recognition systems, based on Support Vector Machines and Markov Models, in the context of different real-world acoustic environments. In addition, we compare different preprocessing robot audition variants (single channel and separated signal from multiple channels) and test different acoustic models, learned a priori, to tackle multiple noise conditions of increasing complexity in the presence of noises of different natures (e.g., robot motion, speech). The results with six different musical genres suggest improved results, in the order of 43.6pp for the most complex conditions, when recurring to Sound Source Separation and acoustic models trained in similar conditions to the testing scenarios. A robot dance demonstration session confirms the applicability of the proposed integration for genre-adaptive dancing robots in real-world noisy environments.",
author = "Oliveira, {Jo{\~a}o Lobato} and Keisuke Nakamura and Thibault Langlois and Fabien Gouyon and Kazuhiro Nakadai and Angelica Lim and Reis, {Luis Paulo} and Okuno, {Hiroshi G.}",
year = "2014",
month = "10",
day = "31",
doi = "10.1109/IROS.2014.6942812",
language = "English",
isbn = "9781479969340",
pages = "1896--1901",
booktitle = "IEEE International Conference on Intelligent Robots and Systems",
publisher = "Institute of Electrical and Electronics Engineers Inc.",

}

TY - GEN

T1 - Making a robot dance to diverse musical genre in noisy environments

AU - Oliveira, João Lobato

AU - Nakamura, Keisuke

AU - Langlois, Thibault

AU - Gouyon, Fabien

AU - Nakadai, Kazuhiro

AU - Lim, Angelica

AU - Reis, Luis Paulo

AU - Okuno, Hiroshi G.

PY - 2014/10/31

Y1 - 2014/10/31

N2 - In this paper we address the problem of musical genre recognition for a dancing robot with embedded microphones capable of distinguishing the genre of a musical piece while moving in a real-world scenario. For this purpose, we assess and compare two state-of-the-art musical genre recognition systems, based on Support Vector Machines and Markov Models, in the context of different real-world acoustic environments. In addition, we compare different preprocessing robot audition variants (single channel and separated signal from multiple channels) and test different acoustic models, learned a priori, to tackle multiple noise conditions of increasing complexity in the presence of noises of different natures (e.g., robot motion, speech). The results with six different musical genres suggest improved results, in the order of 43.6pp for the most complex conditions, when recurring to Sound Source Separation and acoustic models trained in similar conditions to the testing scenarios. A robot dance demonstration session confirms the applicability of the proposed integration for genre-adaptive dancing robots in real-world noisy environments.

AB - In this paper we address the problem of musical genre recognition for a dancing robot with embedded microphones capable of distinguishing the genre of a musical piece while moving in a real-world scenario. For this purpose, we assess and compare two state-of-the-art musical genre recognition systems, based on Support Vector Machines and Markov Models, in the context of different real-world acoustic environments. In addition, we compare different preprocessing robot audition variants (single channel and separated signal from multiple channels) and test different acoustic models, learned a priori, to tackle multiple noise conditions of increasing complexity in the presence of noises of different natures (e.g., robot motion, speech). The results with six different musical genres suggest improved results, in the order of 43.6pp for the most complex conditions, when recurring to Sound Source Separation and acoustic models trained in similar conditions to the testing scenarios. A robot dance demonstration session confirms the applicability of the proposed integration for genre-adaptive dancing robots in real-world noisy environments.

UR - http://www.scopus.com/inward/record.url?scp=84911469221&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=84911469221&partnerID=8YFLogxK

U2 - 10.1109/IROS.2014.6942812

DO - 10.1109/IROS.2014.6942812

M3 - Conference contribution

AN - SCOPUS:84911469221

SN - 9781479969340

SP - 1896

EP - 1901

BT - IEEE International Conference on Intelligent Robots and Systems

PB - Institute of Electrical and Electronics Engineers Inc.

ER -