TY - GEN
T1 - Live assessment of beat tracking for robot audition
AU - Oliveira, Joao Lobato
AU - Ince, Gökhan
AU - Nakamura, Keisuke
AU - Nakadai, Kazuhiro
AU - Okuno, Hiroshi G.
AU - Reis, Luis Paulo
AU - Gouyon, Fabien
PY - 2012
Y1 - 2012
N2 - In this paper we propose the integration of an online audio beat tracking system into the general framework of robot audition, to enable its application in musically-interactive robotic scenarios. To this purpose, we introduced a staterecovery mechanism into our beat tracking algorithm, for handling continuous musical stimuli, and applied different multi-channel preprocessing algorithms (e.g., beamforming, ego noise suppression) to enhance noisy auditory signals lively captured in a real environment. We assessed and compared the robustness of our audio beat tracker through a set of experimental setups, under different live acoustic conditions of incremental complexity. These included the presence of continuous musical stimuli, built of a set of concatenated musical pieces; the presence of noises of different natures (e.g., robot motion, speech); and the simultaneous processing of different audio sources on-the-fly, for music and speech. We successfully tackled all these challenging acoustic conditions and improved the beat tracking accuracy and reaction time to music transitions while simultaneously achieving robust automatic speech recognition.
AB - In this paper we propose the integration of an online audio beat tracking system into the general framework of robot audition, to enable its application in musically-interactive robotic scenarios. To this purpose, we introduced a staterecovery mechanism into our beat tracking algorithm, for handling continuous musical stimuli, and applied different multi-channel preprocessing algorithms (e.g., beamforming, ego noise suppression) to enhance noisy auditory signals lively captured in a real environment. We assessed and compared the robustness of our audio beat tracker through a set of experimental setups, under different live acoustic conditions of incremental complexity. These included the presence of continuous musical stimuli, built of a set of concatenated musical pieces; the presence of noises of different natures (e.g., robot motion, speech); and the simultaneous processing of different audio sources on-the-fly, for music and speech. We successfully tackled all these challenging acoustic conditions and improved the beat tracking accuracy and reaction time to music transitions while simultaneously achieving robust automatic speech recognition.
UR - http://www.scopus.com/inward/record.url?scp=84872314131&partnerID=8YFLogxK
U2 - 10.1109/IROS.2012.6386100
DO - 10.1109/IROS.2012.6386100
M3 - Conference contribution
AN - SCOPUS:84872314131
SN - 9781467317375
T3 - IEEE International Conference on Intelligent Robots and Systems
SP - 992
EP - 997
BT - 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2012
T2 - 25th IEEE/RSJ International Conference on Robotics and Intelligent Systems, IROS 2012
Y2 - 7 October 2012 through 12 October 2012
ER -