Computational Auditory Scene Analysis and Its Application to Robot Audition: Five Years Experience

  • Authors:
  • Hiroshi G. Okuno;Tetsuya Ogata;Kazunori Komatani

  • Affiliations:
  • Kyoto University, Japan;Kyoto University, Japan;Kyoto University, Japan

  • Venue:
  • ICKS '07 Proceedings of the Second International Conference on Informatics Research for Development of Knowledge Society Infrastructure
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

We have been engaged in research on computational auditory scene analysis to attain sophisticated robot/computer human interaction by manipulating real-world sound signals. The objective of our research is the understanding of an arbitrary sound mixture including non-speech sounds and music as well as voiced speech, obtained by robot's ears, that is, microphones embedded in the robot. We have coped with three main issues in computational auditory scene analysis, that is, sound source localization, separation, and recognition of separated sounds for a mixture of speech signals as well as polyphonic music signals. This paper overviews our results in robot audition, in particular, Missing Feature Theory based integration of sound source separation and automatic speech recognition, and those in music information processing, in particular, drum sound equalizer.