Real-time auditory and visual multiple-object tracking for humanoids

  • Authors:
  • Kazuhiro Nakadai;Ken-ichi Hidai;Hiroshi Mizoguchi;Hiroshi G. Okuno;Hiroaki Kitano

  • Affiliations:
  • Kitano Symbiotic Systems Project, ERATO, Japan Science and Technology Corp., Shibuya-ku, Tokyo, Japan;Kitano Symbiotic Systems Project, ERATO, Japan Science and Technology Corp., Shibuya-ku, Tokyo, Japan;Department of Information and Computer Science, Saitama University, Saitama, Japan;Kitano Symbiotic Systems Project, ERATO, Japan Science and Technology Corp., Shibuya-ku, Tokyo, Japan and Department of Intelligence Science and Technology, Kyoto University, Kyoto, Japan;Kitano Symbiotic Systems Project, ERATO, Japan Science and Technology Corp., Shibuya-ku, Tokyo, Japan and Sony Computer Science Laboratories, Inc., Tokyo, Japan

  • Venue:
  • IJCAI'01 Proceedings of the 17th international joint conference on Artificial intelligence - Volume 2
  • Year:
  • 2001

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper presents a real-time auditory and visual tracking of multiple objects for humanoid under real-world environments. Real-time processing is crucial for sensorimotor tasks in tracking, and multiple-object tracking is crucial for real-world applications. Multiple sound source tracking needs perception of a mixture of sounds and cancellation of motor noises caused by body movements. However its real-time processing has not been reported yet. Real-time tracking is attained by fusing information obtained by sound source localization, multiple face recognition, speaker tracking, focus of attention control, and motor control. Auditory streams with sound source direction are extracted by active audition system with motor noise cancellation capability from 48KHz sampling sounds. Visual streams with face ID and 3D-position are extracted by combining skincolor extraction, correlation-based matching, and multiple-scale image generation from a single camera. These auditory and visual streams are associated by comparing the spatial location, and associated streams are used to control focus of attention. Auditory, visual, and association processing are performed asynchronously on different PC's connected by TCP/IP network. The resulting system implemented on an upper-torso humanoid can track multiple objects with the delay of 200 msec, which is forced by visual tracking and network latency.