An extraction method of lip movement images from successive image frames in the speech activity extraction process

  • Authors:
  • Eung-Kyeu Kim;Soo-Jong Lee;Nohpill Park

  • Affiliations:
  • Dept. of Infor. & Commu. Engineering, Hanbat Nat'l Univ., Korea;Automatic Speech Translation Research Team, ETRI, Korea;Dept. of Computer Science, Oklahoma State University

  • Venue:
  • ICEC'10 Proceedings of the 9th international conference on Entertainment computing
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we propose an extraction method of lip movement images from successive image frames and present the possibility to utilize lip movement images in the speech activity extraction process of speech recognition phase. The image frames are acquired from the PC image camera with the assumption that facial movement is limited during talking. First of all, one new lip movement image frame is generated with comparing two successive image frames each other. Second, the fine image noises are removed. Each fitness rate is calculated by comparing the lip feature data as objectly separated images. It is analyzed whether or not there is the lip movement image through verification to the objects and three images which have higher rates in their fitnesses. As a result of linking the speech & image processing system, the interworking rate shows 99.3% even in the various illumination environments. It was visually confirmed that lip movement images are tracked and can be utilized in speech activity extraction process.