Illumination invariant extraction for face recognition using neighboring wavelet coefficients

  • Authors:
  • X. Cao;W. Shen;L. G. Yu;Y. L. Wang;J. Y. Yang;Z. W. Zhang

  • Affiliations:
  • School of Computer Science and Technology, Nanjing University of Science and Technology, Nanjing, Jiangsu 210094, China;Samuel Ginn College of Engineering, Auburn University, 275 Wilmore Labs, Auburn, AL 36849, USA;School of Computer Science and Technology, Nanjing University of Science and Technology, Nanjing, Jiangsu 210094, China;School of Computer Science and Technology, Nanjing University of Science and Technology, Nanjing, Jiangsu 210094, China;School of Computer Science and Technology, Nanjing University of Science and Technology, Nanjing, Jiangsu 210094, China;Samuel Ginn College of Engineering, Auburn University, 275 Wilmore Labs, Auburn, AL 36849, USA

  • Venue:
  • Pattern Recognition
  • Year:
  • 2012

Quantified Score

Hi-index 0.01

Visualization

Abstract

The features of a face can change drastically as the illumination changes. In contrast to pose position and expression, illumination changes present a much greater challenge to face recognition. In this paper, we propose a novel wavelet based approach that considers the correlation of neighboring wavelet coefficients to extract an illumination invariant. This invariant represents the key facial structure needed for face recognition. Our method has better edge preserving ability in low frequency illumination fields and better useful information saving ability in high frequency fields using wavelet based NeighShrink denoise techniques. This method proposes different process approaches for training images and testing images since these images always have different illuminations. More importantly, by having different processes, a simple processing algorithm with low time complexity can be applied to the testing image. This leads to an easy application to real face recognition systems. Experimental results on Yale face database B and CMU PIE Face Database show that excellent recognition rates can be achieved by the proposed method.