Visual attention guided bit allocation in video compression

  • Authors:
  • Zhicheng Li;Shiyin Qin;Laurent Itti

  • Affiliations:
  • School of Automation Science and Electrical Engineering, Beihang University, Beijing, China and Computer Science Department, University of Southern California, Los Angeles, CA, USA;School of Automation Science and Electrical Engineering, Beihang University, Beijing, China;Computer Science Department, University of Southern California, Los Angeles, CA, USA

  • Venue:
  • Image and Vision Computing
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

A visual attention-based bit allocation strategy for video compression is proposed. Saliency-based attention prediction is used to detect interesting regions in video. From the top salient locations from the computed saliency map, a guidance map is generated to guide the bit allocation strategy through a new constrained global optimization approach, which can be solved in a closed form and independently of video frame content. Fifty video sequences (300 frames each) and eye-tracking data from 14 subjects were collected to evaluate both the accuracy of the attention prediction model and the subjective quality of the encoded video. Results show that the area under the curve of the guidance map is 0.773+/-0.002, significantly above chance (0.500). Using a new eye-tracking-weighted PSNR (EWPSNR) measure of subjective quality, more than 90% of the encoded video clips with the proposed method achieve better subjective quality compared to standard encoding with matched bit rate. The improvement in EWPSNR is up to over 2dB and on average 0.79dB.