A comparative study on attention-based rate adaptation for scalable video coding

  • Authors:
  • Chia-Ming Tsai;Chia-Wen Lin;Weisi Lin;Wen-Hsiao Peng

  • Affiliations:
  • Department of Computer Science, National Chung Cheng University;Department of Electrical Engineering, National Tsing Hua University;School of Computer Engineering, Nanyang Technological University;Department of Computer Science, National Chiao Tung University

  • Venue:
  • ICIP'09 Proceedings of the 16th IEEE international conference on Image processing
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

We conduct subjective tests to evaluate the performance of scalable video coding with different spatial-domain bit-allocation methods, visual attention models, and motion feature extractors in the literature. For spatial-domain bit allocation, we use the selective enhancement and quality layer assignment methods. For characterizing visual attention, we use the motion attention model and perceptual quality significant map. For motion features, we adopt motion vectors from hierarchical B-picture coding and optical flow. Experimental results show that a more accurate visual attention model leads to better perceptual quality. In cooperation with a visual attention model, the selective enhancement method, compared to the quality layer assignment, achieves better subjective quality when an ROI has enough bit allocation and its texture is not complex. The quality layer assignment method is suitable for region-wise quality enhancement due to its framebased allocation nature.