Psychovisual aspects of image communication
Signal Processing - Special issue on multidimensional signal processing
International Journal of Computer Vision
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis
IEEE Transactions on Pattern Analysis and Machine Intelligence
Spatiotemporal sensitivity and visual attention for efficient rendering of dynamic environments
ACM Transactions on Graphics (TOG)
Visibility of wavelet quantization noise
IEEE Transactions on Image Processing
Efficient and reliable schemes for nonlinear diffusion filtering
IEEE Transactions on Image Processing
High performance scalable image compression with EBCOT
IEEE Transactions on Image Processing
Locally adaptive perceptual image coding
IEEE Transactions on Image Processing
Embedded foveation image coding
IEEE Transactions on Image Processing
IEEE Transactions on Image Processing
Highly scalable video compression with scalable motion coding
IEEE Transactions on Image Processing
IEEE Transactions on Image Processing
IEEE Transactions on Image Processing
Multigrid Geometric Active Contour Models
IEEE Transactions on Image Processing
IEEE Transactions on Circuits and Systems for Video Technology
Segmentation of the face and hands in sign language video sequences using color and motion cues
IEEE Transactions on Circuits and Systems for Video Technology
Realizing Low-Cost High-Throughput General-Purpose Block Encoder for JPEG2000
IEEE Transactions on Circuits and Systems for Video Technology
TexToons: practical texture mapping for hand-drawn cartoon animations
Proceedings of the ACM SIGGRAPH/Eurographics Symposium on Non-Photorealistic Animation and Rendering
Distortion estimates for adaptive lifting transforms with noise
Image and Vision Computing
Hi-index | 0.00 |
This paper describes a visual optimization strategy for scalable video compression. The challenge scalable coding presents is that truncation of an embedded codestream may induce variable and highly visible distortion. To overcome the deficiencies of visually lossless coding schemes, we propose using an adaptive masking slope to model the perceptual impact of suprathreshold distortion arising from resolution and bit-rate scaling. This allows important scene structures to be better preserved. Following visual masking principles, local sensitivity to distortion is assessed within each frame. To keep the perceptual response uniform against spatiotemporal errors, we mitigate errors compounded by the motion field during temporal synthesis. Visual sensitivity weights are projected into the subband domain along motion trajectories via a process called perceptual mapping. This uses error propagation paths to capture some of the noise-shaping effects attributed to the motion-compensated transform. A key observation is that low contrast regions in the video are generally more susceptible to unmasking of quantization errors. The proposed approach raises the distortion-length slope associated with these critical regions, altering, the bitstream embedding order so that visually sensitive sites may be encoded with higher fidelity. Subjective evaluation demonstrates perceptual improvement with respect to bit-rate, spatial and temporal scalability.