Inverse texture synthesis

  • Authors:
  • Li-Yi Wei;Jianwei Han;Kun Zhou;Hujun Bao;Baining Guo;Heung-Yeung Shum

  • Affiliations:
  • Microsoft Research Asia;Zhejiang University;Microsoft Research Asia;Zhejiang University;Microsoft Research Asia;Microsoft Research Asia

  • Venue:
  • ACM SIGGRAPH 2008 papers
  • Year:
  • 2008

Quantified Score

Hi-index 0.02

Visualization

Abstract

The quality and speed of most texture synthesis algorithms depend on a 2D input sample that is small and contains enough texture variations. However, little research exists on how to acquire such sample. For homogeneous patterns this can be achieved via manual cropping, but no adequate solution exists for inhomogeneous or globally varying textures, i.e. patterns that are local but not stationary, such as rusting over an iron statue with appearance conditioned on varying moisture levels. We present inverse texture synthesis to address this issue. Our inverse synthesis runs in the opposite direction with respect to traditional forward synthesis: given a large globally varying texture, our algorithm automatically produces a small texture compaction that best summarizes the original. This small compaction can be used to reconstruct the original texture or to re-synthesize new textures under user-supplied controls. More important, our technique allows real-time synthesis of globally varying textures on a GPU, where the texture memory is usually too small for large textures. We propose an optimization framework for inverse texture synthesis, ensuring that each input region is properly encoded in the output compaction. Our optimization process also automatically computes orientation fields for anisotropic textures containing both low- and high-frequency regions, a situation difficult to handle via existing techniques.