Learning Big (Image) Data via Coresets for Dictionaries

  • Authors:
  • Dan Feldman;Micha Feigin;Nir Sochen

  • Affiliations:
  • CSAIL, MIT, Cambridge, USA;Media Lab., MIT, Cambridge, USA;Department of Applied Mathematics, Tel Aviv university, Tel Aviv, Israel

  • Venue:
  • Journal of Mathematical Imaging and Vision
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Signal and image processing have seen an explosion of interest in the last few years in a new form of signal/image characterization via the concept of sparsity with respect to a dictionary. An active field of research is dictionary learning: the representation of a given large set of vectors (e.g. signals or images) as linear combinations of only few vectors (patterns). To further reduce the size of the representation, the combinations are usually required to be sparse, i.e., each signal is a linear combination of only a small number of patterns.This paper suggests a new computational approach to the problem of dictionary learning, known in computational geometry as coresets. A coreset for dictionary learning is a small smart non-uniform sample from the input signals such that the quality of any given dictionary with respect to the input can be approximated via the coreset. In particular, the optimal dictionary for the input can be approximated by learning the coreset. Since the coreset is small, the learning is faster. Moreover, using merge-and-reduce, the coreset can be constructed for streaming signals that do not fit in memory and can also be computed in parallel.We apply our coresets for dictionary learning of images using the K-SVD algorithm and bound their size and approximation error analytically. Our simulations demonstrate gain factors of up to 60 in computational time with the same, and even better, performance. We also demonstrate our ability to perform computations on larger patches and high-definition images, where the traditional approach breaks down.