K-hyperline clustering learning for sparse component analysis

  • Authors:
  • Zhaoshui He;Andrzej Cichocki;Yuanqing Li;Shengli Xie;Saeid Sanei

  • Affiliations:
  • Laboratory for Advanced Brain Signal Processing, RIKEN Brain Science Institute, Saitama 3510198, Japan and School of Electronics and Information Engineering, South China University of Technology, ...;Laboratory for Advanced Brain Signal Processing, RIKEN Brain Science Institute, Saitama 3510198, Japan and System Research Institute, Polish Academy of Sciences (PAN), Warsaw 00-901, Poland and Wa ...;Automation Science and Engineering Institute, South China University of Technology, Guangzhou 510641, China;School of Electronics and Information Engineering, South China University of Technology, Guangzhou 510641, China;Centre of Digital Signal Processing, Cardiff University, Cardiff CF24 3AA, Wales, UK

  • Venue:
  • Signal Processing
  • Year:
  • 2009

Quantified Score

Hi-index 0.08

Visualization

Abstract

A two-stage clustering-then-@?"1-optimization approach has been often used for sparse component analysis (SCA). The first challenging task of this approach is to estimate the basis matrix by cluster analysis. In this paper, a robust K-hyperline clustering (K-HLC) algorithm is developed for this task. The novelty of our method is that it is not only able to implement hyperline clustering, but also is capable of detecting the number of hidden hyperlines (or sparse components). K-HLC seamlessly integrates ''the hyperline clustering'' and ''hyperline number detection'' in the same algorithm. In addition, three strategies are proposed to tackle this problem: (1) reject the outliers by overestimating the number of hyperlines; (2) escape from local minima by using a multilayer initialization and (3) suppress the noise by a multilayer K-HLC. By taking these strategies into account, the robust K-HLC procedure can be briefly described as follows: first, we overestimate the number of hyperlines; then, a confidence index is given to evaluate the significance of each hyperline. Subsequently, we determine the number of hyperlines by checking the gap in the sorted confidence indices. Moreover, we select those hyperlines corresponding to large confidence indices with high rank priority and remove spurious ones with small confidence indices. The high performance of our clustering scheme is illustrated by extensive numerical experiments including some challenging benchmarks, e.g., very ill-conditioned basis matrix (Hilbert matrix), or the observations with strong outliers.