IMMC: incremental maximum margin criterion

  • Authors:
  • Jun Yan;Benyu Zhang;Shuicheng Yan;Qiang Yang;Hua Li;Zheng Chen;Wensi Xi;Weiguo Fan;Wei-Ying Ma;Qiansheng Cheng

  • Affiliations:
  • Peking University, Beijing, P. R. China;Microsoft Research Asia, Beijing, P. R. China;Peking University, Beijing, P. R. China;Hong Kong University of Science and Technology;Peking University, Beijing, P. R. China;Microsoft Research Asia, Beijing, P. R. China;Virginia Polytechnic Institute and State University, Blacksburg, VA;Virginia Polytechnic Institute and State University, Blacksburg, VA;Microsoft Research Asia, Beijing, P. R. China;Peking University, Beijing, P. R. China

  • Venue:
  • Proceedings of the tenth ACM SIGKDD international conference on Knowledge discovery and data mining
  • Year:
  • 2004

Quantified Score

Hi-index 0.00

Visualization

Abstract

Subspace learning approaches have attracted much attention in academia recently. However, the classical batch algorithms no longer satisfy the applications on streaming data or large-scale data. To meet this desirability, Incremental Principal Component Analysis (IPCA) algorithm has been well established, but it is an unsupervised subspace learning approach and is not optimal for general classification tasks, such as face recognition and Web document categorization. In this paper, we propose an incremental supervised subspace learning algorithm, called Incremental Maximum Margin Criterion (IMMC), to infer an adaptive subspace by optimizing the Maximum Margin Criterion. We also present the proof for convergence of the proposed algorithm. Experimental results on both synthetic dataset and real world datasets show that IMMC converges to the similar subspace as that of batch approach.