Adaptive Boosting with Leader based Learners for Classification of Large Handwritten Data

  • Authors:
  • T. Ravindra Babu;M. Narasimha Murty;V. K. Agrawal

  • Affiliations:
  • Indian Institute of Science, Bangalore;Indian Institute of Science, Bangalore;ISRO Satellite Centre, Bangalore

  • Venue:
  • HIS '04 Proceedings of the Fourth International Conference on Hybrid Intelligent Systems
  • Year:
  • 2004

Quantified Score

Hi-index 0.00

Visualization

Abstract

Boosting is a general method for improving the accuracy of a learning algorithm. AdaBoost, short form for Adaptive Boosting method, consists of repeated use of a weak or a base learning algorithm to find corresponding weak hypothesis by adapting to the error rates of the individual weak hypotheses. A large, complex handwritten data is under study. A repeated use of weak learner on the huge data results in large amount of processing time. In view of this, instead of using the entire training data for learning, we propose to use only prototypes. Further, in the current work, the base learner consists of a nearest neighbour classifier that employs prototypes generated using "leader" clustering algorithm. The leader algorithm is a single pass algorithm and is linear in terms of time as well as computation complexity. The prototype set alone is used as training data. In the process of developing an algorithm, domain knowledge of the Handwritten data, which is under study, is made use of. With the fusion of clustering, prototype selection, AdaBoost and Nearest Neighbour classifier, a very high classification accuracy, which is better than reported earlier on the considered data, is obtained in less number of iterations. The procedure integrates clustering outcome in terms of prototypes with boosting.