COLT '92 Proceedings of the fifth annual workshop on Computational learning theory
Information-based objective functions for active data selection
Neural Computation
Query Learning Strategies Using Boosting and Bagging
ICML '98 Proceedings of the Fifteenth International Conference on Machine Learning
Incremental Active Learning for Optimal Generalization
Neural Computation
Pool-based active learning in approximate linear regression
Machine Learning
Active graph matching based on pairwise probabilities between nodes
SSPR'12/SPR'12 Proceedings of the 2012 Joint IAPR international conference on Structural, Syntactic, and Statistical Pattern Recognition
Hi-index | 0.01 |
We propose a pool-based active learning algorithm with approximately optimal sampling distributions. An intuitive understanding of the effectiveness of active learning is also illustrated from the viewpoint of the information geometry. In active learning, one can choose informative input points or input distributions. Appropriate choice of data points is expected in order to make prediction performance more accurate than random data selection. Conventional active learning methods, however, yield serious estimation bias, when parametric statistical models do not include the true probability distribution. To correct the bias, we apply the maximum weighted log-likelihood estimator with approximately optimal input distribution. Optimal input distribution for active learning can be obtained by simple regression estimation. Numerical studies show the effectiveness of the proposed learning algorithm.