Training with noise is equivalent to Tikhonov regularization
Neural Computation
Machine Learning
Prior knowledge in support vector kernels
NIPS '97 Proceedings of the 1997 conference on Advances in neural information processing systems 10
Noisy replication in skewed binary classification
Computational Statistics & Data Analysis
A Tutorial on Support Vector Machines for Pattern Recognition
Data Mining and Knowledge Discovery
Recognition and Structure from one 2D Model View: Observations on Prototypes, Object Classes and Symmetries
Expert Systems with Applications: An International Journal
Combining neural networks and semantic feature space for email classification
Knowledge-Based Systems
A study of cross-validation and bootstrap for accuracy estimation and model selection
IJCAI'95 Proceedings of the 14th international joint conference on Artificial intelligence - Volume 2
A neural network approach for solving linear bilevel programming problem
Knowledge-Based Systems
An overview of statistical learning theory
IEEE Transactions on Neural Networks
A new approach for manufacturing forecast problems with insufficient data: the case of TFT---LCDs
Journal of Intelligent Manufacturing
Hi-index | 0.00 |
Traditional machine learning algorithms are not with satisfying generalization ability on noisy, imbalanced, and small sample training set. In this work, a novel virtual sample generation (VSG) method based on Gaussian distribution is proposed. Firstly, the method determines the mean and the standard error of Gaussian distribution. Then, virtual samples can be generated by such Gaussian distribution. Finally, a new training set is constructed by adding the virtual samples to the original training set. This work has shown that training on the new training set is equivalent to a form of regularization regarding small sample problems, or cost-sensitive learning regarding imbalanced sample problems. Experiments show that given a suitable number of virtual sample replicates, the generalization ability of the classifiers on the new training sets can be better than that on the original training sets.