The nature of statistical learning theory
The nature of statistical learning theory
A tutorial on support vector regression
Statistics and Computing
Second-order smo improves svm online and active learning
Neural Computation
Improvements to the SMO algorithm for SVM regression
IEEE Transactions on Neural Networks
Hi-index | 0.00 |
Shevade's et al. Modification 2 is one of the most widely used algorithms to build Support Vector Regression (SVR) models. It selects as a size 2 working set the index pair giving the maximum KKT violation and combines it with the updating heuristics of Smola and Schölkopf enforcing at each training iteration a $\alpha_i \alpha^*_i =0$ condition. In this work we shall present an alternative, much simpler procedure that selects the updating indices as those giving a maximum gain in the SVR dual function. While we do not try to enforce the $\alpha_i \alpha^*_i =0$ condition, we show that it will hold at each iteration provided it does so at the starting multipliers. We will numerically show that the proposed procedure requires essentially the same number of iterations than Modification 2 having thus the same time performance while being much simpler to code.