Hi-index | 0.00 |
Multi-class approaches for SVMs are based on composition of binary SVM classifiers. Due to the numerous binary classifiers to be considered, for large training sets, this approach is known to be time expensive. In our approach, we improve time efficiency using concurrently two strategies: incremental training and reduction of trained binary SVMs. We present the exact migration conditions for the binary SVMs during their incremental training. We rewrite these conditions for the case when the regularization parameter is optimized. The obtained results are applied to a multi-class incremental / decremental SVM based on the Adaptive Directed Acyclic Graph. The regularization parameter is optimized on-line, and not by retraining the SVM with all input samples for each value of the regularization parameter.