Linear support vector machine based on variational inequality
ICNC'09 Proceedings of the 5th international conference on Natural computation
Hi-index | 0.00 |
When applying traditional methods to train approximately linear support vector machine (SVM), we will get a kernel matrix which occupy mass computer memory and lead a slow convergence speed. In order to improve the convergence speed of SVM, a method of training approximately linear support vector machine based on variational inequality (VIALSVM) was proposed. The method turns the convex quadratic programming into variational inequality in the process of training approximately linear support vector machine. The method can find the optimal separating hyperplane by solving variational inequality, and can not produce a great deal of data that occupy the computer memory. So the method can improve the training and test speed of SVM in classification greatly. We apply VIALSVM into the multidimensional iris training samples. Experiments show that VIALSVM not only has less misclassification rate, but also has more fast convergence speed than traditional SVM on the foundation of the equal misclassification, and especially in the high dimension training samples.