作者
SVN Vishwanathan, Alexander J Smola, M Narasimha Murty
发表日期
2003/8/21
图书
Proceedings of the Twentieth International Conference on International Conference on Machine Learning
页码范围
760-767
简介
We present a fast iterative support vector training algorithm for a large variety of different formulations. It works by incrementally changing a candidate support vector set using a greedy approach, until the supporting hyperplane is found within a finite number of iterations.
It is derived from a simple active set method which sweeps through the set of Lagrange multipliers and keeps optimality in the unconstrained variables, while discarding large amounts of bound-constrained variables. The hard-margin version can be viewed as a simple (yet computationally crucial) modification of the incremental SVM training algorithms of Cauwenberghs and Poggio.
Experimental results for various settings are reported. In all cases our algorithm is considerably faster than competing methods such as Sequential Minimal Optimization or the Nearest Point Algorithm.
引用总数
2004200520062007200820092010201120122013201420152016201720182019202020212022202320243119192615139967864332221