The purpose of this paper is to present extended least squares support vector machines (LS‐SVM) where data selection methods are used to get sparse LS‐SVM solution, and to overview and compare the most important data selection approaches.
The selection methods are compared based on their theoretical background and using extensive simulations.
The paper shows that partial reduction is an efficient way of getting a reduced complexity sparse LS‐SVM solution, while partial reduction exploits full knowledge contained in the whole training data set. It also shows that the reduction technique based on reduced row echelon form (RREF) of the kernel matrix is superior when compared to other data selection approaches.
Data selection for getting a sparse LS‐SVM solution can be done in the different representations of the training data: in the input space, in the intermediate feature space, and in the kernel space. Selection in the kernel space can be obtained by finding an approximate basis of the kernel matrix.
The RREF‐based method is a data selection approach with a favorable property: there is a trade‐off tolerance parameter that can be used for balancing complexity and accuracy.
The paper gives contributions to the construction of high‐performance and moderate complexity LS‐SVMs.
Valyon, J. and Horváth, G. (2008), "Selection methods for extended least squares support vector machines", International Journal of Intelligent Computing and Cybernetics, Vol. 1 No. 1, pp. 69-93. https://doi.org/10.1108/17563780810857130Download as .RIS
Emerald Group Publishing Limited
Copyright © 2008, Emerald Group Publishing Limited