Optimizing resources in model selection for support vector machine
作者:
Highlights:
•
摘要
Tuning support vector machine (SVM) hyperparameters is an important step in achieving a high-performance learning machine. It is usually done by minimizing an estimate of generalization error based on the bounds of the leave-one-out (LOO) such as radius-margin bound and on the performance measures such as generalized approximate cross-validation (GACV), empirical error, etc. These usual automatic methods used to tune the hyperparameters require an inversion of the Gram–Schmidt matrix or a resolution of an extra-quadratic programming problem. In the case of a large data set these methods require the addition of huge amounts of memory and a long CPU time to the already significant resources used in SVM training. In this paper, we propose a fast method based on an approximation of the gradient of the empirical error, along with incremental learning, which reduces the resources required both in terms of processing time and of storage space. We tested our method on several benchmarks, which produced promising results confirming our approach. Furthermore, it is worth noting that the gain time increases when the data set is large.
论文关键词:Model selection,SVM,Kernel,Hyperparameters,Optimizing time
论文评审过程:Received 16 November 2005, Revised 26 April 2006, Accepted 6 June 2006, Available online 17 August 2006.
论文官网地址:https://doi.org/10.1016/j.patcog.2006.06.012