{"title":"优化SVM超参数","authors":"Huang Dongyuan, Chen Xiaoyun","doi":"10.1109/CINC.2010.5643857","DOIUrl":null,"url":null,"abstract":"Choosing optimal hyperparameters for Support Vector Machines(SVMs) is quite difficult but extremely essential in SVM design. This is usually done by minimizing estimates of generalization error such as the k-fold cross-validation error or the upper bound of leave-one-out(LOO) error. However, most of the approaches concentrate on the dual optimization problem of SVM. In this paper, we would like to consider the task of tuning hyperparameters in the primal. We derive a smooth validation function from the k-fold cross-validation, then tune hyperparameters by minimizing the smooth validation function using Quasi- Newton optimization technique. Experimental results not only show that our approach is much faster and provides more precise results than grid search method, but also demonstrate that tuning hyperparameters in the primal would be more efficient than in the dual due to advantages provided by the primal.","PeriodicalId":227004,"journal":{"name":"2010 Second International Conference on Computational Intelligence and Natural Computing","volume":"13 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":"{\"title\":\"Tuning SVM hyperparameters in the primal\",\"authors\":\"Huang Dongyuan, Chen Xiaoyun\",\"doi\":\"10.1109/CINC.2010.5643857\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Choosing optimal hyperparameters for Support Vector Machines(SVMs) is quite difficult but extremely essential in SVM design. This is usually done by minimizing estimates of generalization error such as the k-fold cross-validation error or the upper bound of leave-one-out(LOO) error. However, most of the approaches concentrate on the dual optimization problem of SVM. In this paper, we would like to consider the task of tuning hyperparameters in the primal. We derive a smooth validation function from the k-fold cross-validation, then tune hyperparameters by minimizing the smooth validation function using Quasi- Newton optimization technique. Experimental results not only show that our approach is much faster and provides more precise results than grid search method, but also demonstrate that tuning hyperparameters in the primal would be more efficient than in the dual due to advantages provided by the primal.\",\"PeriodicalId\":227004,\"journal\":{\"name\":\"2010 Second International Conference on Computational Intelligence and Natural Computing\",\"volume\":\"13 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2010-11-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"8\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2010 Second International Conference on Computational Intelligence and Natural Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CINC.2010.5643857\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 Second International Conference on Computational Intelligence and Natural Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CINC.2010.5643857","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Choosing optimal hyperparameters for Support Vector Machines(SVMs) is quite difficult but extremely essential in SVM design. This is usually done by minimizing estimates of generalization error such as the k-fold cross-validation error or the upper bound of leave-one-out(LOO) error. However, most of the approaches concentrate on the dual optimization problem of SVM. In this paper, we would like to consider the task of tuning hyperparameters in the primal. We derive a smooth validation function from the k-fold cross-validation, then tune hyperparameters by minimizing the smooth validation function using Quasi- Newton optimization technique. Experimental results not only show that our approach is much faster and provides more precise results than grid search method, but also demonstrate that tuning hyperparameters in the primal would be more efficient than in the dual due to advantages provided by the primal.