{"title":"boosting和正则boosting算法的过度拟合","authors":"Takashi Onoda","doi":"10.1002/ecjc.20344","DOIUrl":null,"url":null,"abstract":"<p>The impressive generalization capacity of AdaBoost has been explained using the concept of a <i>margin</i> introduced in the context of support vector machines. However, this ability to generalize is limited to cases where the data does not include misclassification errors or significant amounts of noise. In addition, the research of Schapire and colleagues has served to provide theoretical support for these results from the perspective of improving margins. In this paper we propose a set of new algorithms, AdaBoost<sub><i>Reg</i></sub>,ν-Arc, and ν-Boost, that attempt to avoid the overfitting that can occur with AdaBoost by introducing a normalization term into the objective function minimized by AdaBoost. © 2007 Wiley Periodicals, Inc. Electron Comm Jpn Pt 3, 90(9): 69– 78, 2007; Published online in Wiley InterScience (www.interscience.wiley.com). DOI 10.1002/ecjc.20344</p>","PeriodicalId":100407,"journal":{"name":"Electronics and Communications in Japan (Part III: Fundamental Electronic Science)","volume":"90 9","pages":"69-78"},"PeriodicalIF":0.0000,"publicationDate":"2007-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1002/ecjc.20344","citationCount":"3","resultStr":"{\"title\":\"Overfitting of boosting and regularized Boosting algorithms\",\"authors\":\"Takashi Onoda\",\"doi\":\"10.1002/ecjc.20344\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>The impressive generalization capacity of AdaBoost has been explained using the concept of a <i>margin</i> introduced in the context of support vector machines. However, this ability to generalize is limited to cases where the data does not include misclassification errors or significant amounts of noise. In addition, the research of Schapire and colleagues has served to provide theoretical support for these results from the perspective of improving margins. In this paper we propose a set of new algorithms, AdaBoost<sub><i>Reg</i></sub>,ν-Arc, and ν-Boost, that attempt to avoid the overfitting that can occur with AdaBoost by introducing a normalization term into the objective function minimized by AdaBoost. © 2007 Wiley Periodicals, Inc. Electron Comm Jpn Pt 3, 90(9): 69– 78, 2007; Published online in Wiley InterScience (www.interscience.wiley.com). DOI 10.1002/ecjc.20344</p>\",\"PeriodicalId\":100407,\"journal\":{\"name\":\"Electronics and Communications in Japan (Part III: Fundamental Electronic Science)\",\"volume\":\"90 9\",\"pages\":\"69-78\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2007-04-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://sci-hub-pdf.com/10.1002/ecjc.20344\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Electronics and Communications in Japan (Part III: Fundamental Electronic Science)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/ecjc.20344\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Electronics and Communications in Japan (Part III: Fundamental Electronic Science)","FirstCategoryId":"1085","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/ecjc.20344","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3