{"title":"新机器学习算法可行性的快速初步评估","authors":"Dustin Baumgartner, G. Serpen","doi":"10.1109/ICMLC.2010.31","DOIUrl":null,"url":null,"abstract":"Traditionally, researchers compare the performance of new machine learning algorithms against those of locally executed simulations that serve as benchmarks. This process requires considerable time, computation resources, and expertise. In this paper, we present a method to quickly evaluate the performance feasibility of new algorithms – offering a preliminary study that either supports or opposes the need to conduct a full-scale traditional evaluation, and possibly saving valuable resources for researchers. The proposed method uses performance benchmarks obtained from results reported in the literature rather than local simulations. Furthermore, an alternate statistical technique is suggested for comparative performance analysis, since traditional statistical significance tests do not fit the problem well. We highlight the use of the proposed evaluation method in a study that compared a new algorithm against 47 other algorithms across 46 datasets.","PeriodicalId":423912,"journal":{"name":"2010 Second International Conference on Machine Learning and Computing","volume":"61 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Fast Preliminary Evaluation of New Machine Learning Algorithms for Feasibility\",\"authors\":\"Dustin Baumgartner, G. Serpen\",\"doi\":\"10.1109/ICMLC.2010.31\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Traditionally, researchers compare the performance of new machine learning algorithms against those of locally executed simulations that serve as benchmarks. This process requires considerable time, computation resources, and expertise. In this paper, we present a method to quickly evaluate the performance feasibility of new algorithms – offering a preliminary study that either supports or opposes the need to conduct a full-scale traditional evaluation, and possibly saving valuable resources for researchers. The proposed method uses performance benchmarks obtained from results reported in the literature rather than local simulations. Furthermore, an alternate statistical technique is suggested for comparative performance analysis, since traditional statistical significance tests do not fit the problem well. We highlight the use of the proposed evaluation method in a study that compared a new algorithm against 47 other algorithms across 46 datasets.\",\"PeriodicalId\":423912,\"journal\":{\"name\":\"2010 Second International Conference on Machine Learning and Computing\",\"volume\":\"61 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2010-02-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2010 Second International Conference on Machine Learning and Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICMLC.2010.31\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 Second International Conference on Machine Learning and Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMLC.2010.31","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Fast Preliminary Evaluation of New Machine Learning Algorithms for Feasibility
Traditionally, researchers compare the performance of new machine learning algorithms against those of locally executed simulations that serve as benchmarks. This process requires considerable time, computation resources, and expertise. In this paper, we present a method to quickly evaluate the performance feasibility of new algorithms – offering a preliminary study that either supports or opposes the need to conduct a full-scale traditional evaluation, and possibly saving valuable resources for researchers. The proposed method uses performance benchmarks obtained from results reported in the literature rather than local simulations. Furthermore, an alternate statistical technique is suggested for comparative performance analysis, since traditional statistical significance tests do not fit the problem well. We highlight the use of the proposed evaluation method in a study that compared a new algorithm against 47 other algorithms across 46 datasets.