{"title":"Research and Application of a Method for Constructing Decision Forests","authors":"Zhu Qiang","doi":"10.1109/ICSSSM.2007.4280168","DOIUrl":null,"url":null,"abstract":"Decision forest (DF) is an implementation of multiple classifier system, which has been introduced to overcome the flaw of a single decision tree. In this paper, a method named the random subspace method (RSM) for constructing DF is investigated. RSM can keep perfect accuracy on training data while having desirable generalization accuracy. Moreover, its accuracy continues to increase as DF becomes larger, exhibiting a characteristic of overtraining resistant. The underlying theory of this model is presented in this paper. A comparison of three combination methods for this model voting, Bayesian method and neural-network is carried out. The power of voting is demonstrated both theoretically and experimentally. As for the application of this method, the superiority of RSM is explored and an advice regarding the size of attribute subsets is given.","PeriodicalId":153603,"journal":{"name":"2007 International Conference on Service Systems and Service Management","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2007-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2007 International Conference on Service Systems and Service Management","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSSSM.2007.4280168","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Decision forest (DF) is an implementation of multiple classifier system, which has been introduced to overcome the flaw of a single decision tree. In this paper, a method named the random subspace method (RSM) for constructing DF is investigated. RSM can keep perfect accuracy on training data while having desirable generalization accuracy. Moreover, its accuracy continues to increase as DF becomes larger, exhibiting a characteristic of overtraining resistant. The underlying theory of this model is presented in this paper. A comparison of three combination methods for this model voting, Bayesian method and neural-network is carried out. The power of voting is demonstrated both theoretically and experimentally. As for the application of this method, the superiority of RSM is explored and an advice regarding the size of attribute subsets is given.