Optimal Ensembles for Deep Learning Classification: Theory and Practice

Wenjing Li, R. Paffenroth
{"title":"Optimal Ensembles for Deep Learning Classification: Theory and Practice","authors":"Wenjing Li, R. Paffenroth","doi":"10.1109/ICMLA.2019.00271","DOIUrl":null,"url":null,"abstract":"Ensemble methods for classification problems construct a set of models, often called \"learners\", and then assign class labels to new data points by taking a combination of the predictions from these models. Ensemble methods are popular and used in a wide range of problem domains because of their good performance. However, a theoretical understanding of the optimality of ensembles is, in many instances, an open problem. In particular, improving the performance of an ensemble requires an understanding of the subtle interplay between the accuracy of the individual learners and the diversity of the learners in the ensemble. For example, if all of the learners in an ensemble were identical, then clearly the accuracy of the ensemble cannot be any better than the accuracy of the individual learning, no matter how many learners one were to use. Accordingly, here we develop a theory for understanding when ensembles are optimal, in an appropriate sense, by balancing individual accuracy against ensemble diversity, from the perspective of statistical correlations. The theory that we derive is applicable for many practical ensembles, and we provide a set of metrics for assessing the optimality of any given ensemble. Perhaps most interestingly, the metrics that we develop lead naturally to a set of novel loss functions that can be optimized using backpropagation giving rise to optimal deep neural network based ensembles. We demonstrate the effectiveness of these deep neural network based ensembles using standard benchmark data sets.","PeriodicalId":436714,"journal":{"name":"2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA)","volume":"102 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMLA.2019.00271","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Ensemble methods for classification problems construct a set of models, often called "learners", and then assign class labels to new data points by taking a combination of the predictions from these models. Ensemble methods are popular and used in a wide range of problem domains because of their good performance. However, a theoretical understanding of the optimality of ensembles is, in many instances, an open problem. In particular, improving the performance of an ensemble requires an understanding of the subtle interplay between the accuracy of the individual learners and the diversity of the learners in the ensemble. For example, if all of the learners in an ensemble were identical, then clearly the accuracy of the ensemble cannot be any better than the accuracy of the individual learning, no matter how many learners one were to use. Accordingly, here we develop a theory for understanding when ensembles are optimal, in an appropriate sense, by balancing individual accuracy against ensemble diversity, from the perspective of statistical correlations. The theory that we derive is applicable for many practical ensembles, and we provide a set of metrics for assessing the optimality of any given ensemble. Perhaps most interestingly, the metrics that we develop lead naturally to a set of novel loss functions that can be optimized using backpropagation giving rise to optimal deep neural network based ensembles. We demonstrate the effectiveness of these deep neural network based ensembles using standard benchmark data sets.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
深度学习分类的最佳集成:理论与实践
用于分类问题的集成方法构建一组模型,通常称为“学习者”,然后通过从这些模型中获得预测的组合为新的数据点分配类标签。集成方法因其良好的性能而广泛应用于各种问题领域。然而,在许多情况下,对集成的最优性的理论理解是一个开放的问题。特别是,提高合奏的性能需要理解单个学习器的准确性和合奏中学习器的多样性之间微妙的相互作用。例如,如果集合中的所有学习器都是相同的,那么很明显,无论使用多少个学习器,集合的准确性都不会比单个学习的准确性好。因此,从统计相关性的角度来看,我们通过平衡个体准确性和集合多样性,在适当的意义上,开发了一种理论来理解何时集合是最佳的。我们推导的理论适用于许多实际的集成,并且我们提供了一组度量来评估任何给定集成的最优性。也许最有趣的是,我们开发的指标自然会导致一组新的损失函数,这些损失函数可以使用反向传播进行优化,从而产生最佳的基于深度神经网络的集成。我们使用标准基准数据集证明了这些基于深度神经网络的集成的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Automated Stenosis Classification of Carotid Artery Sonography using Deep Neural Networks Hybrid Condition Monitoring for Power Electronic Systems Time Series Anomaly Detection from a Markov Chain Perspective Anyone here? Smart Embedded Low-Resolution Omnidirectional Video Sensor to Measure Room Occupancy Deep Learning with Domain Randomization for Optimal Filtering
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1