机器学习中有效不确定性量化的保形预测方法比较研究

Nicolas Dewolf
{"title":"机器学习中有效不确定性量化的保形预测方法比较研究","authors":"Nicolas Dewolf","doi":"arxiv-2405.02082","DOIUrl":null,"url":null,"abstract":"In the past decades, most work in the area of data analysis and machine\nlearning was focused on optimizing predictive models and getting better results\nthan what was possible with existing models. To what extent the metrics with\nwhich such improvements were measured were accurately capturing the intended\ngoal, whether the numerical differences in the resulting values were\nsignificant, or whether uncertainty played a role in this study and if it\nshould have been taken into account, was of secondary importance. Whereas\nprobability theory, be it frequentist or Bayesian, used to be the gold standard\nin science before the advent of the supercomputer, it was quickly replaced in\nfavor of black box models and sheer computing power because of their ability to\nhandle large data sets. This evolution sadly happened at the expense of\ninterpretability and trustworthiness. However, while people are still trying to\nimprove the predictive power of their models, the community is starting to\nrealize that for many applications it is not so much the exact prediction that\nis of importance, but rather the variability or uncertainty. The work in this dissertation tries to further the quest for a world where\neveryone is aware of uncertainty, of how important it is and how to embrace it\ninstead of fearing it. A specific, though general, framework that allows anyone\nto obtain accurate uncertainty estimates is singled out and analysed. Certain\naspects and applications of the framework -- dubbed `conformal prediction' --\nare studied in detail. Whereas many approaches to uncertainty quantification\nmake strong assumptions about the data, conformal prediction is, at the time of\nwriting, the only framework that deserves the title `distribution-free'. No\nparametric assumptions have to be made and the nonparametric results also hold\nwithout having to resort to the law of large numbers in the asymptotic regime.","PeriodicalId":501330,"journal":{"name":"arXiv - MATH - Statistics Theory","volume":"16 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A comparative study of conformal prediction methods for valid uncertainty quantification in machine learning\",\"authors\":\"Nicolas Dewolf\",\"doi\":\"arxiv-2405.02082\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In the past decades, most work in the area of data analysis and machine\\nlearning was focused on optimizing predictive models and getting better results\\nthan what was possible with existing models. To what extent the metrics with\\nwhich such improvements were measured were accurately capturing the intended\\ngoal, whether the numerical differences in the resulting values were\\nsignificant, or whether uncertainty played a role in this study and if it\\nshould have been taken into account, was of secondary importance. Whereas\\nprobability theory, be it frequentist or Bayesian, used to be the gold standard\\nin science before the advent of the supercomputer, it was quickly replaced in\\nfavor of black box models and sheer computing power because of their ability to\\nhandle large data sets. This evolution sadly happened at the expense of\\ninterpretability and trustworthiness. However, while people are still trying to\\nimprove the predictive power of their models, the community is starting to\\nrealize that for many applications it is not so much the exact prediction that\\nis of importance, but rather the variability or uncertainty. The work in this dissertation tries to further the quest for a world where\\neveryone is aware of uncertainty, of how important it is and how to embrace it\\ninstead of fearing it. A specific, though general, framework that allows anyone\\nto obtain accurate uncertainty estimates is singled out and analysed. Certain\\naspects and applications of the framework -- dubbed `conformal prediction' --\\nare studied in detail. Whereas many approaches to uncertainty quantification\\nmake strong assumptions about the data, conformal prediction is, at the time of\\nwriting, the only framework that deserves the title `distribution-free'. No\\nparametric assumptions have to be made and the nonparametric results also hold\\nwithout having to resort to the law of large numbers in the asymptotic regime.\",\"PeriodicalId\":501330,\"journal\":{\"name\":\"arXiv - MATH - Statistics Theory\",\"volume\":\"16 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-05-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - MATH - Statistics Theory\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2405.02082\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - MATH - Statistics Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.02082","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

在过去的几十年中,数据分析和机器学习领域的大部分工作都集中在优化预测模型和获得比现有模型更好的结果上。至于衡量这种改进的指标在多大程度上准确地捕捉到了预期目标,得出的数值差异是否显著,或者不确定性在本研究中是否发挥作用以及是否应该将其考虑在内,这些都是次要的。在超级计算机出现之前,无论是频数理论还是贝叶斯理论,概率论都曾是科学界的黄金标准,但很快就被黑箱模型和纯粹的计算能力所取代,因为它们能够处理大量数据集。可悲的是,这种演变是以牺牲可解释性和可信度为代价的。不过,尽管人们仍在努力提高模型的预测能力,但社会各界已经开始意识到,对于许多应用来说,重要的不是精确预测,而是可变性或不确定性。本论文中的工作试图进一步追求一个人人都能意识到不确定性、不确定性的重要性以及如何拥抱不确定性而不是惧怕不确定性的世界。论文挑出并分析了一个允许任何人获得准确的不确定性估计的具体(尽管是一般)框架。该框架被称为 "共形预测",对其某些方面和应用进行了详细研究。许多不确定性量化方法都对数据做出了强有力的假设,而保形预测在本文撰写时是唯一一个配得上 "无分布 "称号的框架。它不需要任何参数假设,而且非参数结果也是成立的,无需在渐近机制中求助于大数定律。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
A comparative study of conformal prediction methods for valid uncertainty quantification in machine learning
In the past decades, most work in the area of data analysis and machine learning was focused on optimizing predictive models and getting better results than what was possible with existing models. To what extent the metrics with which such improvements were measured were accurately capturing the intended goal, whether the numerical differences in the resulting values were significant, or whether uncertainty played a role in this study and if it should have been taken into account, was of secondary importance. Whereas probability theory, be it frequentist or Bayesian, used to be the gold standard in science before the advent of the supercomputer, it was quickly replaced in favor of black box models and sheer computing power because of their ability to handle large data sets. This evolution sadly happened at the expense of interpretability and trustworthiness. However, while people are still trying to improve the predictive power of their models, the community is starting to realize that for many applications it is not so much the exact prediction that is of importance, but rather the variability or uncertainty. The work in this dissertation tries to further the quest for a world where everyone is aware of uncertainty, of how important it is and how to embrace it instead of fearing it. A specific, though general, framework that allows anyone to obtain accurate uncertainty estimates is singled out and analysed. Certain aspects and applications of the framework -- dubbed `conformal prediction' -- are studied in detail. Whereas many approaches to uncertainty quantification make strong assumptions about the data, conformal prediction is, at the time of writing, the only framework that deserves the title `distribution-free'. No parametric assumptions have to be made and the nonparametric results also hold without having to resort to the law of large numbers in the asymptotic regime.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Precision-based designs for sequential randomized experiments Strang Splitting for Parametric Inference in Second-order Stochastic Differential Equations Stability of a Generalized Debiased Lasso with Applications to Resampling-Based Variable Selection Tuning parameter selection in econometrics Limiting Behavior of Maxima under Dependence
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1