Machine-learned uncertainty quantification is not magic: Lessons learned from emulating radiative transfer with ML

Ryan Lagerquist, Imme Ebert-Uphoff, David D Turner, Jebb Q. Stewart
{"title":"Machine-learned uncertainty quantification is not magic: Lessons learned from emulating radiative transfer with ML","authors":"Ryan Lagerquist, Imme Ebert-Uphoff, David D Turner, Jebb Q. Stewart","doi":"10.22541/essoar.170000340.08902129/v1","DOIUrl":null,"url":null,"abstract":"Machine-learned uncertainty quantification (ML-UQ) has become a hot topic in environmental science, especially for neural networks. Scientists foresee the use of ML-UQ to make better decisions and assess the trustworthiness of the ML model. However, because ML-UQ is a new tool, its limitations are not yet fully appreciated. For example, some types of uncertainty are fundamentally unresolvable, including uncertainty that arises from data being out of sample, i.e. , outside the distribution of the training data. While it is generally recognized that ML-based point predictions (predictions without UQ) do not extrapolate well out of sample, this awareness does not exist for ML-based uncertainty. When point predictions have a large error, instead of accounting for this error by producing a wider confidence interval, ML-UQ often fails just as spectacularly. We demonstrate this problem by training ML with five different UQ methods to predict shortwave radiative transfer. The ML-UQ models are trained with real data but then tasked with generalizing to perturbed data containing, e.g. , fictitious cloud and ozone layers. We show that ML-UQ completely fails on the perturbed data, which are far outside the training distribution. We also show that when the training data are lightly perturbed – so that each basis vector of perturbation has a little variation in the training data – ML-UQ can extrapolate along the basis vectors with some success, leading to much better (but still somewhat concerning) performance on the validation and testing data. Overall, we wish to discourage overreliance on ML-UQ, especially in operational environments.","PeriodicalId":487619,"journal":{"name":"Authorea (Authorea)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2023-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Authorea (Authorea)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.22541/essoar.170000340.08902129/v1","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Machine-learned uncertainty quantification (ML-UQ) has become a hot topic in environmental science, especially for neural networks. Scientists foresee the use of ML-UQ to make better decisions and assess the trustworthiness of the ML model. However, because ML-UQ is a new tool, its limitations are not yet fully appreciated. For example, some types of uncertainty are fundamentally unresolvable, including uncertainty that arises from data being out of sample, i.e. , outside the distribution of the training data. While it is generally recognized that ML-based point predictions (predictions without UQ) do not extrapolate well out of sample, this awareness does not exist for ML-based uncertainty. When point predictions have a large error, instead of accounting for this error by producing a wider confidence interval, ML-UQ often fails just as spectacularly. We demonstrate this problem by training ML with five different UQ methods to predict shortwave radiative transfer. The ML-UQ models are trained with real data but then tasked with generalizing to perturbed data containing, e.g. , fictitious cloud and ozone layers. We show that ML-UQ completely fails on the perturbed data, which are far outside the training distribution. We also show that when the training data are lightly perturbed – so that each basis vector of perturbation has a little variation in the training data – ML-UQ can extrapolate along the basis vectors with some success, leading to much better (but still somewhat concerning) performance on the validation and testing data. Overall, we wish to discourage overreliance on ML-UQ, especially in operational environments.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
机器学习的不确定性量化不是魔法:用ML模拟辐射传递的经验教训
机器学习的不确定性量化(ML-UQ)已成为环境科学领域,尤其是神经网络领域的研究热点。科学家们预计ML- uq将用于做出更好的决策,并评估ML模型的可信度。然而,由于ML-UQ是一种新工具,它的局限性尚未得到充分认识。例如,某些类型的不确定性从根本上是无法解决的,包括由样本外的数据产生的不确定性,即在训练数据的分布之外。虽然人们普遍认为基于机器学习的点预测(没有UQ的预测)不能很好地推断出样本,但这种意识并不存在于基于机器学习的不确定性中。当点预测有很大的误差时,ML-UQ不是通过产生更宽的置信区间来解释这个误差,而是经常以惊人的方式失败。我们通过用五种不同的UQ方法训练ML来预测短波辐射传输来证明这个问题。ML-UQ模型是用真实数据训练的,但随后的任务是将其推广到包含虚构云和臭氧层的扰动数据。我们表明,ML-UQ在远离训练分布的扰动数据上完全失败。我们还表明,当训练数据受到轻微扰动时——这样每个扰动的基向量在训练数据中都有一点变化——ML-UQ可以沿着基向量成功地进行外推,从而在验证和测试数据上获得更好的(但仍然有些令人担忧的)性能。总的来说,我们希望阻止对ML-UQ的过度依赖,特别是在操作环境中。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Heart rate variability biofeedback acutely improves attentional control only in highly stressed individuals Relationship between microRNA-9 and breast cancer The impact of land use change on the diversity and emergence of fungal pathogens Severe seasonal shifts in tropical insect ephemerality drive bat foraging effort Using Circulating MicroRNAs as Noninvasive Cancer Biomarkers in Breast Cancer is a Cutting-Edge Application of MicroRNA Profiling Technology
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1