Improving the Finite Sample Performance of Double/Debiased Machine Learning with Propensity Score Calibration

Daniele Ballinari, Nora Bearth
{"title":"Improving the Finite Sample Performance of Double/Debiased Machine Learning with Propensity Score Calibration","authors":"Daniele Ballinari, Nora Bearth","doi":"arxiv-2409.04874","DOIUrl":null,"url":null,"abstract":"Machine learning techniques are widely used for estimating causal effects.\nDouble/debiased machine learning (DML) (Chernozhukov et al., 2018) uses a\ndouble-robust score function that relies on the prediction of nuisance\nfunctions, such as the propensity score, which is the probability of treatment\nassignment conditional on covariates. Estimators relying on double-robust score\nfunctions are highly sensitive to errors in propensity score predictions.\nMachine learners increase the severity of this problem as they tend to over- or\nunderestimate these probabilities. Several calibration approaches have been\nproposed to improve probabilistic forecasts of machine learners. This paper\ninvestigates the use of probability calibration approaches within the DML\nframework. Simulation results demonstrate that calibrating propensity scores\nmay significantly reduces the root mean squared error of DML estimates of the\naverage treatment effect in finite samples. We showcase it in an empirical\nexample and provide conditions under which calibration does not alter the\nasymptotic properties of the DML estimator.","PeriodicalId":501293,"journal":{"name":"arXiv - ECON - Econometrics","volume":"23 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - ECON - Econometrics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.04874","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Machine learning techniques are widely used for estimating causal effects. Double/debiased machine learning (DML) (Chernozhukov et al., 2018) uses a double-robust score function that relies on the prediction of nuisance functions, such as the propensity score, which is the probability of treatment assignment conditional on covariates. Estimators relying on double-robust score functions are highly sensitive to errors in propensity score predictions. Machine learners increase the severity of this problem as they tend to over- or underestimate these probabilities. Several calibration approaches have been proposed to improve probabilistic forecasts of machine learners. This paper investigates the use of probability calibration approaches within the DML framework. Simulation results demonstrate that calibrating propensity scores may significantly reduces the root mean squared error of DML estimates of the average treatment effect in finite samples. We showcase it in an empirical example and provide conditions under which calibration does not alter the asymptotic properties of the DML estimator.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
利用倾向得分校准提高双重/偏差机器学习的有限样本性能
双重/偏倚机器学习(DML)(Chernozhukov 等人,2018 年)使用双稳健得分函数,该函数依赖于对倾向得分等滋扰函数的预测,倾向得分是以协变量为条件的治疗分配概率。机器学习器往往会高估或低估这些概率,从而加剧了这一问题的严重性。为了改进机器学习器的概率预测,已经提出了几种校准方法。本文研究了在 DML 框架内使用概率校准方法的情况。模拟结果表明,校准倾向得分可以显著降低有限样本中 DML 估计平均治疗效果的均方根误差。我们在一个实证例子中展示了这一方法,并提供了校准不会改变 DML 估计器渐近特性的条件。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Simple robust two-stage estimation and inference for generalized impulse responses and multi-horizon causality GPT takes the SAT: Tracing changes in Test Difficulty and Math Performance of Students A Simple and Adaptive Confidence Interval when Nuisance Parameters Satisfy an Inequality Why you should also use OLS estimation of tail exponents On LASSO Inference for High Dimensional Predictive Regression
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1