Bandit algorithms for policy learning: methods, implementation, and welfare-performance

Toru Kitagawa, Jeff Rowley
{"title":"Bandit algorithms for policy learning: methods, implementation, and welfare-performance","authors":"Toru Kitagawa, Jeff Rowley","doi":"10.1007/s42973-024-00165-6","DOIUrl":null,"url":null,"abstract":"<p>Static supervised learning—in which experimental data serves as a training sample for the estimation of an optimal treatment assignment policy—is a commonly assumed framework of policy learning. An arguably more realistic but challenging scenario is a dynamic setting in which the planner performs experimentation and exploitation simultaneously with subjects that arrive sequentially. This paper studies bandit algorithms for learning an optimal individualised treatment assignment policy. Specifically, we study applicability of the EXP4.P (Exponential weighting for Exploration and Exploitation with Experts) algorithm developed by Beygelzimer et al. (Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, JMLR Workshop and Conference Proceedings, pp 19–26, 2011) to policy learning. Assuming that the class of policies has a finite Vapnik–Chervonenkis dimension and that the number of subjects to be allocated is known, we present a high probability welfare-regret bound of the algorithm. To implement the algorithm, we use an incremental enumeration algorithm for hyperplane arrangements. We perform extensive numerical analysis to assess the algorithm’s sensitivity to its tuning parameters and its welfare-regret performance. Further simulation exercises are calibrated to the National Job Training Partnership Act (JTPA) Study sample to determine how the algorithm performs when applied to economic data. Our findings highlight various computational challenges and suggest that the limited welfare gain from the algorithm is due to substantial heterogeneity in causal effects in the JTPA data.</p>","PeriodicalId":516533,"journal":{"name":"The Japanese Economic Review","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"The Japanese Economic Review","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s42973-024-00165-6","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Static supervised learning—in which experimental data serves as a training sample for the estimation of an optimal treatment assignment policy—is a commonly assumed framework of policy learning. An arguably more realistic but challenging scenario is a dynamic setting in which the planner performs experimentation and exploitation simultaneously with subjects that arrive sequentially. This paper studies bandit algorithms for learning an optimal individualised treatment assignment policy. Specifically, we study applicability of the EXP4.P (Exponential weighting for Exploration and Exploitation with Experts) algorithm developed by Beygelzimer et al. (Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, JMLR Workshop and Conference Proceedings, pp 19–26, 2011) to policy learning. Assuming that the class of policies has a finite Vapnik–Chervonenkis dimension and that the number of subjects to be allocated is known, we present a high probability welfare-regret bound of the algorithm. To implement the algorithm, we use an incremental enumeration algorithm for hyperplane arrangements. We perform extensive numerical analysis to assess the algorithm’s sensitivity to its tuning parameters and its welfare-regret performance. Further simulation exercises are calibrated to the National Job Training Partnership Act (JTPA) Study sample to determine how the algorithm performs when applied to economic data. Our findings highlight various computational challenges and suggest that the limited welfare gain from the algorithm is due to substantial heterogeneity in causal effects in the JTPA data.

Abstract Image

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
政策学习的 Bandit 算法:方法、实施和福利绩效
静态监督学习--即实验数据作为训练样本,用于估计最优处理分配政策--是政策学习的一个常见假设框架。可以说,更现实但更具挑战性的情况是动态设置,即计划者同时对按顺序到达的受试者进行实验和利用。本文研究了学习最优个性化治疗分配政策的匪徒算法。具体来说,我们研究了由 Beygelzimer 等人开发的 EXP4.P(Exponential weighting for Exploration and Exploitation with Experts)算法(第十四届国际人工智能和统计学会议论文集,JMLR 研讨会和会议论文集,第 19-26 页,2011 年)在政策学习中的适用性。假定政策类具有有限的 Vapnik-Chervonenkis 维度,且待分配对象的数量已知,我们提出了该算法的高概率福利-遗憾约束。为了实现该算法,我们使用了超平面排列增量枚举算法。我们进行了大量的数值分析,以评估该算法对其调整参数的敏感性及其福利-遗憾性能。此外,我们还根据《国家就业培训合作法案》(JTPA)研究样本进行了进一步的模拟练习,以确定该算法在应用于经济数据时的表现。我们的研究结果凸显了各种计算挑战,并表明该算法的福利收益有限是由于 JTPA 数据中因果效应的巨大异质性造成的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Bandit algorithms for policy learning: methods, implementation, and welfare-performance Robustness in binary-action supermodular games revisited The impact of the Bank of Japan’s exchange traded fund and corporate bond purchases on firms’ capital structure Evidence on price stickiness in Japan Pass-through and tax incidence in Cournot duopoly with loan commitment
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1