应用控制理论和贝叶斯强化学习在大流行情况下的政策管理

Heena Rathore, A. Samant
{"title":"应用控制理论和贝叶斯强化学习在大流行情况下的政策管理","authors":"Heena Rathore, A. Samant","doi":"10.1109/ICCWorkshops50388.2021.9473604","DOIUrl":null,"url":null,"abstract":"As engineers and scientists, it is our responsibility to learn lessons from the recent pandemic outbreak and see how public health policies can be effectively managed to reduce the severe loss of lives and minimize the impact on people’s livelihood. Non-pharmaceutical interventions, such as in-place sheltering and social distancing, are typically introduced to slow the spread (flatten the curve) and reverse the growth of the virus. However, such approaches have the unintended consequences of causing economic activities to plummet and bringing local businesses to a standstill, thereby putting millions of jobs at risk. City administrators have generally resorted to an open loop, belief-based decision-making process, thereby struggling to manage (identify and enforce) timely and optimal policies. To overcome this challenge, this position paper explores a systematically designed, feedback-based strategy, to modulate parameters that control suppression and mitigation. Our work leverages advances in Bayesian Reinforcement Learning algorithms and known techniques in control theory, to stabilize and diminish the rate of propagation in pandemic situations. This paper discusses how offline exploitation using pre-trigger data, online exploration using observations from the environment, and a careful orchestration between the two using granular control of multiple on-off control signals can be used to modulate policy enforcement based on established metrics, such as reproduction number.","PeriodicalId":127186,"journal":{"name":"2021 IEEE International Conference on Communications Workshops (ICC Workshops)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2021-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Using Control Theory and Bayesian Reinforcement Learning for Policy Management in Pandemic Situations\",\"authors\":\"Heena Rathore, A. Samant\",\"doi\":\"10.1109/ICCWorkshops50388.2021.9473604\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"As engineers and scientists, it is our responsibility to learn lessons from the recent pandemic outbreak and see how public health policies can be effectively managed to reduce the severe loss of lives and minimize the impact on people’s livelihood. Non-pharmaceutical interventions, such as in-place sheltering and social distancing, are typically introduced to slow the spread (flatten the curve) and reverse the growth of the virus. However, such approaches have the unintended consequences of causing economic activities to plummet and bringing local businesses to a standstill, thereby putting millions of jobs at risk. City administrators have generally resorted to an open loop, belief-based decision-making process, thereby struggling to manage (identify and enforce) timely and optimal policies. To overcome this challenge, this position paper explores a systematically designed, feedback-based strategy, to modulate parameters that control suppression and mitigation. Our work leverages advances in Bayesian Reinforcement Learning algorithms and known techniques in control theory, to stabilize and diminish the rate of propagation in pandemic situations. This paper discusses how offline exploitation using pre-trigger data, online exploration using observations from the environment, and a careful orchestration between the two using granular control of multiple on-off control signals can be used to modulate policy enforcement based on established metrics, such as reproduction number.\",\"PeriodicalId\":127186,\"journal\":{\"name\":\"2021 IEEE International Conference on Communications Workshops (ICC Workshops)\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-06-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE International Conference on Communications Workshops (ICC Workshops)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCWorkshops50388.2021.9473604\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Conference on Communications Workshops (ICC Workshops)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCWorkshops50388.2021.9473604","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

作为工程师和科学家,我们有责任从最近的疫情中吸取教训,研究如何有效管理公共卫生政策,减少严重的生命损失,最大限度地减少对民生的影响。通常采取非药物干预措施,如就地庇护和保持社交距离,以减缓传播(使曲线变平)并扭转病毒的增长。然而,这种做法会产生意想不到的后果,导致经济活动急剧下降,使当地企业陷入停滞,从而使数百万个工作岗位面临风险。城市管理者通常采用开环、基于信念的决策过程,因此难以管理(识别和执行)及时和最佳的政策。为了克服这一挑战,本文探讨了一种系统设计的、基于反馈的策略,以调节控制抑制和缓解的参数。我们的工作利用了贝叶斯强化学习算法的进步和控制理论中的已知技术,以稳定和降低大流行情况下的传播速度。本文讨论了如何使用预触发数据进行离线开发,使用来自环境的观察进行在线探索,以及如何使用多个开关控制信号的粒度控制在两者之间进行精心编排,从而根据已建立的度量(如复制数)来调节策略实施。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Using Control Theory and Bayesian Reinforcement Learning for Policy Management in Pandemic Situations
As engineers and scientists, it is our responsibility to learn lessons from the recent pandemic outbreak and see how public health policies can be effectively managed to reduce the severe loss of lives and minimize the impact on people’s livelihood. Non-pharmaceutical interventions, such as in-place sheltering and social distancing, are typically introduced to slow the spread (flatten the curve) and reverse the growth of the virus. However, such approaches have the unintended consequences of causing economic activities to plummet and bringing local businesses to a standstill, thereby putting millions of jobs at risk. City administrators have generally resorted to an open loop, belief-based decision-making process, thereby struggling to manage (identify and enforce) timely and optimal policies. To overcome this challenge, this position paper explores a systematically designed, feedback-based strategy, to modulate parameters that control suppression and mitigation. Our work leverages advances in Bayesian Reinforcement Learning algorithms and known techniques in control theory, to stabilize and diminish the rate of propagation in pandemic situations. This paper discusses how offline exploitation using pre-trigger data, online exploration using observations from the environment, and a careful orchestration between the two using granular control of multiple on-off control signals can be used to modulate policy enforcement based on established metrics, such as reproduction number.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
BML: An Efficient and Versatile Tool for BGP Dataset Collection Efficient and Privacy-Preserving Contact Tracing System for Covid-19 using Blockchain MEC-Based Energy-Aware Distributed Feature Extraction for mHealth Applications with Strict Latency Requirements Distributed Multi-Agent Learning for Service Function Chain Partial Offloading at the Edge A Deep Neural Network Based Environment Sensing in the Presence of Jammers
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1