Adaptive Tuning of Dynamic Matrix Control for Uncertain Industrial Systems With Deep Reinforcement Learning

IF 6.4 2区 计算机科学 Q1 AUTOMATION & CONTROL SYSTEMS IEEE Transactions on Automation Science and Engineering Pub Date : 2024-11-06 DOI:10.1109/TASE.2024.3487878
Yang Zhang;Peng Wang;Liying Yu;Ning Li
{"title":"Adaptive Tuning of Dynamic Matrix Control for Uncertain Industrial Systems With Deep Reinforcement Learning","authors":"Yang Zhang;Peng Wang;Liying Yu;Ning Li","doi":"10.1109/TASE.2024.3487878","DOIUrl":null,"url":null,"abstract":"Dynamic matrix control (DMC) has been field-validated in many industrial practices, highlighting the critical importance of fine-tuning parameters for optimal performance. However, the tuning of well-performed parameters is challenging because the relationship between parameters and the performance of DMC is intricate to characterize for industrial systems with uncertainty. An adaptive tuning approach based on deep reinforcement learning (DRL) is proposed to optimize the performance of DMC for uncertain systems in this paper. The approach can online tune the horizons and weighting matrices of DMC in real time adaptive to the state and uncertainty of the systems. Compared with offline tuning approaches, the proposed approach does not need to tradeoff optimality for robustness. The proposed approach utilizes various state-of-the-art DRL algorithms, e.g., value-based and actor-critic-based, to develop online parameter tuning policies that can adapt to system uncertainty. A piecewise reward function is designed to improve the performance and stability of the agent. A novel predictor-switching criterion is developed to address the horizon inconsistency in the receding optimization process. The proposed approaches are validated by the moisture control task in industrial cigarette drying process. Note to Practitioners—This paper is motivated by the adaptive tuning problem of dynamic matrix control (DMC) in uncertain industrial systems. For other nonlinear industrial scenarios, practitioners should first design a nonlinear model predictive controller suitable for the controlled object. Then, they can refer to the proposed tuning algorithm to improve the controller performance. Specifically, regarding the setting of state and action sets, please refer to the technical details provided in this paper. The reward function can be flexibly set according to the needs of practitioners for the controller, e.g., improving the dynamic performance of the controller or saving controller energy consumption. The design processes of the tuning algorithms can refer to <xref>Algorithm 1</xref> and <xref>Algorithm 2</xref>. There are two reasons why the proposed algorithm can be easily transferred to nonlinear industrial scenarios. First, the proposed algorithm does not restrict the model predictive controller or controlled object type. Second, the requirement for horizons or weighting matrices tuning widely exists in the practical applications of various model predictive algorithms.","PeriodicalId":51060,"journal":{"name":"IEEE Transactions on Automation Science and Engineering","volume":"22 ","pages":"8695-8708"},"PeriodicalIF":6.4000,"publicationDate":"2024-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Automation Science and Engineering","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10745630/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Dynamic matrix control (DMC) has been field-validated in many industrial practices, highlighting the critical importance of fine-tuning parameters for optimal performance. However, the tuning of well-performed parameters is challenging because the relationship between parameters and the performance of DMC is intricate to characterize for industrial systems with uncertainty. An adaptive tuning approach based on deep reinforcement learning (DRL) is proposed to optimize the performance of DMC for uncertain systems in this paper. The approach can online tune the horizons and weighting matrices of DMC in real time adaptive to the state and uncertainty of the systems. Compared with offline tuning approaches, the proposed approach does not need to tradeoff optimality for robustness. The proposed approach utilizes various state-of-the-art DRL algorithms, e.g., value-based and actor-critic-based, to develop online parameter tuning policies that can adapt to system uncertainty. A piecewise reward function is designed to improve the performance and stability of the agent. A novel predictor-switching criterion is developed to address the horizon inconsistency in the receding optimization process. The proposed approaches are validated by the moisture control task in industrial cigarette drying process. Note to Practitioners—This paper is motivated by the adaptive tuning problem of dynamic matrix control (DMC) in uncertain industrial systems. For other nonlinear industrial scenarios, practitioners should first design a nonlinear model predictive controller suitable for the controlled object. Then, they can refer to the proposed tuning algorithm to improve the controller performance. Specifically, regarding the setting of state and action sets, please refer to the technical details provided in this paper. The reward function can be flexibly set according to the needs of practitioners for the controller, e.g., improving the dynamic performance of the controller or saving controller energy consumption. The design processes of the tuning algorithms can refer to Algorithm 1 and Algorithm 2. There are two reasons why the proposed algorithm can be easily transferred to nonlinear industrial scenarios. First, the proposed algorithm does not restrict the model predictive controller or controlled object type. Second, the requirement for horizons or weighting matrices tuning widely exists in the practical applications of various model predictive algorithms.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
利用深度强化学习对不确定工业系统的动态矩阵控制进行自适应调整
动态矩阵控制(DMC)已经在许多工业实践中得到了现场验证,强调了微调参数对优化性能的重要性。然而,性能良好的参数的调整是具有挑战性的,因为参数与DMC性能之间的关系对于具有不确定性的工业系统来说是复杂的。本文提出了一种基于深度强化学习(DRL)的自适应调谐方法来优化不确定系统的DMC性能。该方法可以实时在线调整DMC的视界和权重矩阵,以适应系统的状态和不确定性。与离线调优方法相比,该方法不需要为了鲁棒性而权衡最优性。所提出的方法利用各种最先进的DRL算法,例如,基于值的和基于参与者关键的,来开发能够适应系统不确定性的在线参数调整策略。为了提高智能体的性能和稳定性,设计了分段奖励函数。提出了一种新的预测切换判据,解决了后退优化过程中的水平不一致问题。通过工业卷烟干燥过程中的水分控制任务,验证了所提方法的有效性。本文的灵感来自于不确定工业系统中动态矩阵控制(DMC)的自适应调谐问题。对于其他非线性工业场景,从业者应首先设计适合被控对象的非线性模型预测控制器。然后,他们可以参考所提出的调谐算法来提高控制器的性能。具体来说,关于状态集和动作集的设置,请参考本文提供的技术细节。奖励函数可以根据从业人员对控制器的需求灵活设置,例如改善控制器的动态性能或节省控制器的能耗。调优算法的设计过程可参考算法1和算法2。有两个原因,为什么所提出的算法可以很容易地转移到非线性工业场景。首先,该算法不限制模型预测控制器和被控对象类型。其次,在各种模型预测算法的实际应用中广泛存在对视界或加权矩阵调优的要求。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE Transactions on Automation Science and Engineering
IEEE Transactions on Automation Science and Engineering 工程技术-自动化与控制系统
CiteScore
12.50
自引率
14.30%
发文量
404
审稿时长
3.0 months
期刊介绍: The IEEE Transactions on Automation Science and Engineering (T-ASE) publishes fundamental papers on Automation, emphasizing scientific results that advance efficiency, quality, productivity, and reliability. T-ASE encourages interdisciplinary approaches from computer science, control systems, electrical engineering, mathematics, mechanical engineering, operations research, and other fields. T-ASE welcomes results relevant to industries such as agriculture, biotechnology, healthcare, home automation, maintenance, manufacturing, pharmaceuticals, retail, security, service, supply chains, and transportation. T-ASE addresses a research community willing to integrate knowledge across disciplines and industries. For this purpose, each paper includes a Note to Practitioners that summarizes how its results can be applied or how they might be extended to apply in practice.
期刊最新文献
Automation 5.0: The Step to Systems Intelligence for a Sustainable Future Dual-Layer Bumpless Transfer Control for Markovian Jump Systems and Its Applications: A Weighted Coefficient Mechanism Continuous-time/event-triggered decentralized output feedback prescribed-time control applied to a 2-DOF helicopter High-Precision Tracking Control of Multi-Axis Electro-Optical Systems Based on Accurate Identification of Dynamic Parameters Distributed Adaptive Secondary Control of DC Microgrids with Uncertainties: A Real-time Parameter Estimation Approach
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1