Run-Min Ji, Xiang-Hua Huang, Xing-Long Zhang, Ling-Wei Li
{"title":"基于涡轮螺旋桨发动机动力管理系统和深度强化学习的单杆控制方法设计","authors":"Run-Min Ji, Xiang-Hua Huang, Xing-Long Zhang, Ling-Wei Li","doi":"10.1177/09544100241235824","DOIUrl":null,"url":null,"abstract":"This paper presents a single-lever control method based on Power Management System (PMS) and improved Twin Delayed Deep Deterministic Policy Gradient (TD3) algorithm for turboprop engines. In this approach, power level angle command, which is the single-lever command, is decoupled into controlled variable commands by PMS, and the controller based on improved TD3 algorithm can ensure that controlled variables track their commands rapidly and accurately. To achieve the optimal conversion relationship between different commands, an offline optimization process is used to design PMS. By optimization, specific fuel consumption and propeller efficiency are both improved after conversion. To deal with strong interactions between different control loops of a turboprop engine, TD3 algorithm which is a deep reinforcement learning algorithm is adopted. Two improvements which are the design method of observation state and prioritized experience replay are made to enhance the tracking accuracy. Simulation results show that improved TD3 algorithm can learn an optimal control policy to guarantee good control effect with fast response and small overshoot. The maximum settling time is less than 0.25s and the maximum overshoot is less than 0.1%. It also has a good robustness performance when the plant exists model uncertainties. The maximum fluctuations are less than 0.05%.","PeriodicalId":54566,"journal":{"name":"Proceedings of the Institution of Mechanical Engineers Part G-Journal of Aerospace Engineering","volume":"45 1","pages":""},"PeriodicalIF":1.0000,"publicationDate":"2024-02-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Single-lever control method design based on power management system and deep reinforcement learning for turboprop engines\",\"authors\":\"Run-Min Ji, Xiang-Hua Huang, Xing-Long Zhang, Ling-Wei Li\",\"doi\":\"10.1177/09544100241235824\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper presents a single-lever control method based on Power Management System (PMS) and improved Twin Delayed Deep Deterministic Policy Gradient (TD3) algorithm for turboprop engines. In this approach, power level angle command, which is the single-lever command, is decoupled into controlled variable commands by PMS, and the controller based on improved TD3 algorithm can ensure that controlled variables track their commands rapidly and accurately. To achieve the optimal conversion relationship between different commands, an offline optimization process is used to design PMS. By optimization, specific fuel consumption and propeller efficiency are both improved after conversion. To deal with strong interactions between different control loops of a turboprop engine, TD3 algorithm which is a deep reinforcement learning algorithm is adopted. Two improvements which are the design method of observation state and prioritized experience replay are made to enhance the tracking accuracy. Simulation results show that improved TD3 algorithm can learn an optimal control policy to guarantee good control effect with fast response and small overshoot. The maximum settling time is less than 0.25s and the maximum overshoot is less than 0.1%. It also has a good robustness performance when the plant exists model uncertainties. The maximum fluctuations are less than 0.05%.\",\"PeriodicalId\":54566,\"journal\":{\"name\":\"Proceedings of the Institution of Mechanical Engineers Part G-Journal of Aerospace Engineering\",\"volume\":\"45 1\",\"pages\":\"\"},\"PeriodicalIF\":1.0000,\"publicationDate\":\"2024-02-29\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the Institution of Mechanical Engineers Part G-Journal of Aerospace Engineering\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://doi.org/10.1177/09544100241235824\",\"RegionNum\":4,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"ENGINEERING, AEROSPACE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the Institution of Mechanical Engineers Part G-Journal of Aerospace Engineering","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1177/09544100241235824","RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, AEROSPACE","Score":null,"Total":0}
Single-lever control method design based on power management system and deep reinforcement learning for turboprop engines
This paper presents a single-lever control method based on Power Management System (PMS) and improved Twin Delayed Deep Deterministic Policy Gradient (TD3) algorithm for turboprop engines. In this approach, power level angle command, which is the single-lever command, is decoupled into controlled variable commands by PMS, and the controller based on improved TD3 algorithm can ensure that controlled variables track their commands rapidly and accurately. To achieve the optimal conversion relationship between different commands, an offline optimization process is used to design PMS. By optimization, specific fuel consumption and propeller efficiency are both improved after conversion. To deal with strong interactions between different control loops of a turboprop engine, TD3 algorithm which is a deep reinforcement learning algorithm is adopted. Two improvements which are the design method of observation state and prioritized experience replay are made to enhance the tracking accuracy. Simulation results show that improved TD3 algorithm can learn an optimal control policy to guarantee good control effect with fast response and small overshoot. The maximum settling time is less than 0.25s and the maximum overshoot is less than 0.1%. It also has a good robustness performance when the plant exists model uncertainties. The maximum fluctuations are less than 0.05%.
期刊介绍:
The Journal of Aerospace Engineering is dedicated to the publication of high quality research in all branches of applied sciences and technology dealing with aircraft and spacecraft, and their support systems. "Our authorship is truly international and all efforts are made to ensure that each paper is presented in the best possible way and reaches a wide audience.
"The Editorial Board is composed of recognized experts representing the technical communities of fifteen countries. The Board Members work in close cooperation with the editors, reviewers, and authors to achieve a consistent standard of well written and presented papers."Professor Rodrigo Martinez-Val, Universidad Politécnica de Madrid, Spain
This journal is a member of the Committee on Publication Ethics (COPE).