Chen Wang , Tianyu Dong , Lei Chen , Guixiang Zhu , Yihan Chen
{"title":"通过基于深度强化学习的改进型软演员评判器实现永磁机械的多目标优化方法","authors":"Chen Wang , Tianyu Dong , Lei Chen , Guixiang Zhu , Yihan Chen","doi":"10.1016/j.eswa.2024.125834","DOIUrl":null,"url":null,"abstract":"<div><div>As awareness of environmental protection grows, the development of electric vehicles has emerged as a prominent area of research. Naturally, optimizing Permanent Magnet (PM) machines is critical for enhancing the power performance of electric vehicles, as they are essential components of these vehicles. Benefitting from the advantage of Deep Neural Network (DNN), optimizing the PM machine by leveraging DNN has drawn great attention from industry and academia. To increase the torque of the PM machine and reduce torque ripple from a structural design perspective, this paper proposes a <strong>M</strong>ulti-objective <strong>O</strong>ptimization approach for <strong>PM</strong> machine via improved <strong>S</strong>oft <strong>A</strong>ctor <strong>C</strong>ritic (called MOPM-SAC) based on Deep Reinforcement Learning (DRL). MOPM-SAC consists of two core components: a surrogate model based on DNN and a SAC algorithm. Specifically, the SAC algorithm based on DNN is used to optimize the PM machine. The surrogate model is established by training the DNN to use Finite Element (FE) simulation samples, which possesses higher accuracy than other methods, such as Support Vector Machine (SVM), Random Forest (RF), and so on. Moreover, the SAC algorithm integrates a trained DNN as the state transition function within the DRL environment, which enhances the optimization algorithm’s generalization ability in the context of PM machines. In addition, the reward function is designed based on optimization requirements to guide the agent in the SAC algorithm toward learning the optimal strategy. Finally, a 75 kW prototype is manufactured and tested. The effectiveness of the proposed method is validated through FE simulations and prototype experiments.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"264 ","pages":"Article 125834"},"PeriodicalIF":7.5000,"publicationDate":"2024-11-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multi-objective optimization approach for permanent magnet machine via improved soft actor–critic based on deep reinforcement learning\",\"authors\":\"Chen Wang , Tianyu Dong , Lei Chen , Guixiang Zhu , Yihan Chen\",\"doi\":\"10.1016/j.eswa.2024.125834\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>As awareness of environmental protection grows, the development of electric vehicles has emerged as a prominent area of research. Naturally, optimizing Permanent Magnet (PM) machines is critical for enhancing the power performance of electric vehicles, as they are essential components of these vehicles. Benefitting from the advantage of Deep Neural Network (DNN), optimizing the PM machine by leveraging DNN has drawn great attention from industry and academia. To increase the torque of the PM machine and reduce torque ripple from a structural design perspective, this paper proposes a <strong>M</strong>ulti-objective <strong>O</strong>ptimization approach for <strong>PM</strong> machine via improved <strong>S</strong>oft <strong>A</strong>ctor <strong>C</strong>ritic (called MOPM-SAC) based on Deep Reinforcement Learning (DRL). MOPM-SAC consists of two core components: a surrogate model based on DNN and a SAC algorithm. Specifically, the SAC algorithm based on DNN is used to optimize the PM machine. The surrogate model is established by training the DNN to use Finite Element (FE) simulation samples, which possesses higher accuracy than other methods, such as Support Vector Machine (SVM), Random Forest (RF), and so on. Moreover, the SAC algorithm integrates a trained DNN as the state transition function within the DRL environment, which enhances the optimization algorithm’s generalization ability in the context of PM machines. In addition, the reward function is designed based on optimization requirements to guide the agent in the SAC algorithm toward learning the optimal strategy. Finally, a 75 kW prototype is manufactured and tested. The effectiveness of the proposed method is validated through FE simulations and prototype experiments.</div></div>\",\"PeriodicalId\":50461,\"journal\":{\"name\":\"Expert Systems with Applications\",\"volume\":\"264 \",\"pages\":\"Article 125834\"},\"PeriodicalIF\":7.5000,\"publicationDate\":\"2024-11-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Expert Systems with Applications\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0957417424027015\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Expert Systems with Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0957417424027015","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Multi-objective optimization approach for permanent magnet machine via improved soft actor–critic based on deep reinforcement learning
As awareness of environmental protection grows, the development of electric vehicles has emerged as a prominent area of research. Naturally, optimizing Permanent Magnet (PM) machines is critical for enhancing the power performance of electric vehicles, as they are essential components of these vehicles. Benefitting from the advantage of Deep Neural Network (DNN), optimizing the PM machine by leveraging DNN has drawn great attention from industry and academia. To increase the torque of the PM machine and reduce torque ripple from a structural design perspective, this paper proposes a Multi-objective Optimization approach for PM machine via improved Soft Actor Critic (called MOPM-SAC) based on Deep Reinforcement Learning (DRL). MOPM-SAC consists of two core components: a surrogate model based on DNN and a SAC algorithm. Specifically, the SAC algorithm based on DNN is used to optimize the PM machine. The surrogate model is established by training the DNN to use Finite Element (FE) simulation samples, which possesses higher accuracy than other methods, such as Support Vector Machine (SVM), Random Forest (RF), and so on. Moreover, the SAC algorithm integrates a trained DNN as the state transition function within the DRL environment, which enhances the optimization algorithm’s generalization ability in the context of PM machines. In addition, the reward function is designed based on optimization requirements to guide the agent in the SAC algorithm toward learning the optimal strategy. Finally, a 75 kW prototype is manufactured and tested. The effectiveness of the proposed method is validated through FE simulations and prototype experiments.
期刊介绍:
Expert Systems With Applications is an international journal dedicated to the exchange of information on expert and intelligent systems used globally in industry, government, and universities. The journal emphasizes original papers covering the design, development, testing, implementation, and management of these systems, offering practical guidelines. It spans various sectors such as finance, engineering, marketing, law, project management, information management, medicine, and more. The journal also welcomes papers on multi-agent systems, knowledge management, neural networks, knowledge discovery, data mining, and other related areas, excluding applications to military/defense systems.