{"title":"pymops: A multi-agent simulation-based optimization package for power scheduling","authors":"Awol Seid Ebrie , Young Jin Kim","doi":"10.1016/j.simpa.2024.100616","DOIUrl":null,"url":null,"abstract":"<div><p>In response to the NP-hard power scheduling problem, the <span>pymops</span> package is developed as a robust solution. The package operates within a multi-agent simulation environment, where power-generating units are represented as reinforcement learning (RL) agents. The environment is designed to account for a comprehensive range of constraints. It also accommodates thermal valve point effects (VPEs) within cost and emissions functions. Moreover, in cases of constraint violations, the environment makes real-time contextual adjustments. Within the environment, the power scheduling problem is broken down into sequential Markov decision processes (MDPs), which serve as inputs for training a deep RL model aimed at solving the optimization problem.</p></div>","PeriodicalId":29771,"journal":{"name":"Software Impacts","volume":"19 ","pages":"Article 100616"},"PeriodicalIF":1.2000,"publicationDate":"2024-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2665963824000046/pdfft?md5=651e2ac4e426bc6fbbf5275437df2a6b&pid=1-s2.0-S2665963824000046-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Software Impacts","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2665963824000046","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
引用次数: 0
Abstract
In response to the NP-hard power scheduling problem, the pymops package is developed as a robust solution. The package operates within a multi-agent simulation environment, where power-generating units are represented as reinforcement learning (RL) agents. The environment is designed to account for a comprehensive range of constraints. It also accommodates thermal valve point effects (VPEs) within cost and emissions functions. Moreover, in cases of constraint violations, the environment makes real-time contextual adjustments. Within the environment, the power scheduling problem is broken down into sequential Markov decision processes (MDPs), which serve as inputs for training a deep RL model aimed at solving the optimization problem.