Ramez Abdalla , Nermine Agban , Christian Lüddeke , Dan Sui , Philip Jaeger
{"title":"Multi agent physics informed reinforcement learning for waterflooding optimization","authors":"Ramez Abdalla , Nermine Agban , Christian Lüddeke , Dan Sui , Philip Jaeger","doi":"10.1016/j.fraope.2025.100229","DOIUrl":null,"url":null,"abstract":"<div><div>Waterflooding optimization is a critical process for enhancing oil recovery in mature oil fields, where conventional approaches often rely on fixed injection rates over an extended period. However, this may not be the most efficient strategy due to reservoir heterogeneity and complexity. In this study, we propose a multi-agent physics informed reinforcement learning (MAPIRL) framework to optimize the waterflooding process. The MAPIRL approach utilizes a Markov decision process to formulate the optimization problem, where multiple RL agents are trained to interact with a reservoir simulation model and receive rewards for each action. The proposed approach uses an actor–critic RL architecture to train the agents to find the optimal strategy. The agents interact with the environment during several episodes until convergence is achieved. We evaluated the effectiveness of the MAPIRL approach based on the improvement in net present value (NPV), which reflects the economic benefits of the optimized waterflooding strategy. Then, we compared the MAPIRL approach with the multi-objective particle swarm optimization (MOPSO) algorithm. The comparison revealed that the MAPIRL approach outperformed the MOPSO algorithm in terms of net present value. In conclusion, the MAPIRL approach is a scientifically accurate method for optimizing waterflooding in mature oil fields, providing a more efficient and robust waterflooding strategy that reduces water consumption and associated costs while maximizing the economic benefits. The ability of the MAPIRL approach to optimize the waterflooding process with a high degree of complexity makes it a promising tool for the energy industry, and further research is needed to explore its potential for addressing other complex problems in this domain.</div></div>","PeriodicalId":100554,"journal":{"name":"Franklin Open","volume":"10 ","pages":"Article 100229"},"PeriodicalIF":0.0000,"publicationDate":"2025-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Franklin Open","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2773186325000192","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Waterflooding optimization is a critical process for enhancing oil recovery in mature oil fields, where conventional approaches often rely on fixed injection rates over an extended period. However, this may not be the most efficient strategy due to reservoir heterogeneity and complexity. In this study, we propose a multi-agent physics informed reinforcement learning (MAPIRL) framework to optimize the waterflooding process. The MAPIRL approach utilizes a Markov decision process to formulate the optimization problem, where multiple RL agents are trained to interact with a reservoir simulation model and receive rewards for each action. The proposed approach uses an actor–critic RL architecture to train the agents to find the optimal strategy. The agents interact with the environment during several episodes until convergence is achieved. We evaluated the effectiveness of the MAPIRL approach based on the improvement in net present value (NPV), which reflects the economic benefits of the optimized waterflooding strategy. Then, we compared the MAPIRL approach with the multi-objective particle swarm optimization (MOPSO) algorithm. The comparison revealed that the MAPIRL approach outperformed the MOPSO algorithm in terms of net present value. In conclusion, the MAPIRL approach is a scientifically accurate method for optimizing waterflooding in mature oil fields, providing a more efficient and robust waterflooding strategy that reduces water consumption and associated costs while maximizing the economic benefits. The ability of the MAPIRL approach to optimize the waterflooding process with a high degree of complexity makes it a promising tool for the energy industry, and further research is needed to explore its potential for addressing other complex problems in this domain.