Pub Date : 2024-09-12DOI: 10.1016/j.compchemeng.2024.108877
Yun Zhu , Yucai Zhu , Chao Yang
The wake effect reduces the total power production of wind farms. This paper presents a method for wind farm power optimization through wake effect reduction. The proposed method optimizes the yaw angle offsets and de-rating settings of all turbines to maximize total power generation. The optimization approach is gradient-based, with gradients at each iteration obtained through system identification using field test data, eliminating the need for physical models. In system identification, test signal design, model estimation and model validation problems are solved in a systematic manner; in the gradient-based optimization, in order to achieve fast convergence, methods for initial value and initial step-size determination, variable step-size iteration and iteration termination are developed. The method is verified using the FLORIS wind farm model developed by National Renewable Energy Laboratory (NREL), USA. The studied wind farm consists of 80 wind turbines configured similarly to the Horns Rev I offshore wind farm in Denmark. The result of the developed optimization method is highly consistent with those obtained using FLORIS's built-in optimization tool.
{"title":"Wind farm power optimization using system identification","authors":"Yun Zhu , Yucai Zhu , Chao Yang","doi":"10.1016/j.compchemeng.2024.108877","DOIUrl":"10.1016/j.compchemeng.2024.108877","url":null,"abstract":"<div><p>The wake effect reduces the total power production of wind farms. This paper presents a method for wind farm power optimization through wake effect reduction. The proposed method optimizes the yaw angle offsets and de-rating settings of all turbines to maximize total power generation. The optimization approach is gradient-based, with gradients at each iteration obtained through system identification using field test data, eliminating the need for physical models. In system identification, test signal design, model estimation and model validation problems are solved in a systematic manner; in the gradient-based optimization, in order to achieve fast convergence, methods for initial value and initial step-size determination, variable step-size iteration and iteration termination are developed. The method is verified using the FLORIS wind farm model developed by National Renewable Energy Laboratory (NREL), USA. The studied wind farm consists of 80 wind turbines configured similarly to the Horns Rev I offshore wind farm in Denmark. The result of the developed optimization method is highly consistent with those obtained using FLORIS's built-in optimization tool.</p></div>","PeriodicalId":286,"journal":{"name":"Computers & Chemical Engineering","volume":"192 ","pages":"Article 108877"},"PeriodicalIF":3.9,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0098135424002953/pdfft?md5=b52657f2a6fdc592dabd38ad43caef73&pid=1-s2.0-S0098135424002953-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142241129","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-08DOI: 10.1016/j.compchemeng.2024.108863
Manar Oqbi , Luc Véchot , Dhabia M. Al-Mohannadi
Carbon capture, utilization, and storage supply chains (CCUS) play a pivotal role in achieving sustainability targets but necessitate meticulous risk identification and mitigation measures. Traditional safety assessments often occur post-design, constraining proactive risk management efforts. Hence, there is a pressing need to optimize safety performance during the design stages. To address this challenge, a framework for evaluating and optimizing CCUS supply chain safety performance using inherent safety index system (ISI) is introduced. Recognizing the trade-offs between total cost, environmental impact reduction, and risk mitigation, our approach considers multi-objective optimization to concurrently address these sustainability objectives and generate a Pareto set of solutions. Utilizing the augmented -constraint method, we applied this framework to optimize CCUS networks and develop sustainable designs across three key objectives. The method was applied to a CCUS system that includes various CO2 utilization pathways to minimize the total annual cost, CO2 emissions, and safety risks. The resulting Pareto surface illustrates unique network configurations, each representing a distinct trade-off scenario. Through a case study, we optimized a CCUS network to achieve economic, environmental, and safety objectives. The most economically viable design, with a total annual cost of $97 million and a 40 % net carbon reduction, prioritizes CO2 utilization for value-added products, while limiting CO2 sequestration. Conversely, safety-focused designs shift utilization towards safer routes, including CO2 sequestration and algae production. The proposed framework offers a systematic approach to developing sustainable CCUS supply chain designs, balancing economic viability, environmental sustainability, and safety.
{"title":"Safety-driven design of carbon capture utilization and storage (CCUS) supply chains: A multi-objective optimization approach","authors":"Manar Oqbi , Luc Véchot , Dhabia M. Al-Mohannadi","doi":"10.1016/j.compchemeng.2024.108863","DOIUrl":"10.1016/j.compchemeng.2024.108863","url":null,"abstract":"<div><p>Carbon capture, utilization, and storage supply chains (CCUS) play a pivotal role in achieving sustainability targets but necessitate meticulous risk identification and mitigation measures. Traditional safety assessments often occur post-design, constraining proactive risk management efforts. Hence, there is a pressing need to optimize safety performance during the design stages. To address this challenge, a framework for evaluating and optimizing CCUS supply chain safety performance using inherent safety index system (ISI) is introduced. Recognizing the trade-offs between total cost, environmental impact reduction, and risk mitigation, our approach considers multi-objective optimization to concurrently address these sustainability objectives and generate a Pareto set of solutions. Utilizing the augmented <span><math><mrow><mi>ε</mi></mrow></math></span>-constraint method, we applied this framework to optimize CCUS networks and develop sustainable designs across three key objectives. The method was applied to a CCUS system that includes various CO<sub>2</sub> utilization pathways to minimize the total annual cost, CO<sub>2</sub> emissions, and safety risks. The resulting Pareto surface illustrates unique network configurations, each representing a distinct trade-off scenario. Through a case study, we optimized a CCUS network to achieve economic, environmental, and safety objectives. The most economically viable design, with a total annual cost of $97 million and a 40 % net carbon reduction, prioritizes CO<sub>2</sub> utilization for value-added products, while limiting CO<sub>2</sub> sequestration. Conversely, safety-focused designs shift utilization towards safer routes, including CO<sub>2</sub> sequestration and algae production. The proposed framework offers a systematic approach to developing sustainable CCUS supply chain designs, balancing economic viability, environmental sustainability, and safety.</p></div>","PeriodicalId":286,"journal":{"name":"Computers & Chemical Engineering","volume":"192 ","pages":"Article 108863"},"PeriodicalIF":3.9,"publicationDate":"2024-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0098135424002813/pdfft?md5=bf08b9ddc902551052bee138fa804d6c&pid=1-s2.0-S0098135424002813-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142173260","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-07DOI: 10.1016/j.compchemeng.2024.108864
Haojun Zhong, Zhenlei Wang, Yuzhe Hao
The feeding process of the ethylene cracking furnace necessitates the synchronized adjustment of multiple controlled factors. The process mainly relies on operators to do it manually, which is burdensome and may lead to significant variations in coil out temperature (COT) due to the differing expertise of operators. This paper proposes a method for learning the feeding strategy of the ethylene cracking furnace using offline reinforcement learning. The agent learns and optimizes the operating strategy directly from datasets, eliminating the need for sophisticated process simulator modeling. In addition, the advantage function is incorporated into the Twin Delayed Deep Deterministic Behavioral Cloning (TD3BC) algorithm, which enables the agent to acquire more effective operational experience. The proposed method is initially evaluated using benchmark datasets. Further, the proposed method is validated through comparative experiments on a feeding process validation model, demonstrating superior rewards and outperforming manual operating experience as well as other offline reinforcement learning methods.
{"title":"Offline reinforcement learning based feeding strategy of ethylene cracking furnace","authors":"Haojun Zhong, Zhenlei Wang, Yuzhe Hao","doi":"10.1016/j.compchemeng.2024.108864","DOIUrl":"10.1016/j.compchemeng.2024.108864","url":null,"abstract":"<div><p>The feeding process of the ethylene cracking furnace necessitates the synchronized adjustment of multiple controlled factors. The process mainly relies on operators to do it manually, which is burdensome and may lead to significant variations in coil out temperature (COT) due to the differing expertise of operators. This paper proposes a method for learning the feeding strategy of the ethylene cracking furnace using offline reinforcement learning. The agent learns and optimizes the operating strategy directly from datasets, eliminating the need for sophisticated process simulator modeling. In addition, the advantage function is incorporated into the Twin Delayed Deep Deterministic Behavioral Cloning (TD3BC) algorithm, which enables the agent to acquire more effective operational experience. The proposed method is initially evaluated using benchmark datasets. Further, the proposed method is validated through comparative experiments on a feeding process validation model, demonstrating superior rewards and outperforming manual operating experience as well as other offline reinforcement learning methods.</p></div>","PeriodicalId":286,"journal":{"name":"Computers & Chemical Engineering","volume":"192 ","pages":"Article 108864"},"PeriodicalIF":3.9,"publicationDate":"2024-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0098135424002825/pdfft?md5=85eb0196a921c27b8bd237adcbca4c1d&pid=1-s2.0-S0098135424002825-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142274240","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-07DOI: 10.1016/j.compchemeng.2024.108872
Woohyun Jeong , Namjin Jang , Jay H. Lee
The Simulated Moving Bed (SMB) is a continuous chromatographic separation process that operates on the principle of counter-current movement between the solid and liquid phases. Due to periodic switching of feed and product ports across numerous connected columns, adjusting SMB operating variables such as feed and product flow rates and switching time to achieve desired separations is challenging. While equilibrium theory can help narrow the search space, obtaining essential information such as accurate adsorption isotherms is crucial. This requirement, combined with often highly stringent production specifications, makes it challenging to identify even a feasible operating condition, let alone an optimal one. Trial-and-error-based approaches are often impractical as reaching cyclic steady state can be time-consuming, and any waste produced during this period can lead to significant economic losses. While rigorous dynamic models are available, they are computationally intensive and often do not accurately mirror actual process behavior. To address these challenges, the use of Bayesian Optimization (BO) is proposed to sequentially approach optimal SMB operation. Furthermore, it is suggested to employ the simpler True Moving Bed (TMB) model as a prior for the BO, which significantly accelerates convergence. This approach is demonstrated on an SMB process for cresol separation. Initially, the effectiveness of the BO using the TMB model is examined to gain insights into its behavior. Subsequently, we apply BO to the rigorous SMB model, informed by prior knowledge from the TMB model. Our results show that the developed BO framework rapidly converges to the optimal operating parameters that satisfy the purity constraints. We examine the efficiency improvements over various search algorithms and highlight the advantages of using the TMB model as a prior.
模拟移动床(SMB)是一种连续色谱分离过程,其工作原理是固相和液相之间的逆流运动。由于进料口和产品口会在多个相连的色谱柱之间周期性切换,因此调整 SMB 的操作变量(如进料和产品流速以及切换时间)以实现理想的分离效果非常具有挑战性。虽然平衡理论可以帮助缩小搜索空间,但获得精确的吸附等温线等基本信息至关重要。这一要求加上通常非常严格的生产规格,使得确定可行的操作条件都具有挑战性,更不用说最佳条件了。基于试错的方法往往不切实际,因为达到周期性稳定状态需要耗费大量时间,而在此期间产生的任何废料都可能导致重大经济损失。虽然有严格的动态模型,但这些模型的计算量很大,而且往往不能准确反映实际的工艺行为。为了应对这些挑战,我们建议使用贝叶斯优化法(BO)来依次优化 SMB 的运行。此外,还建议采用更简单的真实移动床(TMB)模型作为贝叶斯优化的先验模型,这将大大加快收敛速度。这种方法在甲酚分离的 SMB 过程中得到了验证。首先,我们考察了使用 TMB 模型的 BO 的有效性,以深入了解其行为。随后,我们根据 TMB 模型的先验知识,将 BO 应用于严格的 SMB 模型。结果表明,所开发的 BO 框架能迅速收敛到满足纯度约束的最佳运行参数。我们检验了与各种搜索算法相比的效率改进,并强调了使用 TMB 模型作为先验知识的优势。
{"title":"Bayesian optimization for quick determination of operating variables of simulated moving bed chromatography","authors":"Woohyun Jeong , Namjin Jang , Jay H. Lee","doi":"10.1016/j.compchemeng.2024.108872","DOIUrl":"10.1016/j.compchemeng.2024.108872","url":null,"abstract":"<div><p>The Simulated Moving Bed (SMB) is a continuous chromatographic separation process that operates on the principle of counter-current movement between the solid and liquid phases. Due to periodic switching of feed and product ports across numerous connected columns, adjusting SMB operating variables such as feed and product flow rates and switching time to achieve desired separations is challenging. While equilibrium theory can help narrow the search space, obtaining essential information such as accurate adsorption isotherms is crucial. This requirement, combined with often highly stringent production specifications, makes it challenging to identify even a feasible operating condition, let alone an optimal one. Trial-and-error-based approaches are often impractical as reaching cyclic steady state can be time-consuming, and any waste produced during this period can lead to significant economic losses. While rigorous dynamic models are available, they are computationally intensive and often do not accurately mirror actual process behavior. To address these challenges, the use of Bayesian Optimization (BO) is proposed to sequentially approach optimal SMB operation. Furthermore, it is suggested to employ the simpler True Moving Bed (TMB) model as a prior for the BO, which significantly accelerates convergence. This approach is demonstrated on an SMB process for cresol separation. Initially, the effectiveness of the BO using the TMB model is examined to gain insights into its behavior. Subsequently, we apply BO to the rigorous SMB model, informed by prior knowledge from the TMB model. Our results show that the developed BO framework rapidly converges to the optimal operating parameters that satisfy the purity constraints. We examine the efficiency improvements over various search algorithms and highlight the advantages of using the TMB model as a prior.</p></div>","PeriodicalId":286,"journal":{"name":"Computers & Chemical Engineering","volume":"192 ","pages":"Article 108872"},"PeriodicalIF":3.9,"publicationDate":"2024-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0098135424002904/pdfft?md5=b35b8203b96305a9afe7ed2b2f470f89&pid=1-s2.0-S0098135424002904-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142241130","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The motility-induced phase separation (MIPS) phenomenon in active matter has been of great interest for the past decade or so. A central conceptual puzzle is that this behavior, which is generally characterized as a nonequilibrium phenomenon, can yet be explained using simple equilibrium models of thermodynamics. Here, we address this problem using a new theory, statistical teleodynamics, which is a conceptual synthesis of game theory and statistical mechanics. In this framework, active agents compete in their pursuit of maximum effective utility, and this self-organizing dynamics results in an arbitrage equilibrium in which all agents have the same effective utility. We show that MIPS is an example of arbitrage equilibrium and that it is mathematically equivalent to other phase-separation phenomena in entirely different domains, such as sociology and economics. As examples, we present the behavior of Janus particles in a potential trap and the effect of chemotaxis on MIPS.
{"title":"Arbitrage equilibria in active matter systems","authors":"Venkat Venkatasubramanian , Abhishek Sivaram , N. Sanjeevrajan , Arun Sankar","doi":"10.1016/j.compchemeng.2024.108861","DOIUrl":"10.1016/j.compchemeng.2024.108861","url":null,"abstract":"<div><p>The motility-induced phase separation (MIPS) phenomenon in active matter has been of great interest for the past decade or so. A central conceptual puzzle is that this behavior, which is generally characterized as a nonequilibrium phenomenon, can yet be explained using simple equilibrium models of thermodynamics. Here, we address this problem using a new theory, <em>statistical teleodynamics</em>, which is a conceptual synthesis of game theory and statistical mechanics. In this framework, active agents compete in their pursuit of <em>maximum effective utility</em>, and this self-organizing dynamics results in an <em>arbitrage equilibrium</em> in which all agents have the same effective utility. We show that MIPS is an example of arbitrage equilibrium and that it is mathematically equivalent to other phase-separation phenomena in entirely different domains, such as sociology and economics. As examples, we present the behavior of Janus particles in a potential trap and the effect of chemotaxis on MIPS.</p></div>","PeriodicalId":286,"journal":{"name":"Computers & Chemical Engineering","volume":"192 ","pages":"Article 108861"},"PeriodicalIF":3.9,"publicationDate":"2024-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0098135424002795/pdfft?md5=a3c048b78dbcc87219cb591648f944fa&pid=1-s2.0-S0098135424002795-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142164189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-07DOI: 10.1016/j.compchemeng.2024.108862
Georgia Ioanna Prokopou , Johannes M.M. Faust , Alexander Mitsos , Dominik Bongartz
Hydrogen refueling stations (HRS) can cause a significant fraction of the hydrogen refueling cost. The main cost contributor is the currently used mechanical compressor. Electrochemical hydrogen compression (EHC) has recently been proposed as an alternative. However, its optimal integration in an HRS has yet to be investigated. In this study, we compare the performance of a gaseous HRS equipped with different compressors. First, we develop dynamic models of three process configurations, which differ in the compressor technology: mechanical vs. electrochemical vs. combined. Then, the design and operation of the compressors are optimized by solving multi-stage dynamic optimization problems. The optimization results show that the three configurations lead to comparable hydrogen dispensing costs, because the electrochemical configuration exhibits lower capital cost but higher energy demand and thus operating cost than the mechanical configuration. The combined configuration is a trade-off with intermediate capital and operating cost.
{"title":"Cost-optimal design and operation of hydrogen refueling stations with mechanical and electrochemical hydrogen compressors","authors":"Georgia Ioanna Prokopou , Johannes M.M. Faust , Alexander Mitsos , Dominik Bongartz","doi":"10.1016/j.compchemeng.2024.108862","DOIUrl":"10.1016/j.compchemeng.2024.108862","url":null,"abstract":"<div><p>Hydrogen refueling stations (HRS) can cause a significant fraction of the hydrogen refueling cost. The main cost contributor is the currently used mechanical compressor. Electrochemical hydrogen compression (EHC) has recently been proposed as an alternative. However, its optimal integration in an HRS has yet to be investigated. In this study, we compare the performance of a gaseous HRS equipped with different compressors. First, we develop dynamic models of three process configurations, which differ in the compressor technology: mechanical vs. electrochemical vs. combined. Then, the design and operation of the compressors are optimized by solving multi-stage dynamic optimization problems. The optimization results show that the three configurations lead to comparable hydrogen dispensing costs, because the electrochemical configuration exhibits lower capital cost but higher energy demand and thus operating cost than the mechanical configuration. The combined configuration is a trade-off with intermediate capital and operating cost.</p></div>","PeriodicalId":286,"journal":{"name":"Computers & Chemical Engineering","volume":"192 ","pages":"Article 108862"},"PeriodicalIF":3.9,"publicationDate":"2024-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0098135424002801/pdfft?md5=f4eaaa70705aaf13f07966a01c9941e4&pid=1-s2.0-S0098135424002801-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142164190","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-06DOI: 10.1016/j.compchemeng.2024.108860
Wendi Zhang , Todd Przybycien , Jan Michael Breuer , Eric von Lieres
A particle size-based Smoluchowski coagulation and fragmentation equation was solved in the free and open source process modeling package CADET. The WFV and MCNP schemes were selected to discretize the internal particle size coordinate. Weights in these schemes were modified to preserve and conserve the zeroth and third moments for size-based equations. Modified propositions and proofs for the scheme are provided. Analytical Jacobians were derived and implemented to reduce the solver’s runtime. A two-dimensional Smoluchowski coagulation and fragmentation equation with axial position as external coordinate was formulated and discretized to support simulations of continuous particulate processes in dispersive plug flow reactors. Five 1D and four 2D test cases were used to validate the implementation and benchmark the solver’s performance. The runtime, L1 error norm, L1 error rate, particle size distribution moments up to sixth order and several scalar metrics were analyzed in detail.
{"title":"Solving crystallization/precipitation population balance models in CADET, Part II: Size-based Smoluchowski coagulation and fragmentation equations in batch and continuous modes","authors":"Wendi Zhang , Todd Przybycien , Jan Michael Breuer , Eric von Lieres","doi":"10.1016/j.compchemeng.2024.108860","DOIUrl":"10.1016/j.compchemeng.2024.108860","url":null,"abstract":"<div><p>A particle size-based Smoluchowski coagulation and fragmentation equation was solved in the free and open source process modeling package CADET. The WFV and MCNP schemes were selected to discretize the internal particle size coordinate. Weights in these schemes were modified to preserve and conserve the zeroth and third moments for size-based equations. Modified propositions and proofs for the scheme are provided. Analytical Jacobians were derived and implemented to reduce the solver’s runtime. A two-dimensional Smoluchowski coagulation and fragmentation equation with axial position as external coordinate was formulated and discretized to support simulations of continuous particulate processes in dispersive plug flow reactors. Five 1D and four 2D test cases were used to validate the implementation and benchmark the solver’s performance. The runtime, L1 error norm, L1 error rate, particle size distribution moments up to sixth order and several scalar metrics were analyzed in detail.</p></div>","PeriodicalId":286,"journal":{"name":"Computers & Chemical Engineering","volume":"192 ","pages":"Article 108860"},"PeriodicalIF":3.9,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0098135424002783/pdfft?md5=fad1470a25e5723a35f04c2126b9b1ad&pid=1-s2.0-S0098135424002783-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142169467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-03DOI: 10.1016/j.compchemeng.2024.108857
Paris Papavasileiou , Dimitrios G. Giovanis , Gabriele Pozzetti , Martin Kathrein , Christoph Czettl , Ioannis G. Kevrekidis , Andreas G. Boudouvis , Stéphane P.A. Bordas , Eleni D. Koronaki
This study introduces a machine learning framework tailored to large-scale industrial processes characterized by a plethora of numerical and categorical inputs. The framework aims to (i) discern critical parameters that influence the output and (ii) generate accurate out-of-sample qualitative and quantitative predictions of production outcomes. Specifically, we address the pivotal question of the significance of each input in shaping the process outcome, using an industrial Chemical Vapor Deposition (CVD) process as an example. The initial objective involves merging subject matter expertise and clustering techniques exclusively on the process output, here, coating thickness measurements at various positions in the reactor. This approach identifies groups of production runs that share similar qualitative characteristics, such as film mean thickness and standard deviation. In particular, the differences of the outcomes represented by the different clusters can be attributed to differences in specific inputs, indicating that these inputs are potentially critical to the production outcome. Shapley value analysis corroborates the formed hypotheses. Leveraging this insight, we subsequently implement supervised classification and regression methods using the identified critical process inputs. The proposed methodology proves to be valuable in scenarios with a multitude of inputs and insufficient data for the direct application of deep learning techniques, providing meaningful insights into the underlying processes.
{"title":"Integrating supervised and unsupervised learning approaches to unveil critical process inputs","authors":"Paris Papavasileiou , Dimitrios G. Giovanis , Gabriele Pozzetti , Martin Kathrein , Christoph Czettl , Ioannis G. Kevrekidis , Andreas G. Boudouvis , Stéphane P.A. Bordas , Eleni D. Koronaki","doi":"10.1016/j.compchemeng.2024.108857","DOIUrl":"10.1016/j.compchemeng.2024.108857","url":null,"abstract":"<div><p>This study introduces a machine learning framework tailored to large-scale industrial processes characterized by a plethora of numerical and categorical inputs. The framework aims to (i) discern critical parameters that influence the output and (ii) generate accurate out-of-sample qualitative and quantitative predictions of production outcomes. Specifically, we address the pivotal question of the significance of each input in shaping the process outcome, using an industrial Chemical Vapor Deposition (CVD) process as an example. The initial objective involves merging subject matter expertise and clustering techniques exclusively on the process output, here, coating thickness measurements at various positions in the reactor. This approach identifies groups of production runs that share similar qualitative characteristics, such as film mean thickness and standard deviation. In particular, the differences of the outcomes represented by the different clusters can be attributed to differences in specific inputs, indicating that these inputs are potentially critical to the production outcome. Shapley value analysis corroborates the formed hypotheses. Leveraging this insight, we subsequently implement supervised classification and regression methods using the identified critical process inputs. The proposed methodology proves to be valuable in scenarios with a multitude of inputs and insufficient data for the direct application of deep learning techniques, providing meaningful insights into the underlying processes.</p></div>","PeriodicalId":286,"journal":{"name":"Computers & Chemical Engineering","volume":"192 ","pages":"Article 108857"},"PeriodicalIF":3.9,"publicationDate":"2024-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0098135424002758/pdfft?md5=5b97cf2052fa37b1ae7b0760796c20ca&pid=1-s2.0-S0098135424002758-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142158021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-02DOI: 10.1016/j.compchemeng.2024.108859
Yen-An Lu , Wei-Shou Hu , Joel A. Paulson , Qi Zhang
Data-driven inverse optimization (IO) aims to estimate unknown parameters in an optimization model from observed decisions. The IO problem is commonly formulated as a large-scale bilevel program that is notoriously difficult to solve. We propose a derivative-free optimization approach based on Bayesian optimization, BO4IO, to solve general IO problems. The main advantages of BO4IO are two-fold: (i) it circumvents the need of complex reformulations or specialized algorithms and can hence enable computational tractability even when the underlying optimization problem is nonconvex or involves discrete variables, and (ii) it allows approximations of the profile likelihood, which provide uncertainty quantification on the IO parameter estimates. Our extensive computational results demonstrate the efficacy and robustness of BO4IO to estimate unknown parameters from small and noisy datasets. In addition, the proposed profile likelihood analysis effectively provides good approximations of the confidence intervals on the parameter estimates and assesses the identifiability of the unknown parameters.
{"title":"BO4IO: A Bayesian optimization approach to inverse optimization with uncertainty quantification","authors":"Yen-An Lu , Wei-Shou Hu , Joel A. Paulson , Qi Zhang","doi":"10.1016/j.compchemeng.2024.108859","DOIUrl":"10.1016/j.compchemeng.2024.108859","url":null,"abstract":"<div><p>Data-driven inverse optimization (IO) aims to estimate unknown parameters in an optimization model from observed decisions. The IO problem is commonly formulated as a large-scale bilevel program that is notoriously difficult to solve. We propose a derivative-free optimization approach based on Bayesian optimization, BO4IO, to solve general IO problems. The main advantages of BO4IO are two-fold: (i) it circumvents the need of complex reformulations or specialized algorithms and can hence enable computational tractability even when the underlying optimization problem is nonconvex or involves discrete variables, and (ii) it allows approximations of the profile likelihood, which provide uncertainty quantification on the IO parameter estimates. Our extensive computational results demonstrate the efficacy and robustness of BO4IO to estimate unknown parameters from small and noisy datasets. In addition, the proposed profile likelihood analysis effectively provides good approximations of the confidence intervals on the parameter estimates and assesses the identifiability of the unknown parameters.</p></div>","PeriodicalId":286,"journal":{"name":"Computers & Chemical Engineering","volume":"192 ","pages":"Article 108859"},"PeriodicalIF":3.9,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0098135424002771/pdfft?md5=2b1fcb630ba3141652b16ea5b79fc168&pid=1-s2.0-S0098135424002771-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142158020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-31DOI: 10.1016/j.compchemeng.2024.108858
Mrunal Sontakke , Lucky E. Yerimah , Andreas Rebmann , Sambit Ghosh , Craig Dory , Ronald Hedden , B. Wayne Bequette
The process systems domain is undergoing the fourth industrial revolution, which is helping industries digitize and optimize their production techniques. Concurrently, the field of data-based modeling has been expanding, leading to the proposal of many fault detection models. However, the rapid expansion has created gaps in the field. For instance, Smart Manufacturing (SM) methodologies have yet to be incorporated into undergraduate chemical engineering education. Additionally, only a few developed fault detection models have been deployed for real-time usage and practical applications. This study takes a crucial step toward bridging the two mentioned gaps by enabling undergraduate students to learn SM techniques and developing a safe and controlled academic environment for deploying fault detection models. The demonstration is implemented on a shell and tube heat exchanger, taught in a senior year laboratory course, using the Smart Manufacturing Innovation Platform (SMIP). The implementation provides an easily customizable pipeline for SM applications involving human-in-the-loop decision-making on a real-life hardware system. Actual data from heat exchanger equipment is used to train and compare the performances of several state-of-the-art fault detection models, including fully connected, convolutional, and recurrent neural networks. Current work also presents tutorials on deploying models for practical real-time applications using the SMIP. The overall architecture is a plug-and-play package that will motivate students to learn about SM and catalyze their interest in developing and deploying fault detection models using real-world data.
过程系统领域正在经历第四次工业革命,这有助于各行业实现生产技术的数字化和优化。与此同时,基于数据的建模领域也在不断扩大,从而提出了许多故障检测模型。然而,快速扩张也造成了该领域的空白。例如,智能制造 (SM) 方法尚未纳入化学工程本科教育。此外,只有少数已开发的故障检测模型被部署到实时使用和实际应用中。本研究通过让本科生学习 SM 技术,并为部署故障检测模型开发安全可控的学术环境,为弥补上述两个差距迈出了关键一步。该演示是在高年级实验课程中使用智能制造创新平台(SMIP)在管壳式热交换器上实施的。该实施方案为智能制造应用提供了一个易于定制的管道,涉及现实生活中硬件系统上的人在环决策。来自热交换器设备的实际数据被用来训练和比较几种最先进的故障检测模型的性能,包括全连接、卷积和递归神经网络。当前工作还介绍了使用 SMIP 为实际实时应用部署模型的教程。整体架构是一个即插即用的软件包,可激发学生学习 SM 的兴趣,并促进他们利用真实世界的数据开发和部署故障检测模型。
{"title":"Integrating smart manufacturing techniques into undergraduate education: A case study with heat exchanger","authors":"Mrunal Sontakke , Lucky E. Yerimah , Andreas Rebmann , Sambit Ghosh , Craig Dory , Ronald Hedden , B. Wayne Bequette","doi":"10.1016/j.compchemeng.2024.108858","DOIUrl":"10.1016/j.compchemeng.2024.108858","url":null,"abstract":"<div><p>The process systems domain is undergoing the fourth industrial revolution, which is helping industries digitize and optimize their production techniques. Concurrently, the field of data-based modeling has been expanding, leading to the proposal of many fault detection models. However, the rapid expansion has created gaps in the field. For instance, Smart Manufacturing (SM) methodologies have yet to be incorporated into undergraduate chemical engineering education. Additionally, only a few developed fault detection models have been deployed for real-time usage and practical applications. This study takes a crucial step toward bridging the two mentioned gaps by enabling undergraduate students to learn SM techniques and developing a safe and controlled academic environment for deploying fault detection models. The demonstration is implemented on a shell and tube heat exchanger, taught in a senior year laboratory course, using the Smart Manufacturing Innovation Platform (SMIP). The implementation provides an easily customizable pipeline for SM applications involving human-in-the-loop decision-making on a real-life hardware system. Actual data from heat exchanger equipment is used to train and compare the performances of several state-of-the-art fault detection models, including fully connected, convolutional, and recurrent neural networks. Current work also presents tutorials on deploying models for practical real-time applications using the SMIP. The overall architecture is a plug-and-play package that will motivate students to learn about SM and catalyze their interest in developing and deploying fault detection models using real-world data.</p></div>","PeriodicalId":286,"journal":{"name":"Computers & Chemical Engineering","volume":"191 ","pages":"Article 108858"},"PeriodicalIF":3.9,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142148362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}