Pub Date : 2025-06-01Epub Date: 2025-05-26DOI: 10.1016/j.dche.2025.100242
Feiyang Ou , Julius Suherman , Chao Zhang , Henrik Wang , Sthitie Bom , James F. Davis , Panagiotis D. Christofides
<div><div>Smart Manufacturing (SM), which is short for “Smart (Predictive, Preventive, Proactive) zero incident, zero emissions Manufacturing,” describes manufacturing’s digital transformation in which factories, supply chains and ecosystems are integrated, interoperable, and interconnected. Smart Manufacturing is rooted in AI, Machine Learned (ML), and Data Synchronized (DS) modeling to tap into invaluable operating data. By making data actionable at larger scales, SM opens new ways to increase productivity, precision, and process performance. Smart Manufacturing applied to front-end wafer manufacturing in the semiconductor industry offers significant opportunity to increase production throughput and ensure precision by increasing staff and operational productivity. Front-end wafer manufacturing involves multi tool operations for complex material processing that requires a high degree of precision and extensive product qualification. There is a high degree of commonality with semiconductor manufacturing tools, for example etching, that are well instrumented. Companies are already collecting large amounts of operational data from these tools that can be aggregated and leveraged for virtual metrology and other control, diagnostic, and management solutions. AI/ML/DS modeling involves monitoring the state of an operation in real-time to continuously learn and improve on human centered, automated, and autonomous actions. This operational data are embedded in invaluable machine, process, product, and material behaviors as interaction complexities, linearities/non-linearities, and dimensional effects. Because of machine commonalities, data can be selected to draw out operational value across machines. Today’s data science offers considerable capability for qualifying, assessing alignment and contribution, aggregating, and engineering data for more robust modeling. We refer to this as a Data-first strategy to process, engineer and model with AI-Ready data. In this paper, we address AI-Ready data for a virtual metrology solution focused on etching measurement PASS/FAIL classification and milling depth prediction regression tasks using operational data from production machine tools. If the quality of the product can be predicted, the productivity of the metrology process can be increased, which in turn increases the productivity of the overall operation. In a previous paper, we considered how to aggregate data from different etch tools in the same processes at different factories within Seagate Technology and proposed a method for data aggregation and demonstrated its value (Ou et al., 2024). The present paper considers how to process and engineer datasets from two different etch tool processes: wafer and slider production. The data processing approaches when used systematically with appropriate ML algorithms demonstrate the potential for reducing metrological interventions in semiconductor manufacturing. Advanced machine learning techniques are used to t
智能制造(SM)是“智能(预测性、预防性、前瞻性)零事故、零排放制造”的缩写,它描述了制造业的数字化转型,在这种转型中,工厂、供应链和生态系统是集成的、可互操作的、互联的。智能制造植根于人工智能、机器学习(ML)和数据同步(DS)建模,以挖掘宝贵的运营数据。通过使数据在更大范围内可操作,SM开辟了提高生产率、精度和流程性能的新途径。智能制造应用于半导体行业的前端晶圆制造,通过提高员工和运营生产率,为提高生产吞吐量和确保精度提供了重要机会。前端晶圆制造涉及复杂材料加工的多工具操作,需要高度精度和广泛的产品认证。与半导体制造工具有高度的共性,例如蚀刻,这是良好的仪器。公司已经从这些工具中收集了大量的操作数据,这些数据可以用于虚拟计量和其他控制、诊断和管理解决方案。AI/ML/DS建模涉及实时监控操作状态,以不断学习和改进以人为中心的自动化和自主操作。这些操作数据嵌入在宝贵的机器、过程、产品和材料行为中,如交互复杂性、线性/非线性和维度效应。由于机器的共性,可以选择数据来跨机器提取操作值。今天的数据科学提供了相当大的能力来鉴定、评估对齐和贡献、聚合和工程数据,以实现更健壮的建模。我们将其称为数据优先策略,以处理、设计和建模ai就绪数据。在本文中,我们解决了虚拟计量解决方案的AI-Ready数据,该解决方案专注于蚀刻测量PASS/FAIL分类和铣削深度预测回归任务,使用生产机床的操作数据。如果可以预测产品的质量,则可以提高计量过程的生产率,从而提高整体操作的生产率。在之前的一篇论文中,我们考虑了如何在希捷科技内部不同工厂的相同流程中聚合来自不同蚀刻工具的数据,并提出了一种数据聚合方法并展示了其价值(Ou et al., 2024)。本文考虑了如何处理和设计来自两种不同蚀刻工具过程的数据集:晶圆和滑块生产。当系统地使用适当的ML算法时,数据处理方法显示了减少半导体制造计量干预的潜力。先进的机器学习技术用于解决低故障率和有限操作可变性的建模挑战。XGBoost是一种基于梯度下降的树算法,在二元分类的训练速度和资源利用率方面优于常用的前馈神经网络(FNN),在ROC-AUC分数(分类)、绝对误差中位数(回归)和R2值方面的性能标准也优于前者。主成分分析(PCA)有效地降低了数据的维数和过拟合,同时保留了重要方差并显著降低了噪声。具有分离缩放的数据聚合协调了来自不同制造工具的输入,并显着提高了组合多个数据集的效率和通用性,从而提高了模型性能。一种实时更新迁移学习方法,该方法使用随机梯度下降(SGD)和单个数据点定期实时更新FNN模型,解决了过程漂移,并显着提高了预测精度。对于滑块生产工具,使用线性Mixup进行数据增强,克服了较短的记录周期,丰富了训练数据集,并显着降低了误差指标。
{"title":"Industrial multi-machine data aggregation, AI-ready data preparation, and machine learning for virtual metrology in semiconductor wafer and slider production","authors":"Feiyang Ou , Julius Suherman , Chao Zhang , Henrik Wang , Sthitie Bom , James F. Davis , Panagiotis D. Christofides","doi":"10.1016/j.dche.2025.100242","DOIUrl":"10.1016/j.dche.2025.100242","url":null,"abstract":"<div><div>Smart Manufacturing (SM), which is short for “Smart (Predictive, Preventive, Proactive) zero incident, zero emissions Manufacturing,” describes manufacturing’s digital transformation in which factories, supply chains and ecosystems are integrated, interoperable, and interconnected. Smart Manufacturing is rooted in AI, Machine Learned (ML), and Data Synchronized (DS) modeling to tap into invaluable operating data. By making data actionable at larger scales, SM opens new ways to increase productivity, precision, and process performance. Smart Manufacturing applied to front-end wafer manufacturing in the semiconductor industry offers significant opportunity to increase production throughput and ensure precision by increasing staff and operational productivity. Front-end wafer manufacturing involves multi tool operations for complex material processing that requires a high degree of precision and extensive product qualification. There is a high degree of commonality with semiconductor manufacturing tools, for example etching, that are well instrumented. Companies are already collecting large amounts of operational data from these tools that can be aggregated and leveraged for virtual metrology and other control, diagnostic, and management solutions. AI/ML/DS modeling involves monitoring the state of an operation in real-time to continuously learn and improve on human centered, automated, and autonomous actions. This operational data are embedded in invaluable machine, process, product, and material behaviors as interaction complexities, linearities/non-linearities, and dimensional effects. Because of machine commonalities, data can be selected to draw out operational value across machines. Today’s data science offers considerable capability for qualifying, assessing alignment and contribution, aggregating, and engineering data for more robust modeling. We refer to this as a Data-first strategy to process, engineer and model with AI-Ready data. In this paper, we address AI-Ready data for a virtual metrology solution focused on etching measurement PASS/FAIL classification and milling depth prediction regression tasks using operational data from production machine tools. If the quality of the product can be predicted, the productivity of the metrology process can be increased, which in turn increases the productivity of the overall operation. In a previous paper, we considered how to aggregate data from different etch tools in the same processes at different factories within Seagate Technology and proposed a method for data aggregation and demonstrated its value (Ou et al., 2024). The present paper considers how to process and engineer datasets from two different etch tool processes: wafer and slider production. The data processing approaches when used systematically with appropriate ML algorithms demonstrate the potential for reducing metrological interventions in semiconductor manufacturing. Advanced machine learning techniques are used to t","PeriodicalId":72815,"journal":{"name":"Digital Chemical Engineering","volume":"15 ","pages":"Article 100242"},"PeriodicalIF":3.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144147463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-06-01Epub Date: 2025-02-19DOI: 10.1016/j.dche.2025.100223
Asma Iqbal, Mohammad Amil Bhat, Qazi Muneeb, Muazam Javid
The Perfumery Ternary Diagram (PTD) is a powerful tool in perfumery for analyzing perfume mixtures comprising three fragrant components and a solvent base. It combines ternary diagrams with perfume pyramids to swiftly evaluate odor characteristics and composition in the headspace across various concentrations, bypassing time-consuming experimental processes. Using a diffusion model to simulate evaporation, this study utilizes PTDs to track changes in the liquid and gas-liquid interface. Using Python, we calculated the OVs of each component at 25 °C, based on molecular weight, saturated vapor pressure, and odor threshold. The data was processed and visualized in MATLAB, producing PTDs that highlighted the component with the highest OV at any given composition. Furthermore, initially as the mole fraction continues to rise, the percentage decrease in odor value is approximately 11.1 %, indicating a diminishing rate of change. The distribution of odor values is elaborated in the MATLAB diagrams that give a comprehensive representation of how the odor value varies with different compositions. The PTDs were effective in representing the critical role of individual components, making them valuable tools for perfumers and researchers. The PTD analysis revealed that limonene (top note) demonstrated the highest odor value (OV) at concentrations above 60 % within the mixture, while vanillin (base note) maintained stability at lower concentrations, supporting its role as a fixative. These findings validate PTDs as predictive tools, accurately reflecting odor value variations across different fragrance compositions. This study investigates whether Perfumery Ternary Diagrams (PTDs) can reliably predict odor value distributions within perfume mixtures, thus providing a practical and efficient tool for optimizing fragrance compositions.
{"title":"Revolutionizing perfume creation: PTD's innovative approach","authors":"Asma Iqbal, Mohammad Amil Bhat, Qazi Muneeb, Muazam Javid","doi":"10.1016/j.dche.2025.100223","DOIUrl":"10.1016/j.dche.2025.100223","url":null,"abstract":"<div><div>The Perfumery Ternary Diagram (PTD) is a powerful tool in perfumery for analyzing perfume mixtures comprising three fragrant components and a solvent base. It combines ternary diagrams with perfume pyramids to swiftly evaluate odor characteristics and composition in the headspace across various concentrations, bypassing time-consuming experimental processes. Using a diffusion model to simulate evaporation, this study utilizes PTDs to track changes in the liquid and gas-liquid interface. Using Python, we calculated the OVs of each component at 25 °C, based on molecular weight, saturated vapor pressure, and odor threshold. The data was processed and visualized in MATLAB, producing PTDs that highlighted the component with the highest OV at any given composition. Furthermore, initially as the mole fraction continues to rise, the percentage decrease in odor value is approximately 11.1 %, indicating a diminishing rate of change. The distribution of odor values is elaborated in the MATLAB diagrams that give a comprehensive representation of how the odor value varies with different compositions. The PTDs were effective in representing the critical role of individual components, making them valuable tools for perfumers and researchers. The PTD analysis revealed that limonene (top note) demonstrated the highest odor value (OV) at concentrations above 60 % within the mixture, while vanillin (base note) maintained stability at lower concentrations, supporting its role as a fixative. These findings validate PTDs as predictive tools, accurately reflecting odor value variations across different fragrance compositions. This study investigates whether Perfumery Ternary Diagrams (PTDs) can reliably predict odor value distributions within perfume mixtures, thus providing a practical and efficient tool for optimizing fragrance compositions.</div></div>","PeriodicalId":72815,"journal":{"name":"Digital Chemical Engineering","volume":"15 ","pages":"Article 100223"},"PeriodicalIF":3.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143510920","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-06-01Epub Date: 2025-03-10DOI: 10.1016/j.dche.2025.100230
Donovan Chaffart , Yue Yuan
The demand for reliable Artificial Intelligence (AI) models within critical domains such as Chemical Engineering has garnered significant attention towards the use and development of transparent AI methodologies. Nevertheless, the field of AI transparency has received an uneven level of attention, such that crucial aspects like explainability (i.e., the transparency of the AI's operational rationales) have remained understudied. To address this challenge, this study investigates the inherent explainability capabilities of Polynomial Neural Networks (PNNs) for applications within Chemical Engineering. PNNs, which implement higher-order polynomials in lieu of linear expressions within their hidden layer neurons, are inherently nonlinear, and thus do not require an activation function to accurately capture the behavior of a system. Accordingly, these neural networks provide continuous, closed-form algebraic expressions that can be used to ascertain the contributions of individual features in the AI architecture towards the network operational behavior. In order to study this behavior, the PNN method was adopted in this work to capture the relationships of noiseless and noisy data derived according to simple mathematical expressions. The PNN polynomials were then extracted and examined to highlight the insights they provide regarding the system operational rationales. The PNN method was furthermore applied to capture the behavior of a circulating fluidized bed reactor to fully showcase the explainative capability of this method within a Chemical Engineering application. These studies highlight the intrinsic explainability capabilities of PNNs and demonstrated their potential for reliable AI implementations for applications in Chemical Engineering.
{"title":"Polynomial Neural Networks for improved AI transparency: An analysis of their inherent explainability (operational rationale) capabilities","authors":"Donovan Chaffart , Yue Yuan","doi":"10.1016/j.dche.2025.100230","DOIUrl":"10.1016/j.dche.2025.100230","url":null,"abstract":"<div><div>The demand for reliable Artificial Intelligence (AI) models within critical domains such as Chemical Engineering has garnered significant attention towards the use and development of transparent AI methodologies. Nevertheless, the field of AI transparency has received an uneven level of attention, such that crucial aspects like <em>explainability</em> (i.e., the transparency of the AI's operational rationales) have remained understudied. To address this challenge, this study investigates the inherent <em>explainability</em> capabilities of Polynomial Neural Networks (PNNs) for applications within Chemical Engineering. PNNs, which implement higher-order polynomials in lieu of linear expressions within their hidden layer neurons, are inherently nonlinear, and thus do not require an activation function to accurately capture the behavior of a system. Accordingly, these neural networks provide continuous, closed-form algebraic expressions that can be used to ascertain the contributions of individual features in the AI architecture towards the network operational behavior. In order to study this behavior, the PNN method was adopted in this work to capture the relationships of noiseless and noisy data derived according to simple mathematical expressions. The PNN polynomials were then extracted and examined to highlight the insights they provide regarding the system operational rationales. The PNN method was furthermore applied to capture the behavior of a circulating fluidized bed reactor to fully showcase the <em>explainative</em> capability of this method within a Chemical Engineering application. These studies highlight the intrinsic <em>explainability</em> capabilities of PNNs and demonstrated their potential for reliable AI implementations for applications in Chemical Engineering.</div></div>","PeriodicalId":72815,"journal":{"name":"Digital Chemical Engineering","volume":"15 ","pages":"Article 100230"},"PeriodicalIF":3.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143684687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-06-01Epub Date: 2025-04-01DOI: 10.1016/j.dche.2025.100233
Arthur Khodaverdian , Dhruv Gohil , Panagiotis D. Christofides
This work proposes a novel two-layer multi-key control architecture to enhance the resilience of nonlinear chemical processes to cyberattacks. The architecture consists of an upper-layer nonlinear controller and a lower-layer of encrypted linear controllers. The nonlinear controllers process unencrypted sensor data to determine optimal control actions, which are then used to estimate the closed-loop state trajectory using a first-principle model of the plant. This trajectory is sampled and mapped to a valid subset before encryption, which can lead to minor inaccuracies. The resulting encrypted state-space data samples are used as set-points for the lower-layer controllers, which can be implemented using encrypted signals, allowing for obfuscation of the computation and transmission of the applied control inputs, thereby enhancing cybersecurity. This study further improves security by taking advantage of the Single-Input-Single-Output nature of some linear control methods to allocate a unique encryption key to each linear controller and its respective sensor data. Two nonlinear chemical process applications, including a benchmark chemical reactor example and one application modeled through the use of Aspen Dynamics, are used to demonstrate the application of the proposed two-layer architecture.
{"title":"Enhancing cybersecurity of nonlinear processes via a two-layer control architecture","authors":"Arthur Khodaverdian , Dhruv Gohil , Panagiotis D. Christofides","doi":"10.1016/j.dche.2025.100233","DOIUrl":"10.1016/j.dche.2025.100233","url":null,"abstract":"<div><div>This work proposes a novel two-layer multi-key control architecture to enhance the resilience of nonlinear chemical processes to cyberattacks. The architecture consists of an upper-layer nonlinear controller and a lower-layer of encrypted linear controllers. The nonlinear controllers process unencrypted sensor data to determine optimal control actions, which are then used to estimate the closed-loop state trajectory using a first-principle model of the plant. This trajectory is sampled and mapped to a valid subset before encryption, which can lead to minor inaccuracies. The resulting encrypted state-space data samples are used as set-points for the lower-layer controllers, which can be implemented using encrypted signals, allowing for obfuscation of the computation and transmission of the applied control inputs, thereby enhancing cybersecurity. This study further improves security by taking advantage of the Single-Input-Single-Output nature of some linear control methods to allocate a unique encryption key to each linear controller and its respective sensor data. Two nonlinear chemical process applications, including a benchmark chemical reactor example and one application modeled through the use of Aspen Dynamics, are used to demonstrate the application of the proposed two-layer architecture.</div></div>","PeriodicalId":72815,"journal":{"name":"Digital Chemical Engineering","volume":"15 ","pages":"Article 100233"},"PeriodicalIF":3.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143759018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-06-01Epub Date: 2025-03-12DOI: 10.1016/j.dche.2025.100227
Austin Braniff , Sahithi Srijana Akundi , Yuanxing Liu , Beatriz Dantas , Shayan S. Niknezhad , Faisal Khan , Efstratios N. Pistikopoulos , Yuhe Tian
The ongoing digital transformation has created new opportunities for chemical manufacturing with increasing plant interconnectivity and data accessibility. This paper reviews state-of-the-art research developments which offer the potential for real-time process safety and systems decision-making in the digital era. An overview is first presented on online process safety management approaches, including dynamic risk analysis and fault diagnosis/prognosis. Advanced operability and control methods are then discussed to achieve safely optimal operations under uncertainty (e.g., flexibility analysis, safety-aware control, fault-tolerant control). We highlight the connections between systems-based operation and process safety management to achieve operational excellence while proactively reducing potential safety losses. We also review the developments and showcases of digital twins paving the way to actual cyber–physical integration. Outstanding challenges and opportunities are identified such as safe data-driven control, integrated operability, safety and control, cyber–physical demonstration, etc. Toward this direction, we present our ongoing developments of the REal-Time Risk-based Optimization (RETRO) framework for safe and smart process operations.
{"title":"Real-time process safety and systems decision-making toward safe and smart chemical manufacturing","authors":"Austin Braniff , Sahithi Srijana Akundi , Yuanxing Liu , Beatriz Dantas , Shayan S. Niknezhad , Faisal Khan , Efstratios N. Pistikopoulos , Yuhe Tian","doi":"10.1016/j.dche.2025.100227","DOIUrl":"10.1016/j.dche.2025.100227","url":null,"abstract":"<div><div>The ongoing digital transformation has created new opportunities for chemical manufacturing with increasing plant interconnectivity and data accessibility. This paper reviews state-of-the-art research developments which offer the potential for real-time process safety and systems decision-making in the digital era. An overview is first presented on online process safety management approaches, including dynamic risk analysis and fault diagnosis/prognosis. Advanced operability and control methods are then discussed to achieve safely optimal operations under uncertainty (e.g., flexibility analysis, safety-aware control, fault-tolerant control). We highlight the connections between systems-based operation and process safety management to achieve operational excellence while proactively reducing potential safety losses. We also review the developments and showcases of digital twins paving the way to actual cyber–physical integration. Outstanding challenges and opportunities are identified such as safe data-driven control, integrated operability, safety and control, cyber–physical demonstration, etc. Toward this direction, we present our ongoing developments of the REal-Time Risk-based Optimization (RETRO) framework for safe and smart process operations.</div></div>","PeriodicalId":72815,"journal":{"name":"Digital Chemical Engineering","volume":"15 ","pages":"Article 100227"},"PeriodicalIF":3.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143628822","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-06-01Epub Date: 2025-04-25DOI: 10.1016/j.dche.2025.100237
Wenlong Wang , Zhe Wu , Dominic Peters , Berkay Citmaci , Carlos G. Morales-Guio , Panagiotis D. Christofides
Electrochemical reactors play a critical role in various industrial sectors, including energy storage, chemical production, and environmental engineering. The complexity of these systems – arising from coupled electrochemical reactions with mass, heat and charge transport phenomena – poses significant challenges in modeling, analysis, and control. Machine learning (ML) has emerged as a promising tool for addressing these challenges by providing data-driven solutions to complex process modeling, optimization, and advanced control. This tutorial review explores the state-of-the-art applications of ML in electrochemical reactor systems, including ML-based modeling techniques and ML-based advanced control strategies, followed by the discussions of practical challenges and their solutions. An electrochemical carbon dioxide (CO2) reduction reactor is used as a demonstration example to show the effectiveness of various modeling and control methods. In addition, an integrated data infrastructure platform is presented for the digitalization and control of the electrochemical CO2 reduction reactor. By identifying current gaps and future opportunities, this article provides a roadmap for leveraging ML tools to improve the analysis and operation of electrochemical reactors.
{"title":"Machine learning in modeling, analysis and control of electrochemical reactors: A tutorial review","authors":"Wenlong Wang , Zhe Wu , Dominic Peters , Berkay Citmaci , Carlos G. Morales-Guio , Panagiotis D. Christofides","doi":"10.1016/j.dche.2025.100237","DOIUrl":"10.1016/j.dche.2025.100237","url":null,"abstract":"<div><div>Electrochemical reactors play a critical role in various industrial sectors, including energy storage, chemical production, and environmental engineering. The complexity of these systems – arising from coupled electrochemical reactions with mass, heat and charge transport phenomena – poses significant challenges in modeling, analysis, and control. Machine learning (ML) has emerged as a promising tool for addressing these challenges by providing data-driven solutions to complex process modeling, optimization, and advanced control. This tutorial review explores the state-of-the-art applications of ML in electrochemical reactor systems, including ML-based modeling techniques and ML-based advanced control strategies, followed by the discussions of practical challenges and their solutions. An electrochemical carbon dioxide (CO<sub>2</sub>) reduction reactor is used as a demonstration example to show the effectiveness of various modeling and control methods. In addition, an integrated data infrastructure platform is presented for the digitalization and control of the electrochemical CO<sub>2</sub> reduction reactor. By identifying current gaps and future opportunities, this article provides a roadmap for leveraging ML tools to improve the analysis and operation of electrochemical reactors.</div></div>","PeriodicalId":72815,"journal":{"name":"Digital Chemical Engineering","volume":"15 ","pages":"Article 100237"},"PeriodicalIF":3.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143887941","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-06-01Epub Date: 2025-05-25DOI: 10.1016/j.dche.2025.100244
Arun K. Sharma, Owen McMillan, Selsela Arsala, Supreet Gandhok, Rylend Young
Asphaltenes are complex polycyclic organic molecules in crude oil that readily aggregate and precipitate under varying thermodynamic conditions. Their structural heterogeneity influences key physicochemical properties, including solubility, stability, and reactivity. Molecular polarizability, a crucial property governing intermolecular interactions and electronic behavior, remains challenging to predict due to this structural diversity. This study employs machine learning models to predict isotropic polarizability using two sets of molecular descriptors: WHIM and GETAWAY. A dataset of 255 asphaltene structures was analyzed using stratified sampling, generating 10 independent training (80 %) and testing (20 %) splits. The Wolfram Language’s Predict function evaluated multiple machine learning algorithms—including Random Forest, Decision Tree, Gradient Boosted Trees, Nearest Neighbors, Linear Regression, Gaussian Process, and Neural Network—through an automated model selection process, serving as an AutoML framework. Linear regression was the best-performing model in 9 out of 10 splits for GETAWAY descriptors. GETAWAY-based models achieved an average mean absolute deviation of 0.0920 ± 0.0030 and standard deviation of 0.113 ± 0.004, significantly outperforming WHIM-based models (MAD = 0.173 ± 0.007, STD = 0.224 ± 0.008) with paired t-tests confirming statistical significance (p < 0.001). While R² values were reported, their interpretability was limited by heterogeneity and narrow property ranges in some test sets. These findings demonstrate the effectiveness of AutoML-guided approaches for predicting molecular properties and identify GETAWAY descriptors as a robust, efficient basis for polarizability prediction. Accurate prediction of polarizability is essential for modeling intermolecular forces and improving force field design in petroleum and materials chemistry, issues that are central to industrial and chemical applications.
{"title":"Machine learning for asphaltene polarizability: Evaluating molecular descriptors","authors":"Arun K. Sharma, Owen McMillan, Selsela Arsala, Supreet Gandhok, Rylend Young","doi":"10.1016/j.dche.2025.100244","DOIUrl":"10.1016/j.dche.2025.100244","url":null,"abstract":"<div><div>Asphaltenes are complex polycyclic organic molecules in crude oil that readily aggregate and precipitate under varying thermodynamic conditions. Their structural heterogeneity influences key physicochemical properties, including solubility, stability, and reactivity. Molecular polarizability, a crucial property governing intermolecular interactions and electronic behavior, remains challenging to predict due to this structural diversity. This study employs machine learning models to predict isotropic polarizability using two sets of molecular descriptors: WHIM and GETAWAY. A dataset of 255 asphaltene structures was analyzed using stratified sampling, generating 10 independent training (80 %) and testing (20 %) splits. The Wolfram Language’s Predict function evaluated multiple machine learning algorithms—including Random Forest, Decision Tree, Gradient Boosted Trees, Nearest Neighbors, Linear Regression, Gaussian Process, and Neural Network—through an automated model selection process, serving as an AutoML framework. Linear regression was the best-performing model in 9 out of 10 splits for GETAWAY descriptors. GETAWAY-based models achieved an average mean absolute deviation of 0.0920 ± 0.0030 and standard deviation of 0.113 ± 0.004, significantly outperforming WHIM-based models (MAD = 0.173 ± 0.007, STD = 0.224 ± 0.008) with paired <em>t</em>-tests confirming statistical significance (<em>p</em> < 0.001). While R² values were reported, their interpretability was limited by heterogeneity and narrow property ranges in some test sets. These findings demonstrate the effectiveness of AutoML-guided approaches for predicting molecular properties and identify GETAWAY descriptors as a robust, efficient basis for polarizability prediction. Accurate prediction of polarizability is essential for modeling intermolecular forces and improving force field design in petroleum and materials chemistry, issues that are central to industrial and chemical applications.</div></div>","PeriodicalId":72815,"journal":{"name":"Digital Chemical Engineering","volume":"15 ","pages":"Article 100244"},"PeriodicalIF":3.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144184928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-06-01Epub Date: 2025-03-31DOI: 10.1016/j.dche.2025.100234
Homa Hamedi, Torsten Brinkmann
Green hydrogen (H₂) is a leading enabler for the decarbonization of hard-to-abate industries where electrification is either uneconomical or infeasible. Establishing an adequate and cost-effective infrastructure for hydrogen distribution remains one of the primary barriers to its widespread adoption. A promising short-term solution to this challenge involves H₂ storage and co-transportation via existing gas grids. For H₂ extraction from distribution gas grids, standalone pressure swing adsorption systems are considered the most viable option, whereas a hybrid process is suggested in the literature for transmission gas networks. This article presents a comprehensive techno-economic model for the proposed hybrid process, developed using an integrated platform based on Aspen Adsorption and Aspen Custom Modeler. The system consists of a single-stage hollow fiber Matrimid membrane module, followed by a 4-bed adsorption process operating in 8 sequential steps to meet H₂ market purity requirements with an acceptable recovery rate. Since the performances of these two separation modules, as an integrated system, significantly influence each other, the study identifies a unique opportunity to minimize separation costs through process optimization. To reduce computational time, a cyclic steady-state approach was employed to simulate the PSA process. Bayesian optimization, facilitated by the integration of Python with Aspen Adsorption, was used to efficiently identify the optimal solution with a minimal number of objective function evaluations. The levelized cost of H₂ separation (99.0 % purity at 10 bar) from natural gas containing 10 % H2 at pressures of 35 bar and 60 bar is estimated to be 2.7310 and, $2.5116/kg-H2, respectively. These estimates correspond to a scenario with 10 identical trains, each handling a feed flowrate of 200 kmol/hr. Increasing the number of trains keeps the cost contribution of PSA constant; however, the total cost decreases as the compression fixed cost is distributed across more trains.
{"title":"Green hydrogen extraction from natural gas transmission grids using hybrid membrane and PSA processes optimized via bayesian techniques","authors":"Homa Hamedi, Torsten Brinkmann","doi":"10.1016/j.dche.2025.100234","DOIUrl":"10.1016/j.dche.2025.100234","url":null,"abstract":"<div><div>Green hydrogen (H₂) is a leading enabler for the decarbonization of hard-to-abate industries where electrification is either uneconomical or infeasible. Establishing an adequate and cost-effective infrastructure for hydrogen distribution remains one of the primary barriers to its widespread adoption. A promising short-term solution to this challenge involves H₂ storage and co-transportation via existing gas grids. For H₂ extraction from distribution gas grids, standalone pressure swing adsorption systems are considered the most viable option, whereas a hybrid process is suggested in the literature for transmission gas networks. This article presents a comprehensive techno-economic model for the proposed hybrid process, developed using an integrated platform based on Aspen Adsorption and Aspen Custom Modeler. The system consists of a single-stage hollow fiber Matrimid membrane module, followed by a 4-bed adsorption process operating in 8 sequential steps to meet H₂ market purity requirements with an acceptable recovery rate. Since the performances of these two separation modules, as an integrated system, significantly influence each other, the study identifies a unique opportunity to minimize separation costs through process optimization. To reduce computational time, a cyclic steady-state approach was employed to simulate the PSA process. Bayesian optimization, facilitated by the integration of Python with Aspen Adsorption, was used to efficiently identify the optimal solution with a minimal number of objective function evaluations. The levelized cost of H₂ separation (99.0 % purity at 10 bar) from natural gas containing 10 % H<sub>2</sub> at pressures of 35 bar and 60 bar is estimated to be 2.7310 and, $2.5116/kg-H<sub>2</sub>, respectively. These estimates correspond to a scenario with 10 identical trains, each handling a feed flowrate of 200 kmol/hr. Increasing the number of trains keeps the cost contribution of PSA constant; however, the total cost decreases as the compression fixed cost is distributed across more trains.</div></div>","PeriodicalId":72815,"journal":{"name":"Digital Chemical Engineering","volume":"15 ","pages":"Article 100234"},"PeriodicalIF":3.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143776826","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-06-01Epub Date: 2025-05-26DOI: 10.1016/j.dche.2025.100239
Marcin Pietrasik , Anna Wilbik , Yannick Damoiseaux , Tessa Derks , Emery Karambiri , Shirley de Koster , Daniel van der Velde , Kim Ragaert , Sin Yong Teng
Plastic mechanical recycling is the conventional technological step towards circularity. In such aspects, complex mixtures of polyolefin blends are often fed into mechanical recycling systems, resulting in moulded products with uncertain quality. To add to the difficulty of heterogeneous feedstocks, the testing of mechanical properties for plastic products often results in stochastic measurements, making connections from material prediction to systems understanding challenging. This research is aimed at providing a framework capable of generalizing stochastic plastic recycling knowledge via interval-based machine learning for the prediction of properties formulation for unrecycled plastics. The framework is made up of two components: a regressor for point estimation and an interval predictor for generating prediction intervals. We compare several competing methods for each of these components through empirical evaluation on a real-world dataset. The results demonstrate the usefulness of interval-based machine learning in the application of stochastic engineering problems such as plastic mechanical recycling, highlighting such approaches towards better model interpretation and (un)certainty prediction regions.
{"title":"Capturing variability in material property predictions for plastics recycling via machine learning","authors":"Marcin Pietrasik , Anna Wilbik , Yannick Damoiseaux , Tessa Derks , Emery Karambiri , Shirley de Koster , Daniel van der Velde , Kim Ragaert , Sin Yong Teng","doi":"10.1016/j.dche.2025.100239","DOIUrl":"10.1016/j.dche.2025.100239","url":null,"abstract":"<div><div>Plastic mechanical recycling is the conventional technological step towards circularity. In such aspects, complex mixtures of polyolefin blends are often fed into mechanical recycling systems, resulting in moulded products with uncertain quality. To add to the difficulty of heterogeneous feedstocks, the testing of mechanical properties for plastic products often results in stochastic measurements, making connections from material prediction to systems understanding challenging. This research is aimed at providing a framework capable of generalizing stochastic plastic recycling knowledge via interval-based machine learning for the prediction of properties formulation for unrecycled plastics. The framework is made up of two components: a regressor for point estimation and an interval predictor for generating prediction intervals. We compare several competing methods for each of these components through empirical evaluation on a real-world dataset. The results demonstrate the usefulness of interval-based machine learning in the application of stochastic engineering problems such as plastic mechanical recycling, highlighting such approaches towards better model interpretation and (un)certainty prediction regions.</div></div>","PeriodicalId":72815,"journal":{"name":"Digital Chemical Engineering","volume":"15 ","pages":"Article 100239"},"PeriodicalIF":3.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144203735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-06-01Epub Date: 2025-05-29DOI: 10.1016/j.dche.2025.100245
Bol Ram, Z Ahmad, N Md Nor
Crude oil remains a vital non-renewable resource that supports numerous industries in the current era of industrial advancement. Consequently, petroleum refineries face increasing challenges, including stringent environmental regulations, fluctuating feedstock quality, rising demand, safety requirements, and the need for cost optimization. These challenges, coupled with the inherent complexity of the Crude Distillation Unit (CDU), demand advanced control strategies to ensure stable and efficient operation. This study investigates the application of Dynamic Matrix Control (DMC), a subset of Model Predictive Control (MPC), using Aspen DMC3 for CDU process control—a novel implementation not previously explored. The methodology involves three main stages: validation of a CDU simulation based on real data from the Basrah refinery, generation of dynamic response data through MATLAB integrated with Aspen Dynamics, and the development of a DMC controller using Aspen DMC3. The performance of the DMC controller is compared against conventional Proportional-Integral-Derivative (PID) controllers implemented in Aspen Dynamics using key indicators such as settling time, offset error, maximum deviation, and response smoothness. Results demonstrate that the DMC controller provides superior control performance, with faster settling times, zero offset, minimal deviations, and smoother responses. Additionally, Aspen DMC3′s AI-assisted capabilities enable streamlined controller configuration and real-time optimization through server connectivity, highlighting its potential for robust and efficient CDU operation.
{"title":"Utilization of aspen DMC3 in process control of crude distillation unit (CDU)","authors":"Bol Ram, Z Ahmad, N Md Nor","doi":"10.1016/j.dche.2025.100245","DOIUrl":"10.1016/j.dche.2025.100245","url":null,"abstract":"<div><div>Crude oil remains a vital non-renewable resource that supports numerous industries in the current era of industrial advancement. Consequently, petroleum refineries face increasing challenges, including stringent environmental regulations, fluctuating feedstock quality, rising demand, safety requirements, and the need for cost optimization. These challenges, coupled with the inherent complexity of the Crude Distillation Unit (CDU), demand advanced control strategies to ensure stable and efficient operation. This study investigates the application of Dynamic Matrix Control (DMC), a subset of Model Predictive Control (MPC), using Aspen DMC3 for CDU process control—a novel implementation not previously explored. The methodology involves three main stages: validation of a CDU simulation based on real data from the Basrah refinery, generation of dynamic response data through MATLAB integrated with Aspen Dynamics, and the development of a DMC controller using Aspen DMC3. The performance of the DMC controller is compared against conventional Proportional-Integral-Derivative (PID) controllers implemented in Aspen Dynamics using key indicators such as settling time, offset error, maximum deviation, and response smoothness. Results demonstrate that the DMC controller provides superior control performance, with faster settling times, zero offset, minimal deviations, and smoother responses. Additionally, Aspen DMC3′s AI-assisted capabilities enable streamlined controller configuration and real-time optimization through server connectivity, highlighting its potential for robust and efficient CDU operation.</div></div>","PeriodicalId":72815,"journal":{"name":"Digital Chemical Engineering","volume":"15 ","pages":"Article 100245"},"PeriodicalIF":3.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144223561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}