Pub Date : 2025-09-01Epub Date: 2025-05-26DOI: 10.1016/j.compind.2025.104313
Mingda Chen , Ruiyun Yu , Zhiyuan Liang , Kun Li , Haifei Qi
In the manufacturing industry, the demand for fault-prediction solutions is increasing to prevent unexpected downtimes and reduce maintenance costs. Although deep-learning methods have demonstrated excellent performance in this domain, the current methods typically overlook the analysis of variable and random processes within mixed-model production, which is a manufacturing strategy that offers flexibility and efficiency in satisfying diverse consumer demands. Hence, we propose the multiscale process-aware retention network (MPRNet), which segments a time series into multiscale patches, thus enabling the model to focus on local information within each production process and correlations across all production processes. Furthermore, the network incorporates a cross-channel interaction module designed to dynamically capture the interactions between various sensor data types using a graph attention network, as well as transmit fault information across processes using state equations. We validate our proposed model on the BBA stud welding gun dataset and four additional open case studies. Compared with other established fault-prediction and time-series models, the MPRNet demonstrates improved F1-score by 13.1% in the BBA case and consistently achieves the best or near-best results in the open case studies.
{"title":"A multiscale process-aware retention network for fault prediction in mixed-model production","authors":"Mingda Chen , Ruiyun Yu , Zhiyuan Liang , Kun Li , Haifei Qi","doi":"10.1016/j.compind.2025.104313","DOIUrl":"10.1016/j.compind.2025.104313","url":null,"abstract":"<div><div>In the manufacturing industry, the demand for fault-prediction solutions is increasing to prevent unexpected downtimes and reduce maintenance costs. Although deep-learning methods have demonstrated excellent performance in this domain, the current methods typically overlook the analysis of variable and random processes within mixed-model production, which is a manufacturing strategy that offers flexibility and efficiency in satisfying diverse consumer demands. Hence, we propose the multiscale process-aware retention network (MPRNet), which segments a time series into multiscale patches, thus enabling the model to focus on local information within each production process and correlations across all production processes. Furthermore, the network incorporates a cross-channel interaction module designed to dynamically capture the interactions between various sensor data types using a graph attention network, as well as transmit fault information across processes using state equations. We validate our proposed model on the BBA stud welding gun dataset and four additional open case studies. Compared with other established fault-prediction and time-series models, the MPRNet demonstrates improved F1-score by 13.1% in the BBA case and consistently achieves the best or near-best results in the open case studies.</div></div>","PeriodicalId":55219,"journal":{"name":"Computers in Industry","volume":"170 ","pages":"Article 104313"},"PeriodicalIF":8.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144139287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-09DOI: 10.1016/j.compind.2025.104347
Jorge Laguna, Mario E. Suaza-Medina, Rubén Béjar, Javier Lacasta, F. Javier Zarazaga-Soria
Industry 4.0 has advanced in agriculture through Smart Agriculture initiatives, yet open-field farming lags in the adoption of digital twins. Although digital twins have transformed manufacturing since 2011, their application in open-field farming remains limited by environmental variability, data scarcity, and financial constraints. This paper addresses four gaps: the lack of affordable platforms for small farms that dominate European agriculture; the need to manage agricultural complexity through data-driven models rather than the physical modelling approaches prevalent in non-agricultural sectors; the absence of open sources solutions adapted to agriculture’s slower innovation pace; the breach between technology and farmers. The platform features innovations in data workflow integration, open data incorporation, a cost-effective shared warehouse, and scalable data pipelines. To validate the proposed platform, a case study with two example digital twins mirroring two fields is conducted. This implementation ran efficiently on modest hardware (2 vCPUs, 4GB RAM). It achieved an average CPU usage of 60%, RAM usage of 2.5 GB, and a deployment time of around one minute. This helps lowering adoption barriers for small holdings and bridging the gap between basic monitoring and complex future systems.
{"title":"A platform to support the fast development of digital twins for agricultural holdings","authors":"Jorge Laguna, Mario E. Suaza-Medina, Rubén Béjar, Javier Lacasta, F. Javier Zarazaga-Soria","doi":"10.1016/j.compind.2025.104347","DOIUrl":"https://doi.org/10.1016/j.compind.2025.104347","url":null,"abstract":"Industry 4.0 has advanced in agriculture through Smart Agriculture initiatives, yet open-field farming lags in the adoption of digital twins. Although digital twins have transformed manufacturing since 2011, their application in open-field farming remains limited by environmental variability, data scarcity, and financial constraints. This paper addresses four gaps: the lack of affordable platforms for small farms that dominate European agriculture; the need to manage agricultural complexity through data-driven models rather than the physical modelling approaches prevalent in non-agricultural sectors; the absence of open sources solutions adapted to agriculture’s slower innovation pace; the breach between technology and farmers. The platform features innovations in data workflow integration, open data incorporation, a cost-effective shared warehouse, and scalable data pipelines. To validate the proposed platform, a case study with two example digital twins mirroring two fields is conducted. This implementation ran efficiently on modest hardware (2 vCPUs, 4GB RAM). It achieved an average CPU usage of 60%, RAM usage of 2.5 GB, and a deployment time of around one minute. This helps lowering adoption barriers for small holdings and bridging the gap between basic monitoring and complex future systems.","PeriodicalId":55219,"journal":{"name":"Computers in Industry","volume":"18 1","pages":""},"PeriodicalIF":10.0,"publicationDate":"2025-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144901449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-01Epub Date: 2025-04-30DOI: 10.1016/j.compind.2025.104302
Caihua Hao , Zhaoyu Wang , Xinyong Mao , Songping He , Bin Li , Hongqi Liu , Fangyu Peng , Weiye Li
Accurately predicting the future wear of cutting tools with variable geometric parameters remains a significant challenge. Existing methods lack the capability to model long-term temporal dependencies and predict future wear values—a key characteristic of world models. To address this challenge, we introduce the Tool-Multimodal Generative Pre-trained Transformer (Tool-MMGPT), a novel and scalable multimodal large language model (MLLM) architecture specifically designed for tool wear prediction. Tool-MMGPT pioneers the first tool wear world model by uniquely unifying multimodal data, extending beyond conventional static dimensions to incorporate dynamic temporal dimensions. This approach extracts modality-specific information and achieves shared spatiotemporal feature fusion through a cross-modal Transformer. Subsequently, alignment and joint interpretation occur within a unified representation space via a multimodal-language projector, which effectively accommodates the comprehensive input characteristics required by world models. This article proposes an effective cross-modal fusion module for vibration signals and images, aiming to fully leverage the advantages of multimodal information. Crucially, Tool-MMGPT transcends the limitations of traditional Large Language Models (LLMs) through an innovative yet generalizable method. By fundamentally reconstructing the output layer and redefining training objectives, we repurpose LLMs for numerical regression tasks, thereby establishing a novel bridge that connects textual representations to continuous numerical predictions. This enables the direct and accurate long-term forecasting of future wear time series. Extensive experiments conducted on a newly developed multimodal dataset for variable geometry tools demonstrate that Tool-MMGPT significantly outperforms state-of-the-art (SOTA) baseline methods. These results highlight the model's superior long-context modeling capabilities and illustrate its potential for effective deployment in environments with limited computational resources.
{"title":"A novel and scalable multimodal large language model architecture Tool-MMGPT for future tool wear prediction in titanium alloy high-speed milling processes","authors":"Caihua Hao , Zhaoyu Wang , Xinyong Mao , Songping He , Bin Li , Hongqi Liu , Fangyu Peng , Weiye Li","doi":"10.1016/j.compind.2025.104302","DOIUrl":"10.1016/j.compind.2025.104302","url":null,"abstract":"<div><div>Accurately predicting the future wear of cutting tools with variable geometric parameters remains a significant challenge. Existing methods lack the capability to model long-term temporal dependencies and predict future wear values—a key characteristic of world models. To address this challenge, we introduce the Tool-Multimodal Generative Pre-trained Transformer (Tool-MMGPT), a novel and scalable multimodal large language model (MLLM) architecture specifically designed for tool wear prediction. Tool-MMGPT pioneers the first tool wear world model by uniquely unifying multimodal data, extending beyond conventional static dimensions to incorporate dynamic temporal dimensions. This approach extracts modality-specific information and achieves shared spatiotemporal feature fusion through a cross-modal Transformer. Subsequently, alignment and joint interpretation occur within a unified representation space via a multimodal-language projector, which effectively accommodates the comprehensive input characteristics required by world models. This article proposes an effective cross-modal fusion module for vibration signals and images, aiming to fully leverage the advantages of multimodal information. Crucially, Tool-MMGPT transcends the limitations of traditional Large Language Models (LLMs) through an innovative yet generalizable method. By fundamentally reconstructing the output layer and redefining training objectives, we repurpose LLMs for numerical regression tasks, thereby establishing a novel bridge that connects textual representations to continuous numerical predictions. This enables the direct and accurate long-term forecasting of future wear time series. Extensive experiments conducted on a newly developed multimodal dataset for variable geometry tools demonstrate that Tool-MMGPT significantly outperforms state-of-the-art (SOTA) baseline methods. These results highlight the model's superior long-context modeling capabilities and illustrate its potential for effective deployment in environments with limited computational resources.</div></div>","PeriodicalId":55219,"journal":{"name":"Computers in Industry","volume":"169 ","pages":"Article 104302"},"PeriodicalIF":8.2,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143886351","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-01Epub Date: 2025-04-07DOI: 10.1016/j.compind.2025.104288
Yusong Zhang , Zhenyu Liu , Guodong Sa , Jiacheng Sun , Mingjie Hou , Yougen Huang , Jianrong Tan
In practical application scenarios such as air quality, traffic and mechanical processing, sensors are often constrained by spatial capacity, geometric structures, extreme environments and other factors, making it impossible to place them in critical monitoring areas. To address this issue, a novel virtual sensor state perception generalization framework, the Virtual-Real Spatial-Temporal Dual Layer Transformer (VR-STDT) model is proposed. It constructs a spatial-temporal correlation model between real sensors and unobservable virtual sensors, to solve the problem of missing information in sensor-restricted zones. Considering the “stop-start” single-operation system with a short time window and high sampling frequency, a historical similar attention mechanism and a convolution-based time patching mechanism are proposed to effectively solve the contradiction between low resolution and information loss. Finally, verification was carried out in practical application scenarios, such as the kitchen particle concentration diffusion experiment platform and the machine tool spindle temperature experiment platform, and then the open urban air quality data set was used for auxiliary verification. The results show that the proposed model achieved an average performance improvement of 10.20 % over existing inter-node spatial-temporal prediction models.
在空气质量、交通、机械加工等实际应用场景中,传感器往往受到空间容量、几何结构、极端环境等因素的限制,无法将其放置在关键的监测区域。针对这一问题,提出了一种新的虚拟传感器状态感知概化框架——虚拟-真实时空双层变压器(virtual - real Spatial-Temporal Dual Layer Transformer, VR-STDT)模型。构建了真实传感器与不可观测虚拟传感器之间的时空关联模型,解决了传感器禁区信息缺失问题。针对短时间窗、高采样频率的“启停”单操作系统,提出了历史相似注意机制和基于卷积的时间补片机制,有效解决了低分辨率与信息丢失之间的矛盾。最后,在厨房颗粒浓度扩散实验平台和机床主轴温度实验平台等实际应用场景中进行验证,然后利用开放的城市空气质量数据集进行辅助验证。结果表明,与现有的节点间时空预测模型相比,该模型的平均性能提高了10.20 %。
{"title":"Virtual-Real Spatial-Temporal Dual Layer Transformer for virtual sensor state perception","authors":"Yusong Zhang , Zhenyu Liu , Guodong Sa , Jiacheng Sun , Mingjie Hou , Yougen Huang , Jianrong Tan","doi":"10.1016/j.compind.2025.104288","DOIUrl":"10.1016/j.compind.2025.104288","url":null,"abstract":"<div><div>In practical application scenarios such as air quality, traffic and mechanical processing, sensors are often constrained by spatial capacity, geometric structures, extreme environments and other factors, making it impossible to place them in critical monitoring areas. To address this issue, a novel virtual sensor state perception generalization framework, the Virtual-Real Spatial-Temporal Dual Layer Transformer (VR-STDT) model is proposed. It constructs a spatial-temporal correlation model between real sensors and unobservable virtual sensors, to solve the problem of missing information in sensor-restricted zones. Considering the “stop-start” single-operation system with a short time window and high sampling frequency, a historical similar attention mechanism and a convolution-based time patching mechanism are proposed to effectively solve the contradiction between low resolution and information loss. Finally, verification was carried out in practical application scenarios, such as the kitchen particle concentration diffusion experiment platform and the machine tool spindle temperature experiment platform, and then the open urban air quality data set was used for auxiliary verification. The results show that the proposed model achieved an average performance improvement of 10.20 % over existing inter-node spatial-temporal prediction models.</div></div>","PeriodicalId":55219,"journal":{"name":"Computers in Industry","volume":"169 ","pages":"Article 104288"},"PeriodicalIF":8.2,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143792594","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-01Epub Date: 2025-04-15DOI: 10.1016/j.compind.2025.104291
Wang Chen , Binhong Yuan , Dongliang Chen , Yong Hu , Feiyu Wang , Jian Zhang
Inspecting the underside of large-span bridges is a major challenge due to the extensive area and inaccessibility. This study developed a system that integrates advanced equipment with intelligent algorithms, designed to achieve precise identification and rapid localization of defects on the underside of bridges. The key components of the system are summarized as follows: (1) The dynamic visual perception system is composed of a perception module, a control and transmission module, and a motion module. It enables automated data collection at any position beneath the bridge structure. (2) A block-based panoramic generation strategy is employed, which uses a spatially ordered block concept to simplify the panorama stitching process and enhance accuracy. (3) Deep learning-driven two-phase synchronous identification and localization method. In the first phase, MobileNetV4 serves as the primary feature representation tool, facilitating the lightweight reconstruction of panoramic images. In the second phase, the YOLOv9 detection framework is employed to perform a precise analysis of the identified defect regions, providing detailed defect information on a localized level. The design of this system significantly enhances the efficiency and accuracy of inspections of large-span bridge undersides, offering robust technical support for bridge health maintenance. Experimental results indicate that the proposed method achieves over 90 % accuracy in defect recognition tasks, alongside millimeter-level precision in localization.
{"title":"Synchronized identification and localization of defect on the bottom of steel box girders based on a dynamic visual perception system","authors":"Wang Chen , Binhong Yuan , Dongliang Chen , Yong Hu , Feiyu Wang , Jian Zhang","doi":"10.1016/j.compind.2025.104291","DOIUrl":"10.1016/j.compind.2025.104291","url":null,"abstract":"<div><div>Inspecting the underside of large-span bridges is a major challenge due to the extensive area and inaccessibility. This study developed a system that integrates advanced equipment with intelligent algorithms, designed to achieve precise identification and rapid localization of defects on the underside of bridges. The key components of the system are summarized as follows: (1) The dynamic visual perception system is composed of a perception module, a control and transmission module, and a motion module. It enables automated data collection at any position beneath the bridge structure. (2) A block-based panoramic generation strategy is employed, which uses a spatially ordered block concept to simplify the panorama stitching process and enhance accuracy. (3) Deep learning-driven two-phase synchronous identification and localization method. In the first phase, MobileNetV4 serves as the primary feature representation tool, facilitating the lightweight reconstruction of panoramic images. In the second phase, the YOLOv9 detection framework is employed to perform a precise analysis of the identified defect regions, providing detailed defect information on a localized level. The design of this system significantly enhances the efficiency and accuracy of inspections of large-span bridge undersides, offering robust technical support for bridge health maintenance. Experimental results indicate that the proposed method achieves over 90 % accuracy in defect recognition tasks, alongside millimeter-level precision in localization.</div></div>","PeriodicalId":55219,"journal":{"name":"Computers in Industry","volume":"169 ","pages":"Article 104291"},"PeriodicalIF":8.2,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143830140","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-01Epub Date: 2025-04-03DOI: 10.1016/j.compind.2025.104289
Heiner Ludwig, Thorsten Schmidt, Mathias Kühn
This paper presents a novel approach to support complex maintenance procedures through a dialogue-driven digital assistant using an ontology-based retrieval augmented generation method. The core of the proposed system relies on the strong formalisation capabilities of the graph-based Web Ontology Language (OWL), combined with various retrieval algorithms and different Large Language Models (LLMs) to determine the most useful context for answering user queries. To do this, we use the popular principle of Retrieval Augmented Generation (RAG). Graph traversal enriches the contextual knowledge, enabling more accurate and context-aware responses. An evaluation using an OWL example ontology and an extensive Q&A dataset demonstrates the improved retrieval quality achieved by combining classical and vector-based semantic matching methods. The community-driven analysis of generation quality illustrates the usability of an OWL-based assistant for maintenance procedures on the basis of contexts and LLMs of varying configurations.
{"title":"An ontology-based retrieval augmented generation procedure for a voice-controlled maintenance assistant","authors":"Heiner Ludwig, Thorsten Schmidt, Mathias Kühn","doi":"10.1016/j.compind.2025.104289","DOIUrl":"10.1016/j.compind.2025.104289","url":null,"abstract":"<div><div>This paper presents a novel approach to support complex maintenance procedures through a dialogue-driven digital assistant using an ontology-based retrieval augmented generation method. The core of the proposed system relies on the strong formalisation capabilities of the graph-based Web Ontology Language (OWL), combined with various retrieval algorithms and different Large Language Models (LLMs) to determine the most useful context for answering user queries. To do this, we use the popular principle of Retrieval Augmented Generation (RAG). Graph traversal enriches the contextual knowledge, enabling more accurate and context-aware responses. An evaluation using an OWL example ontology and an extensive Q&A dataset demonstrates the improved retrieval quality achieved by combining classical and vector-based semantic matching methods. The community-driven analysis of generation quality illustrates the usability of an OWL-based assistant for maintenance procedures on the basis of contexts and LLMs of varying configurations.</div></div>","PeriodicalId":55219,"journal":{"name":"Computers in Industry","volume":"169 ","pages":"Article 104289"},"PeriodicalIF":8.2,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143760341","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-01Epub Date: 2025-04-24DOI: 10.1016/j.compind.2025.104304
Chunhao Jiang , Nian-Zhong Chen
Physics-informed neural networks (PINNs) face significant challenges to predict the vortex shedding in the flow past a two-dimensional cylinder, mainly due to complex loss landscapes, spectral bias, and a lack of inductive bias towards periodic functions. To overcome these challenges, a novel gradient-free PINN (GF-PINN) coupled with a U-Net+ + architecture is proposed. For optimizing the complex loss landscape, the skip pathways in U-Net+ + are redesigned to reduce the semantic gap between encoder and decoder feature maps. Then, the stream function instead of velocity, is used as the input and output for the neural network, ensuring flow incompressibility and reducing output dimensionality. This approach aims to overcome the inherent problems of spectral bias and the lack of inductive bias towards periodic functions in PINNs. Furthermore, gradient-free convolutional filters are employed to approximate the derivative terms in the loss function to further optimize the complex loss landscape. A series of numerical experiments and dynamic mode analyses are conducted and the results show that the vortex shedding in the wake of a square cylinder is successfully captured by the proposed model and the estimated drag coefficients and Strouhal numbers are in a good agreement with those predicted by traditional methods. In addition, numerical experiments also show that the model exhibits great capabilities of generalization and extrapolation. This work demonstrates the potential of PINN-based models to effectively solve complex fluid dynamics problems.
基于物理信息的神经网络(pinn)在预测流过二维圆柱体的流体中的涡落方面面临着重大挑战,这主要是由于复杂的损失、光谱偏倚和缺乏对周期函数的归纳偏倚。为了克服这些挑战,提出了一种新型的无梯度pin - n (GF-PINN)结合U-Net+ +架构。为了优化复杂的损失情况,重新设计了U-Net+ +中的跳过路径,以减少编码器和解码器特征映射之间的语义差距。然后,用流函数代替速度作为神经网络的输入和输出,保证了流不可压缩性,降低了输出维数。该方法旨在克服pinn中固有的频谱偏置和对周期函数缺乏归纳偏置的问题。此外,采用无梯度卷积滤波器对损失函数中的导数项进行近似,进一步优化复杂损失格局。通过一系列的数值实验和动力模态分析,结果表明,该模型成功地捕获了方形圆柱体尾迹的涡脱落,所估计的阻力系数和Strouhal数与传统方法预测的结果吻合较好。此外,数值实验还表明,该模型具有良好的泛化和外推能力。这项工作证明了基于pup模型有效解决复杂流体动力学问题的潜力。
{"title":"Gradient-free physics-informed neural networks (GF-PINNs) for vortex shedding prediction in flow past square cylinders","authors":"Chunhao Jiang , Nian-Zhong Chen","doi":"10.1016/j.compind.2025.104304","DOIUrl":"10.1016/j.compind.2025.104304","url":null,"abstract":"<div><div>Physics-informed neural networks (PINNs) face significant challenges to predict the vortex shedding in the flow past a two-dimensional cylinder, mainly due to complex loss landscapes, spectral bias, and a lack of inductive bias towards periodic functions. To overcome these challenges, a novel gradient-free PINN (GF-PINN) coupled with a U-Net+ + architecture is proposed. For optimizing the complex loss landscape, the skip pathways in U-Net+ + are redesigned to reduce the semantic gap between encoder and decoder feature maps. Then, the stream function instead of velocity, is used as the input and output for the neural network, ensuring flow incompressibility and reducing output dimensionality. This approach aims to overcome the inherent problems of spectral bias and the lack of inductive bias towards periodic functions in PINNs. Furthermore, gradient-free convolutional filters are employed to approximate the derivative terms in the loss function to further optimize the complex loss landscape. A series of numerical experiments and dynamic mode analyses are conducted and the results show that the vortex shedding in the wake of a square cylinder is successfully captured by the proposed model and the estimated drag coefficients and Strouhal numbers are in a good agreement with those predicted by traditional methods. In addition, numerical experiments also show that the model exhibits great capabilities of generalization and extrapolation. This work demonstrates the potential of PINN-based models to effectively solve complex fluid dynamics problems.</div></div>","PeriodicalId":55219,"journal":{"name":"Computers in Industry","volume":"169 ","pages":"Article 104304"},"PeriodicalIF":8.2,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143869738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-01Epub Date: 2025-04-25DOI: 10.1016/j.compind.2025.104305
Jixiang Ding , Zhengding Zheng , Shayu Song , Long Bai , Jianfeng Xu , Jianguo Zhang , Wenjie Chen
Particle-reinforced metal matrix composites (PRMMCs) are widely used because of their exceptional material properties. Online control of the laser field to soften and modify the reinforcing particles on the machined surface of the composites is an effective way to improve the machinability and machining quality of PRMMCs. A real-time segmentation method for reinforcing particles in PRMMCs is proposed. First, real-time acquisition of reinforcing particle images along the processing path is achieved using machine vision, and cutting region images are determined. Next, to improve the model’s ability to effectively segment the reinforcing particles in low-resolution images of the machining region, a reinforcing particle segmentation network (RPSNet) is proposed, incorporating a multimodal fusion and space-to-depth convolution module. Subsequently, position signals along the cutting direction are obtained by using a sliding window method. The effectiveness of each module and the performance of the model are analyzed and verified through comparative and ablation experiments. The results demonstrated that the proposed RPSNet achieved a mean average precision (mAP) of 95.4 % in segmenting reinforcing particles, with an inference time of 5.8 ms. In comparison to other methods, it demonstrated better real-time performance and accuracy. Additionally, the proposed method can convert image information into position signals, thus enabling real-time control of the laser for softening and modifying the reinforcing particles.
{"title":"Toward laser-assisted cutting: A real-time segmentation method for reinforcing particles in particle-reinforced metal matrix composites","authors":"Jixiang Ding , Zhengding Zheng , Shayu Song , Long Bai , Jianfeng Xu , Jianguo Zhang , Wenjie Chen","doi":"10.1016/j.compind.2025.104305","DOIUrl":"10.1016/j.compind.2025.104305","url":null,"abstract":"<div><div>Particle-reinforced metal matrix composites (PRMMCs) are widely used because of their exceptional material properties. Online control of the laser field to soften and modify the reinforcing particles on the machined surface of the composites is an effective way to improve the machinability and machining quality of PRMMCs. A real-time segmentation method for reinforcing particles in PRMMCs is proposed. First, real-time acquisition of reinforcing particle images along the processing path is achieved using machine vision, and cutting region images are determined. Next, to improve the model’s ability to effectively segment the reinforcing particles in low-resolution images of the machining region, a reinforcing particle segmentation network (RPSNet) is proposed, incorporating a multimodal fusion and space-to-depth convolution module. Subsequently, position signals along the cutting direction are obtained by using a sliding window method. The effectiveness of each module and the performance of the model are analyzed and verified through comparative and ablation experiments. The results demonstrated that the proposed RPSNet achieved a mean average precision (mAP) of 95.4 % in segmenting reinforcing particles, with an inference time of 5.8 ms. In comparison to other methods, it demonstrated better real-time performance and accuracy. Additionally, the proposed method can convert image information into position signals, thus enabling real-time control of the laser for softening and modifying the reinforcing particles.</div></div>","PeriodicalId":55219,"journal":{"name":"Computers in Industry","volume":"169 ","pages":"Article 104305"},"PeriodicalIF":8.2,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143869739","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-01Epub Date: 2025-04-24DOI: 10.1016/j.compind.2025.104303
Jieyang Peng , Andreas Kimmig , Simon Kreuzwieser , Zhibin Niu , Xiaoming Tao , Jivka Ovtcharova
In the fast-paced manufacturing industry, rapid and efficient product design is essential for meeting customer demands and maintaining a competitive edge. Despite advancements, transforming 2D design concepts into accurate 3D models remains a complex challenge, primarily due to the non-differentiability of traditional rendering processes that hinder gradient-based optimizations. To address this limitation, this paper introduces an innovative dual-decoder architecture that effectively separates the shape and color components of 3D models. By assigning separate decoders for vertex positions and color assignment, our proposed model enables targeted optimization of each, leading to more refined and authentic 3D reconstructions. Moreover, we have overcome the non-differentiability issue, enabling gradient-based learning through the incorporation of differentiable rendering techniques. These techniques facilitate gradient-based optimization, paving the way for data-driven enhancements in the design process. Our empirical research has demonstrated the effectiveness of our approach in generating high-fidelity 3D models from 2D inputs. Additionally, we have shed light on the sensitivity of hyperparameters within our framework, offering valuable insights for future model refinement and optimization. In summary, our research provides valuable insights into enhancing 3D modeling frameworks, thereby contributing to incremental progress in the field of computer-aided design and manufacturing.
{"title":"3D modeling from a single image via a novel dual-decoder framework for Agile design","authors":"Jieyang Peng , Andreas Kimmig , Simon Kreuzwieser , Zhibin Niu , Xiaoming Tao , Jivka Ovtcharova","doi":"10.1016/j.compind.2025.104303","DOIUrl":"10.1016/j.compind.2025.104303","url":null,"abstract":"<div><div>In the fast-paced manufacturing industry, rapid and efficient product design is essential for meeting customer demands and maintaining a competitive edge. Despite advancements, transforming 2D design concepts into accurate 3D models remains a complex challenge, primarily due to the non-differentiability of traditional rendering processes that hinder gradient-based optimizations. To address this limitation, this paper introduces an innovative dual-decoder architecture that effectively separates the shape and color components of 3D models. By assigning separate decoders for vertex positions and color assignment, our proposed model enables targeted optimization of each, leading to more refined and authentic 3D reconstructions. Moreover, we have overcome the non-differentiability issue, enabling gradient-based learning through the incorporation of differentiable rendering techniques. These techniques facilitate gradient-based optimization, paving the way for data-driven enhancements in the design process. Our empirical research has demonstrated the effectiveness of our approach in generating high-fidelity 3D models from 2D inputs. Additionally, we have shed light on the sensitivity of hyperparameters within our framework, offering valuable insights for future model refinement and optimization. In summary, our research provides valuable insights into enhancing 3D modeling frameworks, thereby contributing to incremental progress in the field of computer-aided design and manufacturing.</div></div>","PeriodicalId":55219,"journal":{"name":"Computers in Industry","volume":"169 ","pages":"Article 104303"},"PeriodicalIF":8.2,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143869740","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-01Epub Date: 2025-04-25DOI: 10.1016/j.compind.2025.104290
Xiangkai Shen , Jinhai Liu , Yifu Ren , Lin Jiang , Lei Wang , He Zhao , Rui Li
Defect diagnosis based on magnetic flux leakage (MFL) signals is an important process for assessing pipeline health, including defect detection and size quantification. However, existing studies suffer from poor consistency of results, because they regard defect detection and size quantification as separate tasks, lacking paradigm harmonization and interaction. In addition, the calibration of experts is required to achieve harmonization between the two, which increases the time cost of data analysis. To address the above challenges, our motivation is to synergistically learn two tasks within a unified framework and utilize their task properties for mutual benefit. Therefore, a novel defect diagnosis method based on a task-oriented physical collaborative network (TOPC-Net) is proposed, which is the first attempt at joint defect detection and size quantification in MFL inspection. First, a feature extraction subnetwork with a heterogeneous focus module is proposed to decompose initial task-specific features from shared spaces. Second, considering the strong correlation between the two tasks, a cross-task information awareness method is proposed to realize the information interaction between the two tasks, so that the task-specific features can be enhanced. Finally, a physical information-guided collaborative decision subnetwork is proposed to jointly optimize two tasks, where MFL domain knowledge is embedded into the subnetwork to provide expert guidance, ensuring the accuracy and stability of predictions. Experimental results show that the proposed method outperforms existing methods, with a detection accuracy of 96.0% and an average improvement of 7.5% in quantification accuracy, which makes it promising for industrial applications.
{"title":"A task-oriented physical collaborative network for pipeline defect diagnosis in a magnetic flux leakage detection system","authors":"Xiangkai Shen , Jinhai Liu , Yifu Ren , Lin Jiang , Lei Wang , He Zhao , Rui Li","doi":"10.1016/j.compind.2025.104290","DOIUrl":"10.1016/j.compind.2025.104290","url":null,"abstract":"<div><div>Defect diagnosis based on magnetic flux leakage (MFL) signals is an important process for assessing pipeline health, including defect detection and size quantification. However, existing studies suffer from poor consistency of results, because they regard defect detection and size quantification as separate tasks, lacking paradigm harmonization and interaction. In addition, the calibration of experts is required to achieve harmonization between the two, which increases the time cost of data analysis. To address the above challenges, our motivation is to synergistically learn two tasks within a unified framework and utilize their task properties for mutual benefit. Therefore, a novel defect diagnosis method based on a task-oriented physical collaborative network (TOPC-Net) is proposed, which is the first attempt at joint defect detection and size quantification in MFL inspection. First, a feature extraction subnetwork with a heterogeneous focus module is proposed to decompose initial task-specific features from shared spaces. Second, considering the strong correlation between the two tasks, a cross-task information awareness method is proposed to realize the information interaction between the two tasks, so that the task-specific features can be enhanced. Finally, a physical information-guided collaborative decision subnetwork is proposed to jointly optimize two tasks, where MFL domain knowledge is embedded into the subnetwork to provide expert guidance, ensuring the accuracy and stability of predictions. Experimental results show that the proposed method outperforms existing methods, with a detection accuracy of 96.0% and an average improvement of 7.5% in quantification accuracy, which makes it promising for industrial applications.</div></div>","PeriodicalId":55219,"journal":{"name":"Computers in Industry","volume":"169 ","pages":"Article 104290"},"PeriodicalIF":8.2,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143869732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}