Pub Date : 2025-11-18DOI: 10.1016/j.jmsy.2025.11.012
Jinhua Xiao , Bo Wang , Kaile Huang , Sergio Terzi , Wei Wang , Marco Macchi
The recycling of end-of-life (EOL) products poses significant challenges due to inefficient and unsafe disassembly processes. To address this, we propose a novel self-perception human-robot collaboration (HRC) system that enhances disassembly efficiency and safety through multi-modal human intention recognition. Our core methodological innovation lies in the real-time fusion of three key perception modules: action recognition using a Spatial-Temporal Graph Convolutional Network (ST-GCN), disassembly tool detection based on an enhanced YOLO algorithm, and facial angle recognition for operator awareness inference. A dedicated dataset for retired power battery disassembly was constructed to support this research, encompassing human skeletal data for action recognition, labeled images for tool detection, and facial expression detection. The proposed system was validated on a physical HRC disassembly platform. Experimental results demonstrate a marked improvement, with our integrated intention recognition method achieving an accuracy of approximately 85 %, significantly outperforming traditional single-modality approaches. Furthermore, the HRC disassembly operation was completed in 238 s, which is 60 s (or 20 %) faster than purely manual disassembly. This work provides a robust and efficient HRC disassembly framework for intelligent disassembly scenario understanding, contributing to advancing circular manufacturing.
{"title":"Intelligent disassembly scenario understanding for human behavior and intention recognition towards self-perception human-robot collaboration system","authors":"Jinhua Xiao , Bo Wang , Kaile Huang , Sergio Terzi , Wei Wang , Marco Macchi","doi":"10.1016/j.jmsy.2025.11.012","DOIUrl":"10.1016/j.jmsy.2025.11.012","url":null,"abstract":"<div><div>The recycling of end-of-life (EOL) products poses significant challenges due to inefficient and unsafe disassembly processes. To address this, we propose a novel self-perception human-robot collaboration (HRC) system that enhances disassembly efficiency and safety through multi-modal human intention recognition. Our core methodological innovation lies in the real-time fusion of three key perception modules: action recognition using a Spatial-Temporal Graph Convolutional Network (ST-GCN), disassembly tool detection based on an enhanced YOLO algorithm, and facial angle recognition for operator awareness inference. A dedicated dataset for retired power battery disassembly was constructed to support this research, encompassing human skeletal data for action recognition, labeled images for tool detection, and facial expression detection. The proposed system was validated on a physical HRC disassembly platform. Experimental results demonstrate a marked improvement, with our integrated intention recognition method achieving an accuracy of approximately 85 %, significantly outperforming traditional single-modality approaches. Furthermore, the HRC disassembly operation was completed in 238 s, which is 60 s (or 20 %) faster than purely manual disassembly. This work provides a robust and efficient HRC disassembly framework for intelligent disassembly scenario understanding, contributing to advancing circular manufacturing.</div></div>","PeriodicalId":16227,"journal":{"name":"Journal of Manufacturing Systems","volume":"83 ","pages":"Pages 937-962"},"PeriodicalIF":14.2,"publicationDate":"2025-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145568611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-18DOI: 10.1016/j.jmsy.2025.11.013
Hengqian Wang, Lei Chen, Kuangrong Hao, Bing Wei
In process manufacturing industries, existing semi-supervised soft sensor methods often exhibit compromised performance under low label rates due to poor generalization, underscoring the urgent need for extracting generalized and transferable features from unlabeled data. Recently, diffusion model-based self-supervised learning (SSL) has shown great potential in addressing this challenge by leveraging powerful generative mechanisms to capture complex data distributions and distill informative representations without extensive supervision. Motivated by this, this paper proposes a variational diffusion representation learning (VDRL) framework for semi-supervised soft sensing modeling under low label rates. First, we propose a self-supervised parametric terminal variational diffusion (PTVD) model for generalized feature extraction. By defining the forward process terminal of the diffusion model as parameterized learnable latent distributions and attempting to recover the original input samples in the reverse process, we successfully achieve a transformation in the role of the diffusion model from the original sample generator to the desired feature extractor. The designed pretext task of patch-level mask reconstruction further improves its capability on temporal data and allows the PTVD model to be trained in a self-supervised manner without the involvement of real labels. Subsequently, we propose a temporal-diffusion joint representation (TDJR) model for the prediction of quality variables based on the extracted generalized multi-granular features. In order to fully exploit the joint dynamic information of multi-granular features in different sequence dimensions, we innovatively extract joint representation information from both time and diffusion dimensions simultaneously and complementarily, and perform supervised training under limited label guidance. A series of experimental evaluations on two real industrial processes validate the framework’s effectiveness and stability in soft sensing applications.
{"title":"Variational diffusion representation learning framework for semi-supervised industrial soft sensing application","authors":"Hengqian Wang, Lei Chen, Kuangrong Hao, Bing Wei","doi":"10.1016/j.jmsy.2025.11.013","DOIUrl":"10.1016/j.jmsy.2025.11.013","url":null,"abstract":"<div><div>In process manufacturing industries, existing semi-supervised soft sensor methods often exhibit compromised performance under low label rates due to poor generalization, underscoring the urgent need for extracting generalized and transferable features from unlabeled data. Recently, diffusion model-based self-supervised learning (SSL) has shown great potential in addressing this challenge by leveraging powerful generative mechanisms to capture complex data distributions and distill informative representations without extensive supervision. Motivated by this, this paper proposes a variational diffusion representation learning (VDRL) framework for semi-supervised soft sensing modeling under low label rates. First, we propose a self-supervised parametric terminal variational diffusion (PTVD) model for generalized feature extraction. By defining the forward process terminal of the diffusion model as parameterized learnable latent distributions and attempting to recover the original input samples in the reverse process, we successfully achieve a transformation in the role of the diffusion model from the original sample generator to the desired feature extractor. The designed pretext task of patch-level mask reconstruction further improves its capability on temporal data and allows the PTVD model to be trained in a self-supervised manner without the involvement of real labels. Subsequently, we propose a temporal-diffusion joint representation (TDJR) model for the prediction of quality variables based on the extracted generalized multi-granular features. In order to fully exploit the joint dynamic information of multi-granular features in different sequence dimensions, we innovatively extract joint representation information from both time and diffusion dimensions simultaneously and complementarily, and perform supervised training under limited label guidance. A series of experimental evaluations on two real industrial processes validate the framework’s effectiveness and stability in soft sensing applications.</div></div>","PeriodicalId":16227,"journal":{"name":"Journal of Manufacturing Systems","volume":"83 ","pages":"Pages 923-936"},"PeriodicalIF":14.2,"publicationDate":"2025-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145568612","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-16DOI: 10.1016/j.jmsy.2025.11.006
Tianze Qiu , Yan Liu , Wenlei Xiao , Gang Zhao , Qiang Liu
With the advancement of intelligent manufacturing, CNC systems are increasingly expected to achieve higher levels of autonomous perception and decision-making. However, conventional CNC systems relying only on sensors have limited online analysis capability in complex machining. STEP-CNC, equipped with a simulation kernel, provides rich semantic information and enables multi-domain data fusion of “simulation + sensing”, offering a novel framework for online process analysis. Taking tool wear as a case study, this paper proposes a comprehensive online monitoring solution integrated into STEP-CNC. First, the geometric simulation is executed to calculate the Cutter Workpiece Engagement (CWE) and quantify instantaneous material removal, effectively characterizing interactions with workpiece. Second, a sparse stacked autoencoder extracts and compresses informative features from multi-sensor signals, yielding compact representations correlated with wear value. Third, an incremental prediction model tailored for online applications is developed, fusing geometric, physical, and process-domain inputs to provide precise wear increment estimates over fixed time windows. Finally, the prediction model is encapsulated as a service and integrated within the STEP-CNC framework, enabling tool wear monitoring with online feedback to the CNC controller. Experimental results demonstrate that the proposed method can accurately track tool wear progression, achieving an online monitoring accuracy exceeding 89%. The monitored wear values can further assist machining decision-making, preventing tool failures and ensuring workpiece quality. The proposed method may also serve as an actionable reference for using STEP-CNC with multi-domain data in intelligent manufacturing applications.
{"title":"STEP-compliant CNC system featuring real-time material removal simulation for online tool wear monitoring","authors":"Tianze Qiu , Yan Liu , Wenlei Xiao , Gang Zhao , Qiang Liu","doi":"10.1016/j.jmsy.2025.11.006","DOIUrl":"10.1016/j.jmsy.2025.11.006","url":null,"abstract":"<div><div>With the advancement of intelligent manufacturing, CNC systems are increasingly expected to achieve higher levels of autonomous perception and decision-making. However, conventional CNC systems relying only on sensors have limited online analysis capability in complex machining. STEP-CNC, equipped with a simulation kernel, provides rich semantic information and enables multi-domain data fusion of “simulation + sensing”, offering a novel framework for online process analysis. Taking tool wear as a case study, this paper proposes a comprehensive online monitoring solution integrated into STEP-CNC. First, the geometric simulation is executed to calculate the Cutter Workpiece Engagement (CWE) and quantify instantaneous material removal, effectively characterizing interactions with workpiece. Second, a sparse stacked autoencoder extracts and compresses informative features from multi-sensor signals, yielding compact representations correlated with wear value. Third, an incremental prediction model tailored for online applications is developed, fusing geometric, physical, and process-domain inputs to provide precise wear increment estimates over fixed time windows. Finally, the prediction model is encapsulated as a service and integrated within the STEP-CNC framework, enabling tool wear monitoring with online feedback to the CNC controller. Experimental results demonstrate that the proposed method can accurately track tool wear progression, achieving an online monitoring accuracy exceeding 89%. The monitored wear values can further assist machining decision-making, preventing tool failures and ensuring workpiece quality. The proposed method may also serve as an actionable reference for using STEP-CNC with multi-domain data in intelligent manufacturing applications.</div></div>","PeriodicalId":16227,"journal":{"name":"Journal of Manufacturing Systems","volume":"83 ","pages":"Pages 904-922"},"PeriodicalIF":14.2,"publicationDate":"2025-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145568613","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-15DOI: 10.1016/j.jmsy.2025.11.008
Yu Liu , Wen Peng , Xuemeng Li , Wenteng Wu , Bingquan Zhu , Xudong Li , Yafeng Ji , Dianhua Zhang , Jie Sun
Strip crown is a critical quality indicator in hot-rolled strip, as inaccurate prediction often leads to downstream flatness defects and reduced production efficiency. Traditional physical models and individual machine learning (ML) algorithms struggle to achieve stable and reliable performance under the noisy and imbalanced conditions of industrial production. To address this challenge, this study proposes an adaptive hybrid prediction framework (A-HPF) based on Dempster-Shafer evidence theory. The framework integrates multiple ML, incorporates the support vector machine synthetic minority over-sampling technique to mitigate class imbalance, and introduces an adaptive conflict resolution mechanism with fuzzy affiliation functions to dynamically optimize evidence fusion. To ensure transparency and trustworthiness, SHapley Additive exPlanations combined with a causal network is employed to interpret the A-HPF’s decision logic and trace potential root causes. Validated on a 2160 mm hot rolling line, the proposed method improves classification accuracy by 5.9–15.6 %, reduces misclassification in edge categories by 63.6 %, and achieves an overall accuracy exceeding 0.93 across eight major steel grades. These improvements are further embedded into an industrial digital platform, enabling real-time quality prediction, interpretable analysis, and decision support, providing a practical solution for intelligent quality management in hot strip rolling.
{"title":"An interpretable industrial digital platform based on Dempster-Shafer theory for pre-diagnosis the quality of hot-rolled strip","authors":"Yu Liu , Wen Peng , Xuemeng Li , Wenteng Wu , Bingquan Zhu , Xudong Li , Yafeng Ji , Dianhua Zhang , Jie Sun","doi":"10.1016/j.jmsy.2025.11.008","DOIUrl":"10.1016/j.jmsy.2025.11.008","url":null,"abstract":"<div><div>Strip crown is a critical quality indicator in hot-rolled strip, as inaccurate prediction often leads to downstream flatness defects and reduced production efficiency. Traditional physical models and individual machine learning (ML) algorithms struggle to achieve stable and reliable performance under the noisy and imbalanced conditions of industrial production. To address this challenge, this study proposes an adaptive hybrid prediction framework (A-HPF) based on Dempster-Shafer evidence theory. The framework integrates multiple ML, incorporates the support vector machine synthetic minority over-sampling technique to mitigate class imbalance, and introduces an adaptive conflict resolution mechanism with fuzzy affiliation functions to dynamically optimize evidence fusion. To ensure transparency and trustworthiness, SHapley Additive exPlanations combined with a causal network is employed to interpret the A-HPF’s decision logic and trace potential root causes. Validated on a 2160 mm hot rolling line, the proposed method improves classification accuracy by 5.9–15.6 %, reduces misclassification in edge categories by 63.6 %, and achieves an overall accuracy exceeding 0.93 across eight major steel grades. These improvements are further embedded into an industrial digital platform, enabling real-time quality prediction, interpretable analysis, and decision support, providing a practical solution for intelligent quality management in hot strip rolling.</div></div>","PeriodicalId":16227,"journal":{"name":"Journal of Manufacturing Systems","volume":"83 ","pages":"Pages 878-903"},"PeriodicalIF":14.2,"publicationDate":"2025-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145516967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-14DOI: 10.1016/j.jmsy.2025.11.009
Ziling Wang , Lai Zou , Wenxi Wang , A.Y.C. Nee , S.K. Ong
The robotic multi-region grinding method is used to machine curved workpieces with uneven surface allowance to further improve their profile accuracy. However, in this method, when the robot executes toolpaths across adjacent grinding regions, force changes and non-uniform grinding often occur. To address the problem, a novel cross-region robotic grinding method that incorporates force control and toolpath planning across adjacent regions is proposed. In this method, a toolpath planning method, with consideration of multiple grinding regions with different expected grinding forces, is developed to generate cutter-contact (CC) points in each toolpath curve based on the point clouds of workpieces, and ensure that there are CC points near the boundary between two adjacent grinding regions. Furthermore, a novel model predictive control scheme with an environment observer is designed to track the grinding force in a single grinding region. In addition, the adaptive impedance model with a novel adaptive update rate is introduced into the control scheme to reduce the changes in the grinding force along the toolpaths across two adjacent regions. Robotic grinding experiments are conducted to verify the superiority of the proposed grinding method. The surface accuracy of the curved workpiece is improved by some 26 %.
{"title":"Cross-region robotic grinding with adaptive toolpath planning and force control for point clouds of complex curved workpieces","authors":"Ziling Wang , Lai Zou , Wenxi Wang , A.Y.C. Nee , S.K. Ong","doi":"10.1016/j.jmsy.2025.11.009","DOIUrl":"10.1016/j.jmsy.2025.11.009","url":null,"abstract":"<div><div>The robotic multi-region grinding method is used to machine curved workpieces with uneven surface allowance to further improve their profile accuracy. However, in this method, when the robot executes toolpaths across adjacent grinding regions, force changes and non-uniform grinding often occur. To address the problem, a novel cross-region robotic grinding method that incorporates force control and toolpath planning across adjacent regions is proposed. In this method, a toolpath planning method, with consideration of multiple grinding regions with different expected grinding forces, is developed to generate cutter-contact (CC) points in each toolpath curve based on the point clouds of workpieces, and ensure that there are CC points near the boundary between two adjacent grinding regions. Furthermore, a novel model predictive control scheme with an environment observer is designed to track the grinding force in a single grinding region. In addition, the adaptive impedance model with a novel adaptive update rate is introduced into the control scheme to reduce the changes in the grinding force along the toolpaths across two adjacent regions. Robotic grinding experiments are conducted to verify the superiority of the proposed grinding method. The surface accuracy of the curved workpiece is improved by some 26 %.</div></div>","PeriodicalId":16227,"journal":{"name":"Journal of Manufacturing Systems","volume":"83 ","pages":"Pages 856-877"},"PeriodicalIF":14.2,"publicationDate":"2025-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145516968","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-12DOI: 10.1016/j.jmsy.2025.11.007
Junhyeong Lee , Joon-Young Kim , Heekyu Kim , Inhyo Lee , Seunghwa Ryu
The injection molding industry faces critical challenges in preserving and transferring field knowledge, particularly as experienced workers retire and multilingual barriers hinder effective communication. This study introduces IM-Chat, a multi-agent framework based on large language models (LLMs), designed to facilitate knowledge transfer in injection molding. IM-Chat integrates both limited documented knowledge (e.g., troubleshooting tables, manuals) and extensive field data modeled through a data-driven process condition generator that infers optimal manufacturing settings from environmental inputs such as temperature and humidity, enabling robust and context-aware task resolution. By adopting a retrieval-augmented generation (RAG) strategy and tool-calling agents within a modular architecture, IM-Chat ensures adaptability without the need for fine-tuning. Performance was assessed across 100 single-tool and 60 hybrid tasks for GPT-4o, GPT-4o-mini, and GPT-3.5-turbo by domain experts using a 10-point rubric focused on relevance and correctness, and was further supplemented by automated evaluation using GPT-4o guided by a domain-adapted instruction prompt. The evaluation results indicate that more capable models tend to achieve higher accuracy, particularly in complex, tool-integrated scenarios. In addition, compared with the fine-tuned single-agent LLM, IM-Chat demonstrated superior accuracy, particularly in quantitative reasoning, and greater scalability in handling multiple information sources. Overall, these findings demonstrate the viability of multi-agent LLM systems for industrial knowledge workflows and establish IM-Chat as a scalable and generalizable approach to AI-assisted decision support in manufacturing. To support reproducibility and practical adoption, supplementary materials including prompts, evaluation data, and video demonstrations are made available.
{"title":"IM-Chat: A multi-agent LLM framework integrating tool-calling and diffusion modeling for knowledge transfer in injection molding industry","authors":"Junhyeong Lee , Joon-Young Kim , Heekyu Kim , Inhyo Lee , Seunghwa Ryu","doi":"10.1016/j.jmsy.2025.11.007","DOIUrl":"10.1016/j.jmsy.2025.11.007","url":null,"abstract":"<div><div>The injection molding industry faces critical challenges in preserving and transferring field knowledge, particularly as experienced workers retire and multilingual barriers hinder effective communication. This study introduces IM-Chat, a multi-agent framework based on large language models (LLMs), designed to facilitate knowledge transfer in injection molding. IM-Chat integrates both limited documented knowledge (e.g., troubleshooting tables, manuals) and extensive field data modeled through a data-driven process condition generator that infers optimal manufacturing settings from environmental inputs such as temperature and humidity, enabling robust and context-aware task resolution. By adopting a retrieval-augmented generation (RAG) strategy and tool-calling agents within a modular architecture, IM-Chat ensures adaptability without the need for fine-tuning. Performance was assessed across 100 single-tool and 60 hybrid tasks for GPT-4o, GPT-4o-mini, and GPT-3.5-turbo by domain experts using a 10-point rubric focused on relevance and correctness, and was further supplemented by automated evaluation using GPT-4o guided by a domain-adapted instruction prompt. The evaluation results indicate that more capable models tend to achieve higher accuracy, particularly in complex, tool-integrated scenarios. In addition, compared with the fine-tuned single-agent LLM, IM-Chat demonstrated superior accuracy, particularly in quantitative reasoning, and greater scalability in handling multiple information sources. Overall, these findings demonstrate the viability of multi-agent LLM systems for industrial knowledge workflows and establish IM-Chat as a scalable and generalizable approach to AI-assisted decision support in manufacturing. To support reproducibility and practical adoption, <span><span>supplementary materials</span></span> including prompts, evaluation data, and video demonstrations are made available.</div></div>","PeriodicalId":16227,"journal":{"name":"Journal of Manufacturing Systems","volume":"83 ","pages":"Pages 839-855"},"PeriodicalIF":14.2,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145516966","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-12DOI: 10.1016/j.jmsy.2025.10.014
Yanying Wang , Zhiheng Zhao , Yujie Han , Ying Cheng , George Q. Huang
Fast fashion platforms such as SHEIN coordinate thousands of small, nearby garment factories to fulfil large numbers of small-lot, fast-switch orders. Dispatching each order operation to the most suitable factory while respecting process routes creates a Flexible Job Shop Scheduling Problem with sequence-dependent setup times (FJSP-SDST), which is NP-hard and must be solved repeatedly as new batches arrive. Current end-to-end deep reinforcement learning (DRL) schedulers either reschedule every batch from scratch or disregard setup costs, which undermines efficiency. We introduce a Cyber-Physical Internet (CPI) scheduling framework that provides routers for every factory and the company, enabling them to cache solved schedules and broadcast real time factory states. This approach skips redundant computations and supplies fresh setup time data. Within this framework, we have developed a Hierarchical Attention Network based Reinforcement Learning (HANRL) scheduler to model the interactions between orders and factories, as well as factory competition and setup costs. Experiments on synthetic and public benchmarks demonstrate that HANRL reduces makespan and improves generalization over state-of-the-art DRL baselines, all while retaining sub-second decision times. This proves the suitability of HANRL for large scale social manufacturing environments.
{"title":"Cyber-physical internet enabled hierarchical attention network based reinforcement learning for order dispatch in fast fashion manufacturing","authors":"Yanying Wang , Zhiheng Zhao , Yujie Han , Ying Cheng , George Q. Huang","doi":"10.1016/j.jmsy.2025.10.014","DOIUrl":"10.1016/j.jmsy.2025.10.014","url":null,"abstract":"<div><div>Fast fashion platforms such as SHEIN coordinate thousands of small, nearby garment factories to fulfil large numbers of small-lot, fast-switch orders. Dispatching each order operation to the most suitable factory while respecting process routes creates a Flexible Job Shop Scheduling Problem with sequence-dependent setup times (FJSP-SDST), which is NP-hard and must be solved repeatedly as new batches arrive. Current end-to-end deep reinforcement learning (DRL) schedulers either reschedule every batch from scratch or disregard setup costs, which undermines efficiency. We introduce a Cyber-Physical Internet (CPI) scheduling framework that provides routers for every factory and the company, enabling them to cache solved schedules and broadcast real time factory states. This approach skips redundant computations and supplies fresh setup time data. Within this framework, we have developed a Hierarchical Attention Network based Reinforcement Learning (HANRL) scheduler to model the interactions between orders and factories, as well as factory competition and setup costs. Experiments on synthetic and public benchmarks demonstrate that HANRL reduces makespan and improves generalization over state-of-the-art DRL baselines, all while retaining sub-second decision times. This proves the suitability of HANRL for large scale social manufacturing environments.</div></div>","PeriodicalId":16227,"journal":{"name":"Journal of Manufacturing Systems","volume":"83 ","pages":"Pages 784-799"},"PeriodicalIF":14.2,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145516964","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-12DOI: 10.1016/j.jmsy.2025.11.005
Hongquan Gui , Zhanpeng Yang , Arjun Rachana Harish , Cheng Ren , Yishu Yang , Ming Li
Customized garment production is hindered by the expert-dependent nature of sewing pattern generation—a skill-intensive process requiring years of training. While recent approaches aim to translate user intent into sewing patterns, they often struggle to interpret multimodal inputs such as text and images. Multimodal large language models (MLLMs) offer a promising path forward, as they can naturally understand diverse user intents. Yet, applying MLLMs to sewing pattern generation is challenging because conventional tokenization methods often lose the structural information of sewing patterns. To address this issue, we propose GenPattern, a novel framework that integrates structured graph modeling with MLLMs to enable more accurate sewing pattern generation. We introduce a scalable vector graphics (SVG)-style pattern tokenizer, which encodes sewing patterns into structured token sequences. Furthermore, we present SewGraphFuser, a dual-graph module that explicitly models geometric and semantic dependencies to inject structural information into MLLMs. This module combines a structure graph convolution module and a sequence graph convolution module to jointly capture multi-scale spatial and sequential features via a geometric consistency graph and a semantic dependency graph. Finally, to bridge the gap between digital design and physical fabrication, our framework drives a human-robot collaborative cutting platform, enabling expert-free, on-demand garment customization. This innovation empowers human-robot collaboration in pattern production, enhancing scalability in real-world manufacturing. Experimental results show that GenPattern achieves 86.7 % stitch accuracy and reduces panel vertex L2 error to 2.9 cm, demonstrating its potential to democratize custom fashion by enabling non-experts to reliably produce physical garments directly from their ideas.
{"title":"GenPattern: dual-graph enhanced sewing pattern generation via multimodal large language model","authors":"Hongquan Gui , Zhanpeng Yang , Arjun Rachana Harish , Cheng Ren , Yishu Yang , Ming Li","doi":"10.1016/j.jmsy.2025.11.005","DOIUrl":"10.1016/j.jmsy.2025.11.005","url":null,"abstract":"<div><div>Customized garment production is hindered by the expert-dependent nature of sewing pattern generation—a skill-intensive process requiring years of training. While recent approaches aim to translate user intent into sewing patterns, they often struggle to interpret multimodal inputs such as text and images. Multimodal large language models (MLLMs) offer a promising path forward, as they can naturally understand diverse user intents. Yet, applying MLLMs to sewing pattern generation is challenging because conventional tokenization methods often lose the structural information of sewing patterns. To address this issue, we propose GenPattern, a novel framework that integrates structured graph modeling with MLLMs to enable more accurate sewing pattern generation. We introduce a scalable vector graphics (SVG)-style pattern tokenizer, which encodes sewing patterns into structured token sequences. Furthermore, we present SewGraphFuser, a dual-graph module that explicitly models geometric and semantic dependencies to inject structural information into MLLMs. This module combines a structure graph convolution module and a sequence graph convolution module to jointly capture multi-scale spatial and sequential features via a geometric consistency graph and a semantic dependency graph. Finally, to bridge the gap between digital design and physical fabrication, our framework drives a human-robot collaborative cutting platform, enabling expert-free, on-demand garment customization. This innovation empowers human-robot collaboration in pattern production, enhancing scalability in real-world manufacturing. Experimental results show that GenPattern achieves 86.7 % stitch accuracy and reduces panel vertex L2 error to 2.9 cm, demonstrating its potential to democratize custom fashion by enabling non-experts to reliably produce physical garments directly from their ideas.</div></div>","PeriodicalId":16227,"journal":{"name":"Journal of Manufacturing Systems","volume":"83 ","pages":"Pages 822-838"},"PeriodicalIF":14.2,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145516965","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-11DOI: 10.1016/j.jmsy.2025.11.010
José Joaquín Peralta Abadía , Fabio Marco Monetti , Sylvia Nathaly Rea Minango , Angela Carrera-Rivera , Miriam Ugarte Querejeta , Mikel Cuesta Zabaljauregui , Felix Larrinaga Barrenechea , Miren Illarramendi Rezabal , Antonio Maffei
Self-diagnosis functionalities, as integral components of advanced manufacturing services within cyber–physical systems (CPSs), are made possible through cloud computing technologies and machine learning techniques. These services play a crucial role in enhancing the autonomy of CPSs and introducing cost-efficient and scalable solutions. Despite the promising outlook, a gap exists in the literature regarding the lack of clear architectural frameworks and requirements for implementing self-diagnosis services in industrial settings. This paper addresses this gap by presenting a comprehensive requirement set and developing a high-level architecture tailored for self-diagnosis services. The proposed approach is validated through a detailed case study of a cloud-based self-diagnosis service, demonstrating alignment with the established architecture and requirements. The anticipated outcome of this research is to offer concrete implementation guidelines to support researchers, engineers, and practitioners in deploying CPS-based self-diagnosis services and improving production processes and system performance.
{"title":"Self-diagnosis service to support analysis of production performance, monitoring and optimisation activities","authors":"José Joaquín Peralta Abadía , Fabio Marco Monetti , Sylvia Nathaly Rea Minango , Angela Carrera-Rivera , Miriam Ugarte Querejeta , Mikel Cuesta Zabaljauregui , Felix Larrinaga Barrenechea , Miren Illarramendi Rezabal , Antonio Maffei","doi":"10.1016/j.jmsy.2025.11.010","DOIUrl":"10.1016/j.jmsy.2025.11.010","url":null,"abstract":"<div><div>Self-diagnosis functionalities, as integral components of advanced manufacturing services within cyber–physical systems (CPSs), are made possible through cloud computing technologies and machine learning techniques. These services play a crucial role in enhancing the autonomy of CPSs and introducing cost-efficient and scalable solutions. Despite the promising outlook, a gap exists in the literature regarding the lack of clear architectural frameworks and requirements for implementing self-diagnosis services in industrial settings. This paper addresses this gap by presenting a comprehensive requirement set and developing a high-level architecture tailored for self-diagnosis services. The proposed approach is validated through a detailed case study of a cloud-based self-diagnosis service, demonstrating alignment with the established architecture and requirements. The anticipated outcome of this research is to offer concrete implementation guidelines to support researchers, engineers, and practitioners in deploying CPS-based self-diagnosis services and improving production processes and system performance.</div></div>","PeriodicalId":16227,"journal":{"name":"Journal of Manufacturing Systems","volume":"83 ","pages":"Pages 800-821"},"PeriodicalIF":14.2,"publicationDate":"2025-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145516962","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-10DOI: 10.1016/j.jmsy.2025.11.003
Yanghao Wu, Paul D. Wilcox, Anthony J. Croxford
Fasteners are widely used in mechanical structures, where stress concentrations around fastener holes can lead to crack initiation and fatigue failures. In the aerospace industry, routine fastener hole inspections are critical to ensure structural integrity. Ultrasonic testing is one of the main inspection approaches. Conventionally, it involves a single-element probe that must be manually placed at multiple locations and orientations so that the ultrasound beam insonifies the area around the hole from different angles. The received ultrasonic time-domain signals at each location are analyzed, which is time-consuming, operator-dependent, and prone to inconsistencies. 2D ultrasonic array probes enable 3D volumetric images of fastener hole defects to be obtained from a single probe position, offering the potential for more efficient automated inspection and data interpretation. To achieve this, the 2D array probe must be accurately located over the centre of the hole and the ultrasonic coupling with the component must be consistent over the entire probe contact surface. This paper presents an automated robotic system for ultrasonic fastener hole inspection, that is designed to address these issues. A 7 degree-of-freedom robot arm is used with a vision module, and a customized probe adapter integrates a 2D ultrasonic array and coupling block into the robot end effector. A novel hybrid probe manipulation method is proposed, which combines camera-based visual localization with real-time ultrasound signal feedback to ensure accurate probe alignment and consistent coupling. The whole inspection workflow is scheduled and a graphical user interface is developed to demonstrate this automatic inspection. Experimental validation demonstrates that the robotic system performs accurate, repeatable inspections, significantly enhancing efficiency and reliability compared to manual techniques. The proposed approach addresses key challenges in robotic ultrasonic inspection and offers a scalable solution for intelligent maintenance in aerospace and other high-reliability industries.
{"title":"Robotic inspection of fastener holes with hybrid visual and ultrasonic motion control","authors":"Yanghao Wu, Paul D. Wilcox, Anthony J. Croxford","doi":"10.1016/j.jmsy.2025.11.003","DOIUrl":"10.1016/j.jmsy.2025.11.003","url":null,"abstract":"<div><div>Fasteners are widely used in mechanical structures, where stress concentrations around fastener holes can lead to crack initiation and fatigue failures. In the aerospace industry, routine fastener hole inspections are critical to ensure structural integrity. Ultrasonic testing is one of the main inspection approaches. Conventionally, it involves a single-element probe that must be manually placed at multiple locations and orientations so that the ultrasound beam insonifies the area around the hole from different angles. The received ultrasonic time-domain signals at each location are analyzed, which is time-consuming, operator-dependent, and prone to inconsistencies. 2D ultrasonic array probes enable 3D volumetric images of fastener hole defects to be obtained from a single probe position, offering the potential for more efficient automated inspection and data interpretation. To achieve this, the 2D array probe must be accurately located over the centre of the hole and the ultrasonic coupling with the component must be consistent over the entire probe contact surface. This paper presents an automated robotic system for ultrasonic fastener hole inspection, that is designed to address these issues. A 7 degree-of-freedom robot arm is used with a vision module, and a customized probe adapter integrates a 2D ultrasonic array and coupling block into the robot end effector. A novel hybrid probe manipulation method is proposed, which combines camera-based visual localization with real-time ultrasound signal feedback to ensure accurate probe alignment and consistent coupling. The whole inspection workflow is scheduled and a graphical user interface is developed to demonstrate this automatic inspection. Experimental validation demonstrates that the robotic system performs accurate, repeatable inspections, significantly enhancing efficiency and reliability compared to manual techniques. The proposed approach addresses key challenges in robotic ultrasonic inspection and offers a scalable solution for intelligent maintenance in aerospace and other high-reliability industries.</div></div>","PeriodicalId":16227,"journal":{"name":"Journal of Manufacturing Systems","volume":"83 ","pages":"Pages 770-783"},"PeriodicalIF":14.2,"publicationDate":"2025-11-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145516963","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}