A skeleton-based assembly action recognition method with feature fusion for human-robot collaborative assembly

IF 12.2 1区 工程技术 Q1 ENGINEERING, INDUSTRIAL Journal of Manufacturing Systems Pub Date : 2024-08-30 DOI:10.1016/j.jmsy.2024.08.019
Daxin Liu , Yu Huang , Zhenyu Liu , Haoyang Mao , Pengcheng Kan , Jianrong Tan
{"title":"A skeleton-based assembly action recognition method with feature fusion for human-robot collaborative assembly","authors":"Daxin Liu ,&nbsp;Yu Huang ,&nbsp;Zhenyu Liu ,&nbsp;Haoyang Mao ,&nbsp;Pengcheng Kan ,&nbsp;Jianrong Tan","doi":"10.1016/j.jmsy.2024.08.019","DOIUrl":null,"url":null,"abstract":"<div><p>Human-robot collaborative assembly (HRCA) is one of the current trends of intelligent manufacturing, and assembly action recognition is the basis of and the key to HRCA. A multi-scale and multi-stream graph convolutional network (2MSGCN) for assembly action recognition is proposed in this paper. 2MSGCN takes the temporal skeleton sample as input and outputs the class of the assembly action to which the sample belongs. RGBD images of the operator performing the assembly actions are captured by three RGBD cameras mounted at different viewpoints and pre-processed to generate the complete human skeleton. A multi-scale and multi-stream (2MS) mechanism and a feature fusion mechanism are proposed to improve the recognition accuracy of 2MSGCN. The 2MS mechanism is designed to input the skeleton data to 2MSGCN in the form of a joint stream, a bone stream and a motion stream, while the joint stream further generates two sets of input with rough scales to represent features in higher dimensional human skeleton, which obtains information of different scales and streams in temporal skeleton samples. And the feature fusion mechanism enables the fused feature to retain the information of the sub-feature while incorporating union information between the sub-features. Also, the improved convolution operation based on Ghost module is introduced to the 2MSGCN to reduce the number of the parameters and floating-point operations (FLOPs) and improve the real-time performance. Considering that there will be transitional actions when the operator switches between assembly actions in the continuous assembly process, a transitional action classification (TAC) method is proposed to distinguish the transitional actions from the assembly actions. Experiments on the public dataset NTU RGB+D 60 (NTU 60) and a self-built assembly action dataset indicate that the proposed 2MSGCN outperforms the mainstream models in recognition accuracy and real-time performance.</p></div>","PeriodicalId":16227,"journal":{"name":"Journal of Manufacturing Systems","volume":"76 ","pages":"Pages 553-566"},"PeriodicalIF":12.2000,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Manufacturing Systems","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0278612524001821","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, INDUSTRIAL","Score":null,"Total":0}
引用次数: 0

Abstract

Human-robot collaborative assembly (HRCA) is one of the current trends of intelligent manufacturing, and assembly action recognition is the basis of and the key to HRCA. A multi-scale and multi-stream graph convolutional network (2MSGCN) for assembly action recognition is proposed in this paper. 2MSGCN takes the temporal skeleton sample as input and outputs the class of the assembly action to which the sample belongs. RGBD images of the operator performing the assembly actions are captured by three RGBD cameras mounted at different viewpoints and pre-processed to generate the complete human skeleton. A multi-scale and multi-stream (2MS) mechanism and a feature fusion mechanism are proposed to improve the recognition accuracy of 2MSGCN. The 2MS mechanism is designed to input the skeleton data to 2MSGCN in the form of a joint stream, a bone stream and a motion stream, while the joint stream further generates two sets of input with rough scales to represent features in higher dimensional human skeleton, which obtains information of different scales and streams in temporal skeleton samples. And the feature fusion mechanism enables the fused feature to retain the information of the sub-feature while incorporating union information between the sub-features. Also, the improved convolution operation based on Ghost module is introduced to the 2MSGCN to reduce the number of the parameters and floating-point operations (FLOPs) and improve the real-time performance. Considering that there will be transitional actions when the operator switches between assembly actions in the continuous assembly process, a transitional action classification (TAC) method is proposed to distinguish the transitional actions from the assembly actions. Experiments on the public dataset NTU RGB+D 60 (NTU 60) and a self-built assembly action dataset indicate that the proposed 2MSGCN outperforms the mainstream models in recognition accuracy and real-time performance.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于骨架的装配动作识别方法与人机协作装配的特征融合
人机协同装配(HRCA)是当前智能制造的发展趋势之一,而装配动作识别是人机协同装配的基础和关键。本文提出了一种用于装配动作识别的多尺度、多流图卷积网络(2MSGCN)。2MSGCN 将时间骨架样本作为输入,并输出样本所属的装配动作类别。操作员执行装配动作的 RGBD 图像由安装在不同视点的三台 RGBD 摄像机拍摄,并经过预处理生成完整的人体骨架。为了提高 2MSGCN 的识别准确率,我们提出了一种多尺度、多流(2MS)机制和一种特征融合机制。2MS 机制的设计是将骨架数据以关节流、骨骼流和运动流的形式输入到 2MSGCN 中,而关节流则进一步生成两组具有粗略尺度的输入,以表示高维人体骨架中的特征,从而获得时空骨架样本中不同尺度和流的信息。而特征融合机制使融合后的特征既保留了子特征的信息,又纳入了子特征之间的联合信息。此外,2MSGCN 还引入了基于 Ghost 模块的改进卷积运算,以减少参数和浮点运算次数(FLOP),提高实时性。考虑到在连续装配过程中,操作员在装配动作之间切换时会存在过渡动作,因此提出了一种过渡动作分类(TAC)方法来区分过渡动作和装配动作。在公开数据集 NTU RGB+D 60(NTU 60)和自建的装配动作数据集上的实验表明,所提出的 2MSGCN 在识别准确率和实时性方面优于主流模型。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Journal of Manufacturing Systems
Journal of Manufacturing Systems 工程技术-工程:工业
CiteScore
23.30
自引率
13.20%
发文量
216
审稿时长
25 days
期刊介绍: The Journal of Manufacturing Systems is dedicated to showcasing cutting-edge fundamental and applied research in manufacturing at the systems level. Encompassing products, equipment, people, information, control, and support functions, manufacturing systems play a pivotal role in the economical and competitive development, production, delivery, and total lifecycle of products, meeting market and societal needs. With a commitment to publishing archival scholarly literature, the journal strives to advance the state of the art in manufacturing systems and foster innovation in crafting efficient, robust, and sustainable manufacturing systems. The focus extends from equipment-level considerations to the broader scope of the extended enterprise. The Journal welcomes research addressing challenges across various scales, including nano, micro, and macro-scale manufacturing, and spanning diverse sectors such as aerospace, automotive, energy, and medical device manufacturing.
期刊最新文献
Material removal rate optimization with bayesian optimized differential evolution based on deep learning in robotic polishing Leveraging AI for energy-efficient manufacturing systems: Review and future prospectives Investigation of assistance systems in assembly in the context of digitalization: A systematic literature review Machining parameter optimization for a batch milling system using multi-task deep reinforcement learning A dynamic artificial bee colony for fuzzy distributed energy-efficient hybrid flow shop scheduling with batch processing machines
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1