AEGLR-Net:用于准确检测车身表面缺陷的注意力增强型全局-局部精细网络

IF 9.1 1区 计算机科学 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Robotics and Computer-integrated Manufacturing Pub Date : 2024-06-17 DOI:10.1016/j.rcim.2024.102806
Yike He , Baotong Wu , Xiao Liu , Baicun Wang , Jianzhong Fu , Songyu Hu
{"title":"AEGLR-Net:用于准确检测车身表面缺陷的注意力增强型全局-局部精细网络","authors":"Yike He ,&nbsp;Baotong Wu ,&nbsp;Xiao Liu ,&nbsp;Baicun Wang ,&nbsp;Jianzhong Fu ,&nbsp;Songyu Hu","doi":"10.1016/j.rcim.2024.102806","DOIUrl":null,"url":null,"abstract":"<div><p>The complex background on the car body surface, such as the orange peel-like texture and shiny metallic powder, poses a considerable challenge to automated defect detection. Two mainstream methods are currently used to tackle this challenge: global information-based and attention mechanism-based methods. However, these methods lack the capability to integrate valuable global-to-local information and explore deeper distinguishable features, thereby affecting the overall detection performance. To address this issue, we propose a novel attention enhanced global–local refined detection network (AEGLR-Net), which can perform effective global-to-local refined feature extraction and fusion. First, we design an adaptive Transformer–CNN tandem backbone (ATCT-backbone) to dynamically aware valuable global information and integrate local details to comprehensively extract specific features between defects and complex backgrounds. Then, we propose a novel refined cross-dimensional aggregation (RCDA) attention to facilitate the point-to-point interaction of multidimensional information, effectively emphasizing the representation of deeper discriminative defect features. Finally, we construct an attention-embedded flexible feature pyramid network (AE-FFPN), which incorporates the RCDA attention to guide the feature pyramid network in targeted feature fusion, thereby enhancing the efficiency of feature fusion in the detection model. Extensive comparative experiments demonstrate that the AEGLR-Net outperforms state-of-the-art approaches, attaining exceptional performance with 89.2 % mAP (mean average precision) and 85.5 FPS (frames per second).</p></div>","PeriodicalId":21452,"journal":{"name":"Robotics and Computer-integrated Manufacturing","volume":"90 ","pages":"Article 102806"},"PeriodicalIF":9.1000,"publicationDate":"2024-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"AEGLR-Net: Attention enhanced global–local refined network for accurate detection of car body surface defects\",\"authors\":\"Yike He ,&nbsp;Baotong Wu ,&nbsp;Xiao Liu ,&nbsp;Baicun Wang ,&nbsp;Jianzhong Fu ,&nbsp;Songyu Hu\",\"doi\":\"10.1016/j.rcim.2024.102806\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>The complex background on the car body surface, such as the orange peel-like texture and shiny metallic powder, poses a considerable challenge to automated defect detection. Two mainstream methods are currently used to tackle this challenge: global information-based and attention mechanism-based methods. However, these methods lack the capability to integrate valuable global-to-local information and explore deeper distinguishable features, thereby affecting the overall detection performance. To address this issue, we propose a novel attention enhanced global–local refined detection network (AEGLR-Net), which can perform effective global-to-local refined feature extraction and fusion. First, we design an adaptive Transformer–CNN tandem backbone (ATCT-backbone) to dynamically aware valuable global information and integrate local details to comprehensively extract specific features between defects and complex backgrounds. Then, we propose a novel refined cross-dimensional aggregation (RCDA) attention to facilitate the point-to-point interaction of multidimensional information, effectively emphasizing the representation of deeper discriminative defect features. Finally, we construct an attention-embedded flexible feature pyramid network (AE-FFPN), which incorporates the RCDA attention to guide the feature pyramid network in targeted feature fusion, thereby enhancing the efficiency of feature fusion in the detection model. Extensive comparative experiments demonstrate that the AEGLR-Net outperforms state-of-the-art approaches, attaining exceptional performance with 89.2 % mAP (mean average precision) and 85.5 FPS (frames per second).</p></div>\",\"PeriodicalId\":21452,\"journal\":{\"name\":\"Robotics and Computer-integrated Manufacturing\",\"volume\":\"90 \",\"pages\":\"Article 102806\"},\"PeriodicalIF\":9.1000,\"publicationDate\":\"2024-06-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Robotics and Computer-integrated Manufacturing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0736584524000930\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Robotics and Computer-integrated Manufacturing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0736584524000930","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 0

摘要

车身表面复杂的背景,如桔皮状纹理和闪亮的金属粉末,给自动缺陷检测带来了相当大的挑战。目前有两种主流方法来应对这一挑战:基于全局信息的方法和基于注意机制的方法。然而,这些方法缺乏整合有价值的全局到局部信息和探索更深层次可区分特征的能力,从而影响了整体检测性能。针对这一问题,我们提出了一种新型注意力增强型全局-局部精细检测网络(AEGLR-Net),它能有效地进行全局-局部精细特征提取和融合。首先,我们设计了一个自适应变换器-CNN 串联骨干网(ATCT-backbone),以动态感知有价值的全局信息并整合局部细节,从而全面提取缺陷和复杂背景之间的特定特征。然后,我们提出了一种新颖的精细跨维聚合(RCDA)注意力,以促进多维信息的点对点交互,有效地强调了更深层次的缺陷判别特征的表示。最后,我们构建了一种嵌入注意力的柔性特征金字塔网络(AE-FPN),它结合了 RCDA 注意力,引导特征金字塔网络进行有针对性的特征融合,从而提高了检测模型中特征融合的效率。广泛的对比实验证明,AEGLR-Net 的性能优于最先进的方法,达到了 89.2 % mAP(平均精度)和 85.5 FPS(每秒帧数)的卓越性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
AEGLR-Net: Attention enhanced global–local refined network for accurate detection of car body surface defects

The complex background on the car body surface, such as the orange peel-like texture and shiny metallic powder, poses a considerable challenge to automated defect detection. Two mainstream methods are currently used to tackle this challenge: global information-based and attention mechanism-based methods. However, these methods lack the capability to integrate valuable global-to-local information and explore deeper distinguishable features, thereby affecting the overall detection performance. To address this issue, we propose a novel attention enhanced global–local refined detection network (AEGLR-Net), which can perform effective global-to-local refined feature extraction and fusion. First, we design an adaptive Transformer–CNN tandem backbone (ATCT-backbone) to dynamically aware valuable global information and integrate local details to comprehensively extract specific features between defects and complex backgrounds. Then, we propose a novel refined cross-dimensional aggregation (RCDA) attention to facilitate the point-to-point interaction of multidimensional information, effectively emphasizing the representation of deeper discriminative defect features. Finally, we construct an attention-embedded flexible feature pyramid network (AE-FFPN), which incorporates the RCDA attention to guide the feature pyramid network in targeted feature fusion, thereby enhancing the efficiency of feature fusion in the detection model. Extensive comparative experiments demonstrate that the AEGLR-Net outperforms state-of-the-art approaches, attaining exceptional performance with 89.2 % mAP (mean average precision) and 85.5 FPS (frames per second).

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Robotics and Computer-integrated Manufacturing
Robotics and Computer-integrated Manufacturing 工程技术-工程:制造
CiteScore
24.10
自引率
13.50%
发文量
160
审稿时长
50 days
期刊介绍: The journal, Robotics and Computer-Integrated Manufacturing, focuses on sharing research applications that contribute to the development of new or enhanced robotics, manufacturing technologies, and innovative manufacturing strategies that are relevant to industry. Papers that combine theory and experimental validation are preferred, while review papers on current robotics and manufacturing issues are also considered. However, papers on traditional machining processes, modeling and simulation, supply chain management, and resource optimization are generally not within the scope of the journal, as there are more appropriate journals for these topics. Similarly, papers that are overly theoretical or mathematical will be directed to other suitable journals. The journal welcomes original papers in areas such as industrial robotics, human-robot collaboration in manufacturing, cloud-based manufacturing, cyber-physical production systems, big data analytics in manufacturing, smart mechatronics, machine learning, adaptive and sustainable manufacturing, and other fields involving unique manufacturing technologies.
期刊最新文献
A dual knowledge embedded hybrid model based on augmented data and improved loss function for tool wear monitoring A real-time collision avoidance method for redundant dual-arm robots in an open operational environment Less gets more attention: A novel human-centered MR remote collaboration assembly method with information recommendation and visual enhancement Drilling task planning and offline programming of a robotic multi-spindle drilling system for aero-engine nacelle acoustic liners Human-in-the-loop Multi-objective Bayesian Optimization for Directed Energy Deposition with in-situ monitoring
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1