Gang Xu , Yingshui Zhang , Qingrui Yue , Xiaogang Liu
{"title":"A deep learning framework for real-time multi-task recognition and measurement of concrete cracks","authors":"Gang Xu , Yingshui Zhang , Qingrui Yue , Xiaogang Liu","doi":"10.1016/j.aei.2025.103127","DOIUrl":null,"url":null,"abstract":"<div><div>This study presents an innovative deep learning framework, YOLO-DL, for automatic multi-task recognition of concrete cracks. The framework integrates the You Only Look Once (YOLO) object detection algorithm with the encoder-decoder architecture of the DeepLabv3 + model, incorporating an attention mechanism and a calibration module, resulting in three distinct branches for crack classification, localization detection, and semantic segmentation. The YOLO-DL model achieves a detection precision of 84.87 %, an [email protected] of 83.55 %, and a mean intersection-over-union (mIoU) of 94.94 % for crack segmentation. The model’s segmentation inference time is significantly shorter than that of the DeepLabv3+, fully convolutional networks (FCN), U-Net, and SegNet models, making it suitable for real-time concrete crack recognition. The model effectively handles classification, detection, and segmentation tasks, demonstrating enhanced performance and robustness, particularly with the inclusion of the attention mechanism. Additionally, a novel crack width measurement method based on the local element grid method is presented, achieving sub-pixel precision. This method provides comprehensive crack width information, including the maximum width of each crack and its corresponding location, with a maximum relative error of less than 10 %. The findings highlight the model’s strong inference performance, robust generalization ability, and promising real-time crack recognition capabilities.</div></div>","PeriodicalId":50941,"journal":{"name":"Advanced Engineering Informatics","volume":"65 ","pages":"Article 103127"},"PeriodicalIF":9.9000,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advanced Engineering Informatics","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1474034625000205","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/1/17 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
This study presents an innovative deep learning framework, YOLO-DL, for automatic multi-task recognition of concrete cracks. The framework integrates the You Only Look Once (YOLO) object detection algorithm with the encoder-decoder architecture of the DeepLabv3 + model, incorporating an attention mechanism and a calibration module, resulting in three distinct branches for crack classification, localization detection, and semantic segmentation. The YOLO-DL model achieves a detection precision of 84.87 %, an [email protected] of 83.55 %, and a mean intersection-over-union (mIoU) of 94.94 % for crack segmentation. The model’s segmentation inference time is significantly shorter than that of the DeepLabv3+, fully convolutional networks (FCN), U-Net, and SegNet models, making it suitable for real-time concrete crack recognition. The model effectively handles classification, detection, and segmentation tasks, demonstrating enhanced performance and robustness, particularly with the inclusion of the attention mechanism. Additionally, a novel crack width measurement method based on the local element grid method is presented, achieving sub-pixel precision. This method provides comprehensive crack width information, including the maximum width of each crack and its corresponding location, with a maximum relative error of less than 10 %. The findings highlight the model’s strong inference performance, robust generalization ability, and promising real-time crack recognition capabilities.
本文提出了一种创新的深度学习框架YOLO-DL,用于混凝土裂缝的多任务自动识别。该框架将You Only Look Once (YOLO)目标检测算法与DeepLabv3 +模型的编码器-解码器架构集成在一起,结合了注意机制和校准模块,形成了三个不同的分支,分别用于裂缝分类、定位检测和语义分割。YOLO-DL模型的检测精度为84.87%,[email protected]的检测精度为83.55%,裂缝分割的平均相交-超并度(mIoU)为94.94%。该模型的分割推理时间明显短于DeepLabv3+、全卷积网络(FCN)、U-Net和SegNet模型,适用于实时混凝土裂缝识别。该模型有效地处理分类、检测和分割任务,展示了增强的性能和鲁棒性,特别是在包含注意机制的情况下。此外,提出了一种基于局部单元网格法的裂缝宽度测量方法,该方法可达到亚像素精度。该方法提供了全面的裂缝宽度信息,包括每条裂缝的最大宽度及其对应位置,最大相对误差小于10%。研究结果表明,该模型具有较强的推理性能、较强的泛化能力和较好的实时裂缝识别能力。
期刊介绍:
Advanced Engineering Informatics is an international Journal that solicits research papers with an emphasis on 'knowledge' and 'engineering applications'. The Journal seeks original papers that report progress in applying methods of engineering informatics. These papers should have engineering relevance and help provide a scientific base for more reliable, spontaneous, and creative engineering decision-making. Additionally, papers should demonstrate the science of supporting knowledge-intensive engineering tasks and validate the generality, power, and scalability of new methods through rigorous evaluation, preferably both qualitatively and quantitatively. Abstracting and indexing for Advanced Engineering Informatics include Science Citation Index Expanded, Scopus and INSPEC.