Trimming the Risk: Towards Reliable Continuous Training for Deep Learning Inspection Systems

Altaf Allah Abbassi, Houssem Ben Braiek, Foutse Khomh, Thomas Reid
{"title":"Trimming the Risk: Towards Reliable Continuous Training for Deep Learning Inspection Systems","authors":"Altaf Allah Abbassi, Houssem Ben Braiek, Foutse Khomh, Thomas Reid","doi":"arxiv-2409.09108","DOIUrl":null,"url":null,"abstract":"The industry increasingly relies on deep learning (DL) technology for\nmanufacturing inspections, which are challenging to automate with rule-based\nmachine vision algorithms. DL-powered inspection systems derive defect patterns\nfrom labeled images, combining human-like agility with the consistency of a\ncomputerized system. However, finite labeled datasets often fail to encompass\nall natural variations necessitating Continuous Training (CT) to regularly\nadjust their models with recent data. Effective CT requires fresh labeled\nsamples from the original distribution; otherwise, selfgenerated labels can\nlead to silent performance degradation. To mitigate this risk, we develop a\nrobust CT-based maintenance approach that updates DL models using reliable data\nselections through a two-stage filtering process. The initial stage filters out\nlow-confidence predictions, as the model inherently discredits them. The second\nstage uses variational auto-encoders and histograms to generate image\nembeddings that capture latent and pixel characteristics, then rejects the\ninputs of substantially shifted embeddings as drifted data with erroneous\noverconfidence. Then, a fine-tuning of the original DL model is executed on the\nfiltered inputs while validating on a mixture of recent production and original\ndatasets. This strategy mitigates catastrophic forgetting and ensures the model\nadapts effectively to new operational conditions. Evaluations on industrial\ninspection systems for popsicle stick prints and glass bottles using critical\nreal-world datasets showed less than 9% of erroneous self-labeled data are\nretained after filtering and used for fine-tuning, improving model performance\non production data by up to 14% without compromising its results on original\nvalidation data.","PeriodicalId":501278,"journal":{"name":"arXiv - CS - Software Engineering","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Software Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.09108","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The industry increasingly relies on deep learning (DL) technology for manufacturing inspections, which are challenging to automate with rule-based machine vision algorithms. DL-powered inspection systems derive defect patterns from labeled images, combining human-like agility with the consistency of a computerized system. However, finite labeled datasets often fail to encompass all natural variations necessitating Continuous Training (CT) to regularly adjust their models with recent data. Effective CT requires fresh labeled samples from the original distribution; otherwise, selfgenerated labels can lead to silent performance degradation. To mitigate this risk, we develop a robust CT-based maintenance approach that updates DL models using reliable data selections through a two-stage filtering process. The initial stage filters out low-confidence predictions, as the model inherently discredits them. The second stage uses variational auto-encoders and histograms to generate image embeddings that capture latent and pixel characteristics, then rejects the inputs of substantially shifted embeddings as drifted data with erroneous overconfidence. Then, a fine-tuning of the original DL model is executed on the filtered inputs while validating on a mixture of recent production and original datasets. This strategy mitigates catastrophic forgetting and ensures the model adapts effectively to new operational conditions. Evaluations on industrial inspection systems for popsicle stick prints and glass bottles using critical real-world datasets showed less than 9% of erroneous self-labeled data are retained after filtering and used for fine-tuning, improving model performance on production data by up to 14% without compromising its results on original validation data.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
降低风险:为深度学习检测系统提供可靠的持续训练
制造业越来越依赖于深度学习(DL)技术来进行制造检测,而使用基于规则的机器视觉算法来实现自动化具有挑战性。由深度学习驱动的检测系统从标记图像中提取缺陷模式,将人类的敏捷性与计算机化系统的一致性结合起来。然而,有限的标注数据集往往无法涵盖所有自然变化,因此需要进行持续训练(CT),利用最新数据定期调整模型。有效的持续训练需要来自原始分布的新鲜标签样本;否则,自生成的标签会导致无声的性能下降。为了降低这种风险,我们开发了一种基于 CT 的稳健维护方法,通过两阶段过滤过程,使用可靠的数据选择更新 DL 模型。第一阶段过滤掉置信度低的预测,因为模型本身就会否定这些预测。第二阶段使用变异自动编码器和直方图来生成图像嵌入,以捕捉潜在特征和像素特征,然后将大幅偏移的嵌入作为具有错误过高置信度的漂移数据剔除。然后,对过滤后的输入执行原始 DL 模型的微调,同时在最新生产数据集和原始数据集的混合数据上进行验证。这种策略可以减少灾难性遗忘,确保模型有效适应新的运行条件。利用重要的真实数据集对冰棒棍印花和玻璃瓶的工业检测系统进行的评估表明,经过过滤并用于微调后,错误的自标注数据只保留了不到 9%,从而将模型在生产数据上的性能提高了 14%,而不会影响其在原始验证数据上的结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Promise and Peril of Collaborative Code Generation Models: Balancing Effectiveness and Memorization Shannon Entropy is better Feature than Category and Sentiment in User Feedback Processing Motivations, Challenges, Best Practices, and Benefits for Bots and Conversational Agents in Software Engineering: A Multivocal Literature Review A Taxonomy of Self-Admitted Technical Debt in Deep Learning Systems Investigating team maturity in an agile automotive reorganization
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1