Pub Date : 2025-12-27DOI: 10.1007/s10489-025-07058-0
Xing Wu, Deyu Gao, Zhi Li, Junfeng Yao, Quan Qian, Jun Song
Cultural relic image restoration presents unique challenges due to irregular damage and historically specific textures, which standard deep learning methods struggle to address. This paper proposes a novel two-stage Transformer-CNN framework tailored for this task. The first stage leverages a Transformer to capture global structural dependencies from low-resolution priors, generating coherent coarse proposals. The second stage employs a specialized CNN to refine fine-grained textures from these proposals, optimized by a compound perceptual loss function. Validated on a new large-scale dataset of 88,000 East Asian cultural relic images, our approach demonstrates state-of-the-art performance. A key contribution is the generation of diversified restoration outputs, providing conservators with multiple valid references for decision-making. This work establishes an effective paradigm for digital heritage conservation that balances global structural integrity with local texture fidelity.
{"title":"Cultural relic image restoration using two-stage transformer-CNN framework","authors":"Xing Wu, Deyu Gao, Zhi Li, Junfeng Yao, Quan Qian, Jun Song","doi":"10.1007/s10489-025-07058-0","DOIUrl":"10.1007/s10489-025-07058-0","url":null,"abstract":"<div><p>Cultural relic image restoration presents unique challenges due to irregular damage and historically specific textures, which standard deep learning methods struggle to address. This paper proposes a novel two-stage Transformer-CNN framework tailored for this task. The first stage leverages a Transformer to capture global structural dependencies from low-resolution priors, generating coherent coarse proposals. The second stage employs a specialized CNN to refine fine-grained textures from these proposals, optimized by a compound perceptual loss function. Validated on a new large-scale dataset of 88,000 East Asian cultural relic images, our approach demonstrates state-of-the-art performance. A key contribution is the generation of diversified restoration outputs, providing conservators with multiple valid references for decision-making. This work establishes an effective paradigm for digital heritage conservation that balances global structural integrity with local texture fidelity.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"56 1","pages":""},"PeriodicalIF":3.5,"publicationDate":"2025-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145831046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-27DOI: 10.1007/s10489-025-07051-7
Qian Zhang, Qiu Chen
Deep neural networks suffer from overfitting when training samples contain inaccurate annotations (noisy labels), leading to suboptimal performance. In addressing this challenge, current methods for learning with noisy labels employ specific criteria, such as small loss, historical prediction, etc., to distinguish clean and noisy instances. Subsequently, semi-supervised learning techniques are introduced to boost performance. Most of them are one-stage frameworks that aim to achieve optimal sample partitioning and robust SSL training within a single iteration, thereby increasing training difficulty and complexity. To address this limitation, we propose a novel two-stage noisy label learning framework called UCRT, which consists of uniform consistency selection and robust training. In the first stage, the emphasis lies on creating a more uniform and accurate clean set, while the second stage uniformly extends this clean set to improve model performance by introducing SSL techniques. Comprehensive experiments conducted on both synthetic and real-world noisy datasets demonstrate the stability of UCRT across various noise types, showcasing superior performance compared with state-of-the-art methods. The code will be available at: https://github.com/LanXiaoPang613/UCRT.
{"title":"UCRT: a two-stage noisy label learning framework with uniform consistency selection and robust training","authors":"Qian Zhang, Qiu Chen","doi":"10.1007/s10489-025-07051-7","DOIUrl":"10.1007/s10489-025-07051-7","url":null,"abstract":"<div><p>Deep neural networks suffer from overfitting when training samples contain inaccurate annotations (noisy labels), leading to suboptimal performance. In addressing this challenge, current methods for learning with noisy labels employ specific criteria, such as small loss, historical prediction, etc., to distinguish clean and noisy instances. Subsequently, semi-supervised learning techniques are introduced to boost performance. Most of them are one-stage frameworks that aim to achieve optimal sample partitioning and robust SSL training within a single iteration, thereby increasing training difficulty and complexity. To address this limitation, we propose a novel two-stage noisy label learning framework called UCRT, which consists of uniform consistency selection and robust training. In the first stage, the emphasis lies on creating a more uniform and accurate clean set, while the second stage uniformly extends this clean set to improve model performance by introducing SSL techniques. Comprehensive experiments conducted on both synthetic and real-world noisy datasets demonstrate the stability of UCRT across various noise types, showcasing superior performance compared with state-of-the-art methods. The code will be available at: https://github.com/LanXiaoPang613/UCRT.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"56 1","pages":""},"PeriodicalIF":3.5,"publicationDate":"2025-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145831045","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}