Zhiqun Pan , Yongxiong Wang , Jiapeng Zhang , Yihan Shan , Zhe Wang , Jin Peng
{"title":"Distilled fine-grained domain adversarial network prompted by normalization and regularization","authors":"Zhiqun Pan , Yongxiong Wang , Jiapeng Zhang , Yihan Shan , Zhe Wang , Jin Peng","doi":"10.1016/j.ins.2025.121970","DOIUrl":null,"url":null,"abstract":"<div><div>Fine-grained domain adaptation presents a significant challenge in machine learning, where subtle differences between classes and domains often lead to poor generalization. Addressing this issue is crucial for improving the accuracy and robustness of models when applied to unseen data from different but related domains. We propose a distilled fine-grained domain adversarial network in which feature distributions across diverse domains are effectively aligned, and classification discrimination is enhanced during the mini-batch training phase. Based on knowledge distillation, the capability for fine-grained feature extraction is transferred from the teacher network to the student feature extractor, all while preserving the original data and prediction distributions. Consequently, low-entropy prediction distributions can be effectively leveraged by the domain adversarial network for unsupervised training in the target domain, leading to notable improvements in fine-grained recognition performance. In order to train the domain adversarial network more effectively, a linear transformation bottleneck layer for feature normalization is introduced after the feature extractor. This layer helps alleviate the normalization conflict between training and inference, which arises due to domain shift in mini-batches of data. Moreover, to strike a balance between the competing effects of the classifier and discriminator, we devise a double bilinear dropout module in the domain discriminator.</div></div>","PeriodicalId":51063,"journal":{"name":"Information Sciences","volume":"704 ","pages":"Article 121970"},"PeriodicalIF":8.1000,"publicationDate":"2025-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Sciences","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0020025525001021","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"0","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Fine-grained domain adaptation presents a significant challenge in machine learning, where subtle differences between classes and domains often lead to poor generalization. Addressing this issue is crucial for improving the accuracy and robustness of models when applied to unseen data from different but related domains. We propose a distilled fine-grained domain adversarial network in which feature distributions across diverse domains are effectively aligned, and classification discrimination is enhanced during the mini-batch training phase. Based on knowledge distillation, the capability for fine-grained feature extraction is transferred from the teacher network to the student feature extractor, all while preserving the original data and prediction distributions. Consequently, low-entropy prediction distributions can be effectively leveraged by the domain adversarial network for unsupervised training in the target domain, leading to notable improvements in fine-grained recognition performance. In order to train the domain adversarial network more effectively, a linear transformation bottleneck layer for feature normalization is introduced after the feature extractor. This layer helps alleviate the normalization conflict between training and inference, which arises due to domain shift in mini-batches of data. Moreover, to strike a balance between the competing effects of the classifier and discriminator, we devise a double bilinear dropout module in the domain discriminator.
期刊介绍:
Informatics and Computer Science Intelligent Systems Applications is an esteemed international journal that focuses on publishing original and creative research findings in the field of information sciences. We also feature a limited number of timely tutorial and surveying contributions.
Our journal aims to cater to a diverse audience, including researchers, developers, managers, strategic planners, graduate students, and anyone interested in staying up-to-date with cutting-edge research in information science, knowledge engineering, and intelligent systems. While readers are expected to share a common interest in information science, they come from varying backgrounds such as engineering, mathematics, statistics, physics, computer science, cell biology, molecular biology, management science, cognitive science, neurobiology, behavioral sciences, and biochemistry.