Zhanxuan Hu, Yu Duan, Yaming Zhang, Rong Wang, Feiping Nie
{"title":"用于广义类别发现的带有分布一致性正则化的原型分类器:强大的基线","authors":"Zhanxuan Hu, Yu Duan, Yaming Zhang, Rong Wang, Feiping Nie","doi":"10.1016/j.neunet.2024.106908","DOIUrl":null,"url":null,"abstract":"<p><p>Generalized Category Discovery (GCD) addresses a more realistic and challenging setting in semi-supervised visual recognition, where unlabeled data contains samples from both known and novel categories. Recently, prototypical classifier has shown prominent performance on this issue, with the Softmax-based Cross-Entropy loss (SCE) commonly employed to optimize the distance between a sample and prototypes. However, the inherent non-bijectiveness of SCE prevents it from resolving intraclass relations among samples, resulting in semantic ambiguity. To mitigate this issue, we propose Distribution Consistency Regularization (DCR) for the prototypical classifier. By leveraging a simple intraclass consistency loss, we enforce the classifier to yield consistent distributions for samples belonging to the same class. In doing so, we equip the classifier to better capture local structures and alleviate semantic ambiguity. Additionally, we propose using partial labels, rather than hard pseudo labels, to explore potential positive pairs in unlabeled data, thereby reducing the risk of introducing noisy supervisory signals. DCR requires no external sophisticated module, rendering the enhanced model concise and efficient. Extensive experiments validate consistent performance benefits of DCR while achieving competitive or better performance on six benchmarks. Hence, our method can serve as a strong baseline for GCD. Our code is available at: https://github.com/yichenwang231/DCR.</p>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"182 ","pages":"106908"},"PeriodicalIF":6.0000,"publicationDate":"2024-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Prototypical classifier with distribution consistency regularization for generalized category discovery: A strong baseline.\",\"authors\":\"Zhanxuan Hu, Yu Duan, Yaming Zhang, Rong Wang, Feiping Nie\",\"doi\":\"10.1016/j.neunet.2024.106908\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Generalized Category Discovery (GCD) addresses a more realistic and challenging setting in semi-supervised visual recognition, where unlabeled data contains samples from both known and novel categories. Recently, prototypical classifier has shown prominent performance on this issue, with the Softmax-based Cross-Entropy loss (SCE) commonly employed to optimize the distance between a sample and prototypes. However, the inherent non-bijectiveness of SCE prevents it from resolving intraclass relations among samples, resulting in semantic ambiguity. To mitigate this issue, we propose Distribution Consistency Regularization (DCR) for the prototypical classifier. By leveraging a simple intraclass consistency loss, we enforce the classifier to yield consistent distributions for samples belonging to the same class. In doing so, we equip the classifier to better capture local structures and alleviate semantic ambiguity. Additionally, we propose using partial labels, rather than hard pseudo labels, to explore potential positive pairs in unlabeled data, thereby reducing the risk of introducing noisy supervisory signals. DCR requires no external sophisticated module, rendering the enhanced model concise and efficient. Extensive experiments validate consistent performance benefits of DCR while achieving competitive or better performance on six benchmarks. Hence, our method can serve as a strong baseline for GCD. Our code is available at: https://github.com/yichenwang231/DCR.</p>\",\"PeriodicalId\":49763,\"journal\":{\"name\":\"Neural Networks\",\"volume\":\"182 \",\"pages\":\"106908\"},\"PeriodicalIF\":6.0000,\"publicationDate\":\"2024-11-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neural Networks\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://doi.org/10.1016/j.neunet.2024.106908\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1016/j.neunet.2024.106908","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Prototypical classifier with distribution consistency regularization for generalized category discovery: A strong baseline.
Generalized Category Discovery (GCD) addresses a more realistic and challenging setting in semi-supervised visual recognition, where unlabeled data contains samples from both known and novel categories. Recently, prototypical classifier has shown prominent performance on this issue, with the Softmax-based Cross-Entropy loss (SCE) commonly employed to optimize the distance between a sample and prototypes. However, the inherent non-bijectiveness of SCE prevents it from resolving intraclass relations among samples, resulting in semantic ambiguity. To mitigate this issue, we propose Distribution Consistency Regularization (DCR) for the prototypical classifier. By leveraging a simple intraclass consistency loss, we enforce the classifier to yield consistent distributions for samples belonging to the same class. In doing so, we equip the classifier to better capture local structures and alleviate semantic ambiguity. Additionally, we propose using partial labels, rather than hard pseudo labels, to explore potential positive pairs in unlabeled data, thereby reducing the risk of introducing noisy supervisory signals. DCR requires no external sophisticated module, rendering the enhanced model concise and efficient. Extensive experiments validate consistent performance benefits of DCR while achieving competitive or better performance on six benchmarks. Hence, our method can serve as a strong baseline for GCD. Our code is available at: https://github.com/yichenwang231/DCR.
期刊介绍:
Neural Networks is a platform that aims to foster an international community of scholars and practitioners interested in neural networks, deep learning, and other approaches to artificial intelligence and machine learning. Our journal invites submissions covering various aspects of neural networks research, from computational neuroscience and cognitive modeling to mathematical analyses and engineering applications. By providing a forum for interdisciplinary discussions between biology and technology, we aim to encourage the development of biologically-inspired artificial intelligence.