Jing Yang , Xingjiang Ma , Yuankai Wu , Chengjiang Li , Zhidong Su , Ji Xu , Yixiong Feng
{"title":"AOGN-CZSL: An Attribute- and Object-Guided Network for Compositional Zero-Shot Learning","authors":"Jing Yang , Xingjiang Ma , Yuankai Wu , Chengjiang Li , Zhidong Su , Ji Xu , Yixiong Feng","doi":"10.1016/j.inffus.2025.103096","DOIUrl":null,"url":null,"abstract":"<div><div>Humans are able to readily acquire knowledge about unfamiliar and unknown objects. However, it is extremely challenging for artificial intelligence to achieve this skill. With the rapid development of artificial intelligence, compositional zero-shot learning (CZSL) can generalize unseen compositions by learning prior knowledge of seen attributes and object compositions during training. Although existing composition-based and relationship-based methods show great potential for addressing this challenge, they still exhibit some limitations. Composition-based methods often ignore the intrinsic correlations between attributes, objects, and images, which may lead the model to perform poorly when it is generalized to unseen compositions. Some relationship-based methods can better capture the relationships between attributes, objects, and images but may overlook the interdependencies between distinct attributes and objects. Therefore, the advantages of the composition-based and relationship-based methods are combined, and a new method is proposed for learning attribute and object dependencies (AOGN-CZSL). AOGN-CZSL learns the dependencies between different attributes or objects. It also learns all attributes and objects simultaneously. Different from traditional composition-based methods, it typically address each attribute-object compositions separately. Moreover, unlike general relation-based approaches, this paper adopts learned textual and visual modality features of attributes and objects for attribute scoring and object scoring, respectively. The code and dataset are available at: <span><span>https://github.com/ybyangjing/AOGN-CZSL</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"120 ","pages":"Article 103096"},"PeriodicalIF":15.5000,"publicationDate":"2025-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525001691","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Humans are able to readily acquire knowledge about unfamiliar and unknown objects. However, it is extremely challenging for artificial intelligence to achieve this skill. With the rapid development of artificial intelligence, compositional zero-shot learning (CZSL) can generalize unseen compositions by learning prior knowledge of seen attributes and object compositions during training. Although existing composition-based and relationship-based methods show great potential for addressing this challenge, they still exhibit some limitations. Composition-based methods often ignore the intrinsic correlations between attributes, objects, and images, which may lead the model to perform poorly when it is generalized to unseen compositions. Some relationship-based methods can better capture the relationships between attributes, objects, and images but may overlook the interdependencies between distinct attributes and objects. Therefore, the advantages of the composition-based and relationship-based methods are combined, and a new method is proposed for learning attribute and object dependencies (AOGN-CZSL). AOGN-CZSL learns the dependencies between different attributes or objects. It also learns all attributes and objects simultaneously. Different from traditional composition-based methods, it typically address each attribute-object compositions separately. Moreover, unlike general relation-based approaches, this paper adopts learned textual and visual modality features of attributes and objects for attribute scoring and object scoring, respectively. The code and dataset are available at: https://github.com/ybyangjing/AOGN-CZSL.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.