Ning Xu;Congyu Qiao;Yuchen Zhao;Xin Geng;Min-Ling Zhang
{"title":"针对依赖于实例的部分标签学习的变量标签增强。","authors":"Ning Xu;Congyu Qiao;Yuchen Zhao;Xin Geng;Min-Ling Zhang","doi":"10.1109/TPAMI.2024.3455260","DOIUrl":null,"url":null,"abstract":"Partial label learning (PLL) is a form of weakly supervised learning, where each training example is linked to a set of candidate labels, among which only one label is correct. Most existing PLL approaches assume that the incorrect labels in each training example are randomly picked as the candidate labels. However, in practice, this assumption may not hold true, as the candidate labels are often instance-dependent. In this paper, we address the instance-dependent PLL problem and assume that each example is associated with a latent \n<italic>label distribution</i>\n where the incorrect label with a high degree is more likely to be annotated as a candidate label. Motivated by this consideration, we propose two methods \n<sc>Valen</small>\n and \n<sc>Milen</small>\n, which train the predictive model via utilizing the latent label distributions recovered by the label enhancement process. Specifically, \n<sc>Valen</small>\n recovers the latent label distributions via inferring the variational posterior density parameterized by an inference model with the deduced evidence lower bound. \n<sc>Milen</small>\n recovers the latent label distribution by adopting the variational approximation to bound the mutual information among the latent label distribution, observed labels and augmented instances. Experiments on benchmark and real-world datasets validate the effectiveness of the proposed methods.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"46 12","pages":"11298-11313"},"PeriodicalIF":0.0000,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Variational Label Enhancement for Instance-Dependent Partial Label Learning\",\"authors\":\"Ning Xu;Congyu Qiao;Yuchen Zhao;Xin Geng;Min-Ling Zhang\",\"doi\":\"10.1109/TPAMI.2024.3455260\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Partial label learning (PLL) is a form of weakly supervised learning, where each training example is linked to a set of candidate labels, among which only one label is correct. Most existing PLL approaches assume that the incorrect labels in each training example are randomly picked as the candidate labels. However, in practice, this assumption may not hold true, as the candidate labels are often instance-dependent. In this paper, we address the instance-dependent PLL problem and assume that each example is associated with a latent \\n<italic>label distribution</i>\\n where the incorrect label with a high degree is more likely to be annotated as a candidate label. Motivated by this consideration, we propose two methods \\n<sc>Valen</small>\\n and \\n<sc>Milen</small>\\n, which train the predictive model via utilizing the latent label distributions recovered by the label enhancement process. Specifically, \\n<sc>Valen</small>\\n recovers the latent label distributions via inferring the variational posterior density parameterized by an inference model with the deduced evidence lower bound. \\n<sc>Milen</small>\\n recovers the latent label distribution by adopting the variational approximation to bound the mutual information among the latent label distribution, observed labels and augmented instances. Experiments on benchmark and real-world datasets validate the effectiveness of the proposed methods.\",\"PeriodicalId\":94034,\"journal\":{\"name\":\"IEEE transactions on pattern analysis and machine intelligence\",\"volume\":\"46 12\",\"pages\":\"11298-11313\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE transactions on pattern analysis and machine intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10667671/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on pattern analysis and machine intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10667671/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Variational Label Enhancement for Instance-Dependent Partial Label Learning
Partial label learning (PLL) is a form of weakly supervised learning, where each training example is linked to a set of candidate labels, among which only one label is correct. Most existing PLL approaches assume that the incorrect labels in each training example are randomly picked as the candidate labels. However, in practice, this assumption may not hold true, as the candidate labels are often instance-dependent. In this paper, we address the instance-dependent PLL problem and assume that each example is associated with a latent
label distribution
where the incorrect label with a high degree is more likely to be annotated as a candidate label. Motivated by this consideration, we propose two methods
Valen
and
Milen
, which train the predictive model via utilizing the latent label distributions recovered by the label enhancement process. Specifically,
Valen
recovers the latent label distributions via inferring the variational posterior density parameterized by an inference model with the deduced evidence lower bound.
Milen
recovers the latent label distribution by adopting the variational approximation to bound the mutual information among the latent label distribution, observed labels and augmented instances. Experiments on benchmark and real-world datasets validate the effectiveness of the proposed methods.