DyLas: A dynamic label alignment strategy for large-scale multi-label text classification

IF 15.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Information Fusion Pub Date : 2025-03-11 DOI:10.1016/j.inffus.2025.103081
Lin Ren, Yongbin Liu, Chunping Ouyang, Ying Yu, Shuda Zhou, Yidong He, Yaping Wan
{"title":"DyLas: A dynamic label alignment strategy for large-scale multi-label text classification","authors":"Lin Ren,&nbsp;Yongbin Liu,&nbsp;Chunping Ouyang,&nbsp;Ying Yu,&nbsp;Shuda Zhou,&nbsp;Yidong He,&nbsp;Yaping Wan","doi":"10.1016/j.inffus.2025.103081","DOIUrl":null,"url":null,"abstract":"<div><div>Large-scale multi-label Text Classification (LMTC) is an advanced facet of NLP that entails assigning multiple labels to text documents from an extensive label space, often comprising thousands to millions of possible categories. This classification task is pivotal across various domains, including e-commerce product tagging, news categorization, medical code assignment, and legal document analysis, where accurate multi-label predictions drive search efficiency, recommendation systems, and regulatory compliance. However, LMTC poses significant challenges, the dynamic nature of label sets, which traditional supervised learning approaches find difficult to address due to their reliance on annotated data. In light of this challenge, this work introduces a novel approach leveraging Large Language Models (LLMs) for dynamic label alignment in LMTC tasks, based on counterfactual analysis, called DyLas (<u>Dy</u>namic <u>L</u>abel <u>A</u>lignment <u>S</u>trategy). Through a multi-step strategy, we aim to mitigate the issues arising from dynamic label sets. We evaluate the performance of LMTC on the 8 LLMs by 4 datasets and apply DyLas to 3 closed-source and 3 open-weight LLMs. Compared to the single-step approach, our method, DyLas, achieves improvements in almost all metrics across the datasets. Our method can also work well in dynamic label set environments. This work not only demonstrates the potential of LLMs to address complex classification challenges, but is also, to the best of our knowledge, the first to address dynamic label set challenges in LMTC tasks with LLMs without requiring additional model training.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"120 ","pages":"Article 103081"},"PeriodicalIF":15.5000,"publicationDate":"2025-03-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S156625352500154X","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Large-scale multi-label Text Classification (LMTC) is an advanced facet of NLP that entails assigning multiple labels to text documents from an extensive label space, often comprising thousands to millions of possible categories. This classification task is pivotal across various domains, including e-commerce product tagging, news categorization, medical code assignment, and legal document analysis, where accurate multi-label predictions drive search efficiency, recommendation systems, and regulatory compliance. However, LMTC poses significant challenges, the dynamic nature of label sets, which traditional supervised learning approaches find difficult to address due to their reliance on annotated data. In light of this challenge, this work introduces a novel approach leveraging Large Language Models (LLMs) for dynamic label alignment in LMTC tasks, based on counterfactual analysis, called DyLas (Dynamic Label Alignment Strategy). Through a multi-step strategy, we aim to mitigate the issues arising from dynamic label sets. We evaluate the performance of LMTC on the 8 LLMs by 4 datasets and apply DyLas to 3 closed-source and 3 open-weight LLMs. Compared to the single-step approach, our method, DyLas, achieves improvements in almost all metrics across the datasets. Our method can also work well in dynamic label set environments. This work not only demonstrates the potential of LLMs to address complex classification challenges, but is also, to the best of our knowledge, the first to address dynamic label set challenges in LMTC tasks with LLMs without requiring additional model training.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
DyLas:大规模多标签文本分类的动态标签对齐策略
大规模多标签文本分类(LMTC)是NLP的一个高级方面,它需要从广泛的标签空间为文本文档分配多个标签,通常包含数千到数百万个可能的类别。这个分类任务在各个领域都很关键,包括电子商务产品标记、新闻分类、医疗代码分配和法律文档分析,其中精确的多标签预测可以提高搜索效率、推荐系统和法规遵从性。然而,LMTC提出了重大挑战,即标签集的动态性,传统的监督学习方法由于依赖于注释数据而难以解决。鉴于这一挑战,本工作介绍了一种利用大型语言模型(llm)在LMTC任务中进行动态标签对齐的新方法,该方法基于反事实分析,称为DyLas(动态标签对齐策略)。通过多步骤策略,我们的目标是减轻由动态标签集引起的问题。我们通过4个数据集评估了LMTC在8个llm上的性能,并将DyLas应用于3个闭源和3个开权llm。与单步方法相比,我们的方法,DyLas,实现了跨数据集几乎所有指标的改进。我们的方法在动态标签集环境中也能很好地工作。这项工作不仅展示了llm解决复杂分类挑战的潜力,而且据我们所知,这是第一个在不需要额外模型训练的情况下,使用llm解决LMTC任务中动态标签集挑战的研究。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Information Fusion
Information Fusion 工程技术-计算机:理论方法
CiteScore
33.20
自引率
4.30%
发文量
161
审稿时长
7.9 months
期刊介绍: Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.
期刊最新文献
GeoCraft: A Diffusion Model-Based 3D Reconstruction Method Driven by Image and Point Cloud Fusion GC-Fed: Gradient Centralized Federated Learning with Partial Client Participation Editorial Board Arbitrary‑Scale Spatial–Spectral Fusion using Kernel Integral and Progressive Resampling Data science: a natural ecosystem
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1