Zhe Wang, Qingbiao Li, Bin Wang, Tong Wu, Chengwei Chang
{"title":"通过预关注机制衍生词典改进文本分类","authors":"Zhe Wang, Qingbiao Li, Bin Wang, Tong Wu, Chengwei Chang","doi":"10.1007/s10489-024-05742-1","DOIUrl":null,"url":null,"abstract":"<p>A comprehensive and high-quality lexicon plays a crucial role in traditional text classification approaches. It improves the utilization of linguistic knowledge. Although it is helpful for this task, the lexicon has received little attention in current neural network models. First, obtaining a high-quality lexicon is not easy. Second, an effective automated lexicon extraction method is lacking, and most lexicons are handcrafted, which is very inefficient for big data. Finally, there is no effective way to use a lexicon in a neural network. To address these limitations, we propose a pre-attention mechanism for text classification in this study, which can learn the attention values of various words based on their effects on classification tasks. Words with different attention values can form a domain lexicon. Experiments on three publicly available and authoritative benchmark text classification tasks show that our models obtain competitive results compared with state-of-the-art models. For the same dataset, when we use the pre-attention mechanism to obtain attention values, followed by different neural networks, words with high attention values have a high degree of coincidence, which proves the versatility and portability of the pre-attention mechanism. We can obtain stable lexicons using attention values, which is an inspiring method of information extraction.</p>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 22","pages":"11765 - 11778"},"PeriodicalIF":3.4000,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Improving text classification through pre-attention mechanism-derived lexicons\",\"authors\":\"Zhe Wang, Qingbiao Li, Bin Wang, Tong Wu, Chengwei Chang\",\"doi\":\"10.1007/s10489-024-05742-1\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>A comprehensive and high-quality lexicon plays a crucial role in traditional text classification approaches. It improves the utilization of linguistic knowledge. Although it is helpful for this task, the lexicon has received little attention in current neural network models. First, obtaining a high-quality lexicon is not easy. Second, an effective automated lexicon extraction method is lacking, and most lexicons are handcrafted, which is very inefficient for big data. Finally, there is no effective way to use a lexicon in a neural network. To address these limitations, we propose a pre-attention mechanism for text classification in this study, which can learn the attention values of various words based on their effects on classification tasks. Words with different attention values can form a domain lexicon. Experiments on three publicly available and authoritative benchmark text classification tasks show that our models obtain competitive results compared with state-of-the-art models. For the same dataset, when we use the pre-attention mechanism to obtain attention values, followed by different neural networks, words with high attention values have a high degree of coincidence, which proves the versatility and portability of the pre-attention mechanism. We can obtain stable lexicons using attention values, which is an inspiring method of information extraction.</p>\",\"PeriodicalId\":8041,\"journal\":{\"name\":\"Applied Intelligence\",\"volume\":\"54 22\",\"pages\":\"11765 - 11778\"},\"PeriodicalIF\":3.4000,\"publicationDate\":\"2024-09-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Applied Intelligence\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://link.springer.com/article/10.1007/s10489-024-05742-1\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Intelligence","FirstCategoryId":"94","ListUrlMain":"https://link.springer.com/article/10.1007/s10489-024-05742-1","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Improving text classification through pre-attention mechanism-derived lexicons
A comprehensive and high-quality lexicon plays a crucial role in traditional text classification approaches. It improves the utilization of linguistic knowledge. Although it is helpful for this task, the lexicon has received little attention in current neural network models. First, obtaining a high-quality lexicon is not easy. Second, an effective automated lexicon extraction method is lacking, and most lexicons are handcrafted, which is very inefficient for big data. Finally, there is no effective way to use a lexicon in a neural network. To address these limitations, we propose a pre-attention mechanism for text classification in this study, which can learn the attention values of various words based on their effects on classification tasks. Words with different attention values can form a domain lexicon. Experiments on three publicly available and authoritative benchmark text classification tasks show that our models obtain competitive results compared with state-of-the-art models. For the same dataset, when we use the pre-attention mechanism to obtain attention values, followed by different neural networks, words with high attention values have a high degree of coincidence, which proves the versatility and portability of the pre-attention mechanism. We can obtain stable lexicons using attention values, which is an inspiring method of information extraction.
期刊介绍:
With a focus on research in artificial intelligence and neural networks, this journal addresses issues involving solutions of real-life manufacturing, defense, management, government and industrial problems which are too complex to be solved through conventional approaches and require the simulation of intelligent thought processes, heuristics, applications of knowledge, and distributed and parallel processing. The integration of these multiple approaches in solving complex problems is of particular importance.
The journal presents new and original research and technological developments, addressing real and complex issues applicable to difficult problems. It provides a medium for exchanging scientific research and technological achievements accomplished by the international community.