{"title":"基于多头关注的分层门控卷积网络文本分类","authors":"Haizhou Du, Jingu Qian","doi":"10.1109/ICSAI.2018.8599366","DOIUrl":null,"url":null,"abstract":"Text classification is a fundamental problem in natural language processing. Recently, neural network models have been demonstrated to be capable of achieving remarkable performance in this domain. However, none of existing method can achieve excellent classification accuracy while concerning of computational cost. To solve this problem, we proposed hierarchical gated convolutional networks with multi-head attention which reduces computational cost through its two distinctive characteristics to save considerable model parameters. First, it has a hierarchical structure the same as the hierarchical structure of documents that has word-level and sentence-level, which not only benefits to classification performance but also reduces computational cost significantly by reusing parameters of the model in each sentence. Second, we apply gated convolutional network on both levels that enables our model achieved comparable performance to very deep networks with relatively shallow network depth. To further improve the performance of our model, multi-head attention mechanism is employed to differentiate more or less importance of words or sentences for better construction of document representation. Experiments conducted on the commonly used Yelp reviews datasets demonstrate that the proposed architecture obtains competitive performance against the state-of-the-art methods.","PeriodicalId":375852,"journal":{"name":"2018 5th International Conference on Systems and Informatics (ICSAI)","volume":"85 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"10","resultStr":"{\"title\":\"Hierarchical Gated Convolutional Networks with Multi-Head Attention for Text Classification\",\"authors\":\"Haizhou Du, Jingu Qian\",\"doi\":\"10.1109/ICSAI.2018.8599366\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Text classification is a fundamental problem in natural language processing. Recently, neural network models have been demonstrated to be capable of achieving remarkable performance in this domain. However, none of existing method can achieve excellent classification accuracy while concerning of computational cost. To solve this problem, we proposed hierarchical gated convolutional networks with multi-head attention which reduces computational cost through its two distinctive characteristics to save considerable model parameters. First, it has a hierarchical structure the same as the hierarchical structure of documents that has word-level and sentence-level, which not only benefits to classification performance but also reduces computational cost significantly by reusing parameters of the model in each sentence. Second, we apply gated convolutional network on both levels that enables our model achieved comparable performance to very deep networks with relatively shallow network depth. To further improve the performance of our model, multi-head attention mechanism is employed to differentiate more or less importance of words or sentences for better construction of document representation. Experiments conducted on the commonly used Yelp reviews datasets demonstrate that the proposed architecture obtains competitive performance against the state-of-the-art methods.\",\"PeriodicalId\":375852,\"journal\":{\"name\":\"2018 5th International Conference on Systems and Informatics (ICSAI)\",\"volume\":\"85 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"10\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2018 5th International Conference on Systems and Informatics (ICSAI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICSAI.2018.8599366\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 5th International Conference on Systems and Informatics (ICSAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSAI.2018.8599366","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Hierarchical Gated Convolutional Networks with Multi-Head Attention for Text Classification
Text classification is a fundamental problem in natural language processing. Recently, neural network models have been demonstrated to be capable of achieving remarkable performance in this domain. However, none of existing method can achieve excellent classification accuracy while concerning of computational cost. To solve this problem, we proposed hierarchical gated convolutional networks with multi-head attention which reduces computational cost through its two distinctive characteristics to save considerable model parameters. First, it has a hierarchical structure the same as the hierarchical structure of documents that has word-level and sentence-level, which not only benefits to classification performance but also reduces computational cost significantly by reusing parameters of the model in each sentence. Second, we apply gated convolutional network on both levels that enables our model achieved comparable performance to very deep networks with relatively shallow network depth. To further improve the performance of our model, multi-head attention mechanism is employed to differentiate more or less importance of words or sentences for better construction of document representation. Experiments conducted on the commonly used Yelp reviews datasets demonstrate that the proposed architecture obtains competitive performance against the state-of-the-art methods.