{"title":"An Improved Adaptive and Structured Sentence Embedding","authors":"Ke Fan, Hong Li, Xinyue Jiang","doi":"10.1109/ICSGEA.2019.00053","DOIUrl":null,"url":null,"abstract":"Recently, attention mechanism has aroused great interest in various fields of Natural Language Processing (NLP). In this paper, we propose a new model for extracting an interpretable sentence embedding by introducing an \"Adaptive self-attention\". Instead of using a vector, we use a 2-D matrix to represent the embedding and each valid row of the matrix represents a part of sentence. In addition, a length hierarchy mechanism with a unique loss function is applied to adaptively adjust the number of the valid rows of the matrix, which can solve the problem of attention redundancy in short sentences and lack of attention in long sentences. We evaluate our model on text classification tasks: news categorization, review categorization and opinion classification. The results show that our model, compared with other sentence embedding methods, achieve significant improvement in terms of performance when there exists a large amount of data and the length of the data is evenly distributed.","PeriodicalId":201721,"journal":{"name":"2019 International Conference on Smart Grid and Electrical Automation (ICSGEA)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 International Conference on Smart Grid and Electrical Automation (ICSGEA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSGEA.2019.00053","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Recently, attention mechanism has aroused great interest in various fields of Natural Language Processing (NLP). In this paper, we propose a new model for extracting an interpretable sentence embedding by introducing an "Adaptive self-attention". Instead of using a vector, we use a 2-D matrix to represent the embedding and each valid row of the matrix represents a part of sentence. In addition, a length hierarchy mechanism with a unique loss function is applied to adaptively adjust the number of the valid rows of the matrix, which can solve the problem of attention redundancy in short sentences and lack of attention in long sentences. We evaluate our model on text classification tasks: news categorization, review categorization and opinion classification. The results show that our model, compared with other sentence embedding methods, achieve significant improvement in terms of performance when there exists a large amount of data and the length of the data is evenly distributed.