{"title":"Hierarchical Attention-based BiLSTM Network for Document Similarity Calculation","authors":"Jiang Zhang, Qun Zhu, Yanlin He","doi":"10.1145/3440084.3441188","DOIUrl":null,"url":null,"abstract":"Neural network model is a momentous method to calculate semantic similarity. Taking into account the complexity of document structure, introducing hierarchical structure and attention mechanism into neural network can calculate document semantic representation more precisely. In order to verify the validity of the model, LP50 dataset was tested. The experimental results reveal that accurate document representation can be obtained by using the attention mechanism at two levels of words and sentences. Since this method has taken both the influence of context information and the contribution of components to the document into consideration. Compared with several conventional methods, there is a significant improvement of performance in our model.","PeriodicalId":250100,"journal":{"name":"Proceedings of the 2020 4th International Symposium on Computer Science and Intelligent Control","volume":"2 6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2020 4th International Symposium on Computer Science and Intelligent Control","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3440084.3441188","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Neural network model is a momentous method to calculate semantic similarity. Taking into account the complexity of document structure, introducing hierarchical structure and attention mechanism into neural network can calculate document semantic representation more precisely. In order to verify the validity of the model, LP50 dataset was tested. The experimental results reveal that accurate document representation can be obtained by using the attention mechanism at two levels of words and sentences. Since this method has taken both the influence of context information and the contribution of components to the document into consideration. Compared with several conventional methods, there is a significant improvement of performance in our model.