{"title":"Named Entity Recognition Based on Pre-training Model and Multi-head Attention Mechanism","authors":"GuoHua Zhu, Jian Wang","doi":"10.1109/icnlp58431.2023.00040","DOIUrl":null,"url":null,"abstract":"When processing Chinese named entity recognition, the traditional algorithm model have been having the ambiguity of word segmentation and the singleness of the word vector, and the training consequence of algorithm models was not well. To solve this problem, a BERT-BiLSTM Multi-Attention (PMA-CNER) model was proposed to improve the accuracy of Chinese named entity recognition (CNER). This model used BERT model to embed words based on BiLSTM model, which can extract global context semantic features more effectively. Next, a layer of Multi-head attention mechanism was added behind the BiLSTM layer, which can effectively extract multiple semantic features and overcome the shortage of BiLSTM in obtaining local features. Finally, the experimental results on the CLUSER2020 dataset and the Yudu-S4K dataset show that the accuracy rate is significantly improved, reaching 93.94% and 91.83% respectively.","PeriodicalId":53637,"journal":{"name":"Icon","volume":"36 1","pages":"187-190"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Icon","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/icnlp58431.2023.00040","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Arts and Humanities","Score":null,"Total":0}
引用次数: 0
Abstract
When processing Chinese named entity recognition, the traditional algorithm model have been having the ambiguity of word segmentation and the singleness of the word vector, and the training consequence of algorithm models was not well. To solve this problem, a BERT-BiLSTM Multi-Attention (PMA-CNER) model was proposed to improve the accuracy of Chinese named entity recognition (CNER). This model used BERT model to embed words based on BiLSTM model, which can extract global context semantic features more effectively. Next, a layer of Multi-head attention mechanism was added behind the BiLSTM layer, which can effectively extract multiple semantic features and overcome the shortage of BiLSTM in obtaining local features. Finally, the experimental results on the CLUSER2020 dataset and the Yudu-S4K dataset show that the accuracy rate is significantly improved, reaching 93.94% and 91.83% respectively.