基于预训练模型和多头注意机制的命名实体识别

Q3 Arts and Humanities Icon Pub Date : 2023-03-01 DOI:10.1109/icnlp58431.2023.00040
GuoHua Zhu, Jian Wang
{"title":"基于预训练模型和多头注意机制的命名实体识别","authors":"GuoHua Zhu, Jian Wang","doi":"10.1109/icnlp58431.2023.00040","DOIUrl":null,"url":null,"abstract":"When processing Chinese named entity recognition, the traditional algorithm model have been having the ambiguity of word segmentation and the singleness of the word vector, and the training consequence of algorithm models was not well. To solve this problem, a BERT-BiLSTM Multi-Attention (PMA-CNER) model was proposed to improve the accuracy of Chinese named entity recognition (CNER). This model used BERT model to embed words based on BiLSTM model, which can extract global context semantic features more effectively. Next, a layer of Multi-head attention mechanism was added behind the BiLSTM layer, which can effectively extract multiple semantic features and overcome the shortage of BiLSTM in obtaining local features. Finally, the experimental results on the CLUSER2020 dataset and the Yudu-S4K dataset show that the accuracy rate is significantly improved, reaching 93.94% and 91.83% respectively.","PeriodicalId":53637,"journal":{"name":"Icon","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Named Entity Recognition Based on Pre-training Model and Multi-head Attention Mechanism\",\"authors\":\"GuoHua Zhu, Jian Wang\",\"doi\":\"10.1109/icnlp58431.2023.00040\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"When processing Chinese named entity recognition, the traditional algorithm model have been having the ambiguity of word segmentation and the singleness of the word vector, and the training consequence of algorithm models was not well. To solve this problem, a BERT-BiLSTM Multi-Attention (PMA-CNER) model was proposed to improve the accuracy of Chinese named entity recognition (CNER). This model used BERT model to embed words based on BiLSTM model, which can extract global context semantic features more effectively. Next, a layer of Multi-head attention mechanism was added behind the BiLSTM layer, which can effectively extract multiple semantic features and overcome the shortage of BiLSTM in obtaining local features. Finally, the experimental results on the CLUSER2020 dataset and the Yudu-S4K dataset show that the accuracy rate is significantly improved, reaching 93.94% and 91.83% respectively.\",\"PeriodicalId\":53637,\"journal\":{\"name\":\"Icon\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Icon\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/icnlp58431.2023.00040\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"Arts and Humanities\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Icon","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/icnlp58431.2023.00040","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Arts and Humanities","Score":null,"Total":0}
引用次数: 0

摘要

在处理中文命名实体识别时,传统算法模型存在分词模糊和词向量单一的问题,算法模型的训练效果不佳。为了解决这一问题,提出了BERT-BiLSTM多注意(PMA-CNER)模型,以提高中文命名实体识别(CNER)的准确率。该模型在BiLSTM模型的基础上使用BERT模型嵌入词,可以更有效地提取全局上下文语义特征。其次,在BiLSTM层后面增加了一层多头注意机制,可以有效地提取多个语义特征,克服了BiLSTM在获取局部特征方面的不足;最后,在CLUSER2020数据集和Yudu-S4K数据集上的实验结果表明,准确率显著提高,分别达到93.94%和91.83%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Named Entity Recognition Based on Pre-training Model and Multi-head Attention Mechanism
When processing Chinese named entity recognition, the traditional algorithm model have been having the ambiguity of word segmentation and the singleness of the word vector, and the training consequence of algorithm models was not well. To solve this problem, a BERT-BiLSTM Multi-Attention (PMA-CNER) model was proposed to improve the accuracy of Chinese named entity recognition (CNER). This model used BERT model to embed words based on BiLSTM model, which can extract global context semantic features more effectively. Next, a layer of Multi-head attention mechanism was added behind the BiLSTM layer, which can effectively extract multiple semantic features and overcome the shortage of BiLSTM in obtaining local features. Finally, the experimental results on the CLUSER2020 dataset and the Yudu-S4K dataset show that the accuracy rate is significantly improved, reaching 93.94% and 91.83% respectively.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Icon
Icon Arts and Humanities-History and Philosophy of Science
CiteScore
0.30
自引率
0.00%
发文量
0
期刊最新文献
Long-term Coherent Accumulation Algorithm Based on Radar Altimeter Deep Composite Kernels ELM Based on Spatial Feature Extraction for Hyperspectral Vegetation Image Classification Research based on improved SSD target detection algorithm CON-GAN-BERT: combining Contrastive Learning with Generative Adversarial Nets for Few-Shot Sentiment Classification A Two Stage Learning Algorithm for Hyperspectral Image Classification
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1