One Improved Model of Named Entity Recognition by Combining BERT and BiLSTM-CNN for Domain of Chinese Railway Construction

Xiaojun Wu, Tianqi Zhang, Sheng Yuan, Yuanfeng Yan
{"title":"One Improved Model of Named Entity Recognition by Combining BERT and BiLSTM-CNN for Domain of Chinese Railway Construction","authors":"Xiaojun Wu, Tianqi Zhang, Sheng Yuan, Yuanfeng Yan","doi":"10.1109/ICSP54964.2022.9778794","DOIUrl":null,"url":null,"abstract":"There are currently few named entity recognition (NER) models in domain of Chinese railway construction. To mitigate such awkward situation, this paper uses the neural network method to sort out the basic information from Chinese text about Chinese railway construction. Concretely, this paper proposes one improved model of NER by combining bidirectional encoder representation from transformers (BERT) and convolutional long short-term memory (LSTM) network model so as to promote the NER performance of Chinese text about Chinese railway construction. Based on deep understandings of domain knowledge about Chinese railway construction, the proposed model performs targeted processing on the input, and designs a novel masking algorithm based on Chinese placenames and numbers. The proposed model further uses bidirectional LSTM (BiLSTM) network as the encoding layer, which can leverage the feature extraction capability of the convolution neural network (CNN) to improve the NER performance. Experimental results show that the F1 value of the proposed model is 7.28% higher than the traditional conditional random field (CRF) model, and the F1 value of the BERT model with mask of Chinese placenames and numbers is 3.43% higher than the original BERT model.","PeriodicalId":363766,"journal":{"name":"2022 7th International Conference on Intelligent Computing and Signal Processing (ICSP)","volume":"120 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 7th International Conference on Intelligent Computing and Signal Processing (ICSP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSP54964.2022.9778794","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

Abstract

There are currently few named entity recognition (NER) models in domain of Chinese railway construction. To mitigate such awkward situation, this paper uses the neural network method to sort out the basic information from Chinese text about Chinese railway construction. Concretely, this paper proposes one improved model of NER by combining bidirectional encoder representation from transformers (BERT) and convolutional long short-term memory (LSTM) network model so as to promote the NER performance of Chinese text about Chinese railway construction. Based on deep understandings of domain knowledge about Chinese railway construction, the proposed model performs targeted processing on the input, and designs a novel masking algorithm based on Chinese placenames and numbers. The proposed model further uses bidirectional LSTM (BiLSTM) network as the encoding layer, which can leverage the feature extraction capability of the convolution neural network (CNN) to improve the NER performance. Experimental results show that the F1 value of the proposed model is 7.28% higher than the traditional conditional random field (CRF) model, and the F1 value of the BERT model with mask of Chinese placenames and numbers is 3.43% higher than the original BERT model.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
结合BERT和BiLSTM-CNN的中国铁路建设领域命名实体识别改进模型
目前,中国铁路建设领域的命名实体识别(NER)模型很少。为了缓解这一尴尬局面,本文采用神经网络方法对中国铁路建设中文文本中的基本信息进行了梳理。具体而言,本文提出了一种结合变压器双向编码器表示(BERT)和卷积长短期记忆(LSTM)网络模型的改进神经网络模型,以提高中国铁路建设中文文本的神经网络性能。该模型基于对中国铁路建设领域知识的深入理解,对输入进行有针对性的处理,设计了一种基于中文地名和数字的掩蔽算法。该模型进一步采用双向LSTM (BiLSTM)网络作为编码层,利用卷积神经网络(CNN)的特征提取能力来提高NER性能。实验结果表明,该模型的F1值比传统的条件随机场(CRF)模型高7.28%,带中文地名和数字掩码的BERT模型的F1值比原BERT模型高3.43%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Research on Retailer Churn Prediction Based on Spatial-Temporal Features Non-sinusoidal harmonic signal detection method for energy meter measurement Deep Intra-Class Similarity Measured Semi-Supervised Learning Adaptive Persymmetric Subspace Detector for Distributed Target Deblurring Reconstruction of Monitoring Video in Smart Grid Based on Depth-wise Separable Convolutional Neural Network
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1