{"title":"一种基于注意机制的字词融合中文命名实体识别模型","authors":"Jinshang Luo, Xinchun Zou, Mengshu Hou","doi":"10.1109/CCET55412.2022.9906333","DOIUrl":null,"url":null,"abstract":"Named Entity Recognition (NER) is a fundamental task in natural language processing. Compared with English NER, the difficulty of Chinese NER lies in word segmentation ambiguity and polysemy. Aiming at the issue, a novel character-word fusion Long Short-Term Memory (LSTM) model combined with the sentence-level attention mechanism (CWSA-LSTM) is proposed. Firstly, the method encodes the representations of characters and words through the pretrained models. The word information is incorporated into the character sequence by matching the potential word with a lexicon. Then the feature vectors are fed into the LSTM layer to learn contextual information. The attention mechanism is utilized to capture the tightness of the correlation in the sentence. Experiments on benchmark datasets demonstrate that CWSA-LSTM outperforms other state-of-the-art methods, and verify the effectiveness of character-word fusion. For the MSRA dataset, CWSA-LSTM achieves a 2.46% improvement in F1 score over baseline Lattice LSTM.","PeriodicalId":329327,"journal":{"name":"2022 IEEE 5th International Conference on Computer and Communication Engineering Technology (CCET)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A Novel Character-Word Fusion Chinese Named Entity Recognition Model Based on Attention Mechanism\",\"authors\":\"Jinshang Luo, Xinchun Zou, Mengshu Hou\",\"doi\":\"10.1109/CCET55412.2022.9906333\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Named Entity Recognition (NER) is a fundamental task in natural language processing. Compared with English NER, the difficulty of Chinese NER lies in word segmentation ambiguity and polysemy. Aiming at the issue, a novel character-word fusion Long Short-Term Memory (LSTM) model combined with the sentence-level attention mechanism (CWSA-LSTM) is proposed. Firstly, the method encodes the representations of characters and words through the pretrained models. The word information is incorporated into the character sequence by matching the potential word with a lexicon. Then the feature vectors are fed into the LSTM layer to learn contextual information. The attention mechanism is utilized to capture the tightness of the correlation in the sentence. Experiments on benchmark datasets demonstrate that CWSA-LSTM outperforms other state-of-the-art methods, and verify the effectiveness of character-word fusion. For the MSRA dataset, CWSA-LSTM achieves a 2.46% improvement in F1 score over baseline Lattice LSTM.\",\"PeriodicalId\":329327,\"journal\":{\"name\":\"2022 IEEE 5th International Conference on Computer and Communication Engineering Technology (CCET)\",\"volume\":\"67 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-08-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE 5th International Conference on Computer and Communication Engineering Technology (CCET)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CCET55412.2022.9906333\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 5th International Conference on Computer and Communication Engineering Technology (CCET)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCET55412.2022.9906333","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Novel Character-Word Fusion Chinese Named Entity Recognition Model Based on Attention Mechanism
Named Entity Recognition (NER) is a fundamental task in natural language processing. Compared with English NER, the difficulty of Chinese NER lies in word segmentation ambiguity and polysemy. Aiming at the issue, a novel character-word fusion Long Short-Term Memory (LSTM) model combined with the sentence-level attention mechanism (CWSA-LSTM) is proposed. Firstly, the method encodes the representations of characters and words through the pretrained models. The word information is incorporated into the character sequence by matching the potential word with a lexicon. Then the feature vectors are fed into the LSTM layer to learn contextual information. The attention mechanism is utilized to capture the tightness of the correlation in the sentence. Experiments on benchmark datasets demonstrate that CWSA-LSTM outperforms other state-of-the-art methods, and verify the effectiveness of character-word fusion. For the MSRA dataset, CWSA-LSTM achieves a 2.46% improvement in F1 score over baseline Lattice LSTM.