Min Zhang, Yanqing Zhao, Xiaosong Qiao, Song Peng, Shimin Tao, Hao Yang, Ying Qin, Yanfei Jiang
{"title":"FastSpanNER: Speeding up SpanNER by Named Entity Head Prediction","authors":"Min Zhang, Yanqing Zhao, Xiaosong Qiao, Song Peng, Shimin Tao, Hao Yang, Ying Qin, Yanfei Jiang","doi":"10.1109/ICNLP58431.2023.00042","DOIUrl":null,"url":null,"abstract":"Named Entity Recognition (NER) is one of the most fundamental tasks in natural language processing (NLP). Different from the widely-used sequence labeling framework in NER, span prediction based methods are more naturally suitable for the nested NER problem and have received a lot of attention recently. However, classifying the samples generated by traversing all sub-sequences is computational expensive during training and very ineffective at inference. In this paper, we propose the FastSpanNER approach to reduce the computation of both training and inferring. We introduce a task of Named Entity Head (NEH) prediction for each word in given sequence, and perform multi-task learning together with the task of span classification, which uses no more than half of the samples in SpanNER. In the inference phase, only the words predicted as NEHs are used to generate candidate spans for named entity classification. Experimental results on the four standard benchmark datasets (CoNLL2003, MSRA, CNERTA and GENIA) show that our FastSpanNER method not only greatly reduces the computation of training and inferring but also achieves better F1 scores compared with the SpanNER method.","PeriodicalId":53637,"journal":{"name":"Icon","volume":"199 1","pages":"198-202"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Icon","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICNLP58431.2023.00042","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Arts and Humanities","Score":null,"Total":0}
引用次数: 0
Abstract
Named Entity Recognition (NER) is one of the most fundamental tasks in natural language processing (NLP). Different from the widely-used sequence labeling framework in NER, span prediction based methods are more naturally suitable for the nested NER problem and have received a lot of attention recently. However, classifying the samples generated by traversing all sub-sequences is computational expensive during training and very ineffective at inference. In this paper, we propose the FastSpanNER approach to reduce the computation of both training and inferring. We introduce a task of Named Entity Head (NEH) prediction for each word in given sequence, and perform multi-task learning together with the task of span classification, which uses no more than half of the samples in SpanNER. In the inference phase, only the words predicted as NEHs are used to generate candidate spans for named entity classification. Experimental results on the four standard benchmark datasets (CoNLL2003, MSRA, CNERTA and GENIA) show that our FastSpanNER method not only greatly reduces the computation of training and inferring but also achieves better F1 scores compared with the SpanNER method.