Yiqi Tong, Fuzhen Zhuang, Deqing Wang, Haochao Ying, Binling Wang
{"title":"Improving Biomedical Named Entity Recognition with a Unified Multi-Task MRC Framework","authors":"Yiqi Tong, Fuzhen Zhuang, Deqing Wang, Haochao Ying, Binling Wang","doi":"10.1109/ICASSP43922.2022.9746482","DOIUrl":null,"url":null,"abstract":"The prior knowledge, such as expert rules and knowledge base, has been proven effective in the traditional Biomedical Named Entity Recognition (BioNER). Most current neural BioNER systems use this external knowledge for pre-processing or post-editing instead of incorporate it into the training process, which cannot be learned by the model. To encode prior knowledge into the model, we present a unified multi-task Machine Reading Comprehension (MRC) framework for BioNER. Specifically, in the MRC task, the question sequences are derived from the standard BioNER dataset. We introduce three kinds of prior knowledge at query sequences, including Wikipedia, annotation scheme, entity dictionary. Then, our model adopts a multi-task learning strategy to joint training the main task BioNER and the auxiliary task MRC. Finally, experimental results on three benchmark datasets validate the superiority of our BioNER model compared with various state-of-the-art baselines.","PeriodicalId":272439,"journal":{"name":"ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"62 4","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICASSP43922.2022.9746482","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
The prior knowledge, such as expert rules and knowledge base, has been proven effective in the traditional Biomedical Named Entity Recognition (BioNER). Most current neural BioNER systems use this external knowledge for pre-processing or post-editing instead of incorporate it into the training process, which cannot be learned by the model. To encode prior knowledge into the model, we present a unified multi-task Machine Reading Comprehension (MRC) framework for BioNER. Specifically, in the MRC task, the question sequences are derived from the standard BioNER dataset. We introduce three kinds of prior knowledge at query sequences, including Wikipedia, annotation scheme, entity dictionary. Then, our model adopts a multi-task learning strategy to joint training the main task BioNER and the auxiliary task MRC. Finally, experimental results on three benchmark datasets validate the superiority of our BioNER model compared with various state-of-the-art baselines.