{"title":"从多语言BERT模型到单语言BERT模型的越南语问答系统","authors":"Nguyen Thi Mai Trang, M. Shcherbakov","doi":"10.1109/SMART50582.2020.9337155","DOIUrl":null,"url":null,"abstract":"A question answering (QA) system based on natural language processing and deep learning gets more attention from AI communities. Many companies and organizations are interested in developing automated question answering systems which are being researched widely. Recently, the new model named Bidirectional Encoder Representation from Transformer (BERT) was proposed to solve the restrictions of NLP tasks. BERT achieved the best results in almost tasks that include QA tasks. In this work, we tried applying the multilingual BERT models (multilingual BERT [1], DeepPavlov multilingual BERT, multilingual BERT fine-tuned on XQuAD) and the language-specific BERT model for Vietnamese (PhoBERT). The obtained result has shown that the monolingual model outperforms the multilingual models. We also recommend multilingual BERT fine-tuned on XQuAD model as an option to build a Vietnamese QA system if the system is built from a multilingual BERT based model.","PeriodicalId":129946,"journal":{"name":"2020 9th International Conference System Modeling and Advancement in Research Trends (SMART)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":"{\"title\":\"Vietnamese Question Answering System f rom Multilingual BERT Models to Monolingual BERT Model\",\"authors\":\"Nguyen Thi Mai Trang, M. Shcherbakov\",\"doi\":\"10.1109/SMART50582.2020.9337155\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A question answering (QA) system based on natural language processing and deep learning gets more attention from AI communities. Many companies and organizations are interested in developing automated question answering systems which are being researched widely. Recently, the new model named Bidirectional Encoder Representation from Transformer (BERT) was proposed to solve the restrictions of NLP tasks. BERT achieved the best results in almost tasks that include QA tasks. In this work, we tried applying the multilingual BERT models (multilingual BERT [1], DeepPavlov multilingual BERT, multilingual BERT fine-tuned on XQuAD) and the language-specific BERT model for Vietnamese (PhoBERT). The obtained result has shown that the monolingual model outperforms the multilingual models. We also recommend multilingual BERT fine-tuned on XQuAD model as an option to build a Vietnamese QA system if the system is built from a multilingual BERT based model.\",\"PeriodicalId\":129946,\"journal\":{\"name\":\"2020 9th International Conference System Modeling and Advancement in Research Trends (SMART)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-12-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"7\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 9th International Conference System Modeling and Advancement in Research Trends (SMART)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SMART50582.2020.9337155\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 9th International Conference System Modeling and Advancement in Research Trends (SMART)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SMART50582.2020.9337155","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Vietnamese Question Answering System f rom Multilingual BERT Models to Monolingual BERT Model
A question answering (QA) system based on natural language processing and deep learning gets more attention from AI communities. Many companies and organizations are interested in developing automated question answering systems which are being researched widely. Recently, the new model named Bidirectional Encoder Representation from Transformer (BERT) was proposed to solve the restrictions of NLP tasks. BERT achieved the best results in almost tasks that include QA tasks. In this work, we tried applying the multilingual BERT models (multilingual BERT [1], DeepPavlov multilingual BERT, multilingual BERT fine-tuned on XQuAD) and the language-specific BERT model for Vietnamese (PhoBERT). The obtained result has shown that the monolingual model outperforms the multilingual models. We also recommend multilingual BERT fine-tuned on XQuAD model as an option to build a Vietnamese QA system if the system is built from a multilingual BERT based model.