{"title":"基于新型BERT预训练的中文文本分类研究","authors":"Youyao Liu, Haimei Huang, Jialei Gao, Shihao Gai","doi":"10.1109/ICNLP58431.2023.00062","DOIUrl":null,"url":null,"abstract":"Chinese Text Classification (TC) is the process of mapping text to a pre-given topics category. In recent years, it has been found that TC is mainly based on RNN and BERT, so the development of different novel pre-training applied to Chinese TC is described as based on BERT pre-training. BERT combined with convolutional neural network is proposed to extend the BERT-CNN model for the problem of lack of semantic knowledge of BERT to derive a good classification effect. The second RoBERTa model performs feature extraction and fusion to obtain the feature word vector as the text output vector, which can solve the problem of insufficient BERT extracted features. The BERT-BiGRU model, on the other hand, mainly avoids the increase in the number of texts leading to long training time and overfitting, and uses a simpler GRU bi-word network as the main network to fully extract the contextual information of Chinese texts and finally complete the classification of Chinese texts; at the same time, it makes an outlook and conclusion on the new pre-training model for Chinese TC.","PeriodicalId":53637,"journal":{"name":"Icon","volume":"1 1","pages":"303-307"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A study of Chinese Text Classification based on a new type of BERT pre-training\",\"authors\":\"Youyao Liu, Haimei Huang, Jialei Gao, Shihao Gai\",\"doi\":\"10.1109/ICNLP58431.2023.00062\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Chinese Text Classification (TC) is the process of mapping text to a pre-given topics category. In recent years, it has been found that TC is mainly based on RNN and BERT, so the development of different novel pre-training applied to Chinese TC is described as based on BERT pre-training. BERT combined with convolutional neural network is proposed to extend the BERT-CNN model for the problem of lack of semantic knowledge of BERT to derive a good classification effect. The second RoBERTa model performs feature extraction and fusion to obtain the feature word vector as the text output vector, which can solve the problem of insufficient BERT extracted features. The BERT-BiGRU model, on the other hand, mainly avoids the increase in the number of texts leading to long training time and overfitting, and uses a simpler GRU bi-word network as the main network to fully extract the contextual information of Chinese texts and finally complete the classification of Chinese texts; at the same time, it makes an outlook and conclusion on the new pre-training model for Chinese TC.\",\"PeriodicalId\":53637,\"journal\":{\"name\":\"Icon\",\"volume\":\"1 1\",\"pages\":\"303-307\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Icon\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICNLP58431.2023.00062\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"Arts and Humanities\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Icon","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICNLP58431.2023.00062","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Arts and Humanities","Score":null,"Total":0}
A study of Chinese Text Classification based on a new type of BERT pre-training
Chinese Text Classification (TC) is the process of mapping text to a pre-given topics category. In recent years, it has been found that TC is mainly based on RNN and BERT, so the development of different novel pre-training applied to Chinese TC is described as based on BERT pre-training. BERT combined with convolutional neural network is proposed to extend the BERT-CNN model for the problem of lack of semantic knowledge of BERT to derive a good classification effect. The second RoBERTa model performs feature extraction and fusion to obtain the feature word vector as the text output vector, which can solve the problem of insufficient BERT extracted features. The BERT-BiGRU model, on the other hand, mainly avoids the increase in the number of texts leading to long training time and overfitting, and uses a simpler GRU bi-word network as the main network to fully extract the contextual information of Chinese texts and finally complete the classification of Chinese texts; at the same time, it makes an outlook and conclusion on the new pre-training model for Chinese TC.