{"title":"分类灾难相关推文的混合模型","authors":"Nayan Ranjan Paul, Deepak Sahoo, R. Balabantaray","doi":"10.1109/OCIT56763.2022.00021","DOIUrl":null,"url":null,"abstract":"When it comes to classifying tweets about disasters, Deep Neural Network-based models have shown great potential over conventional machine learning models. In particular, Bidirectional Encoder Representations from Transformers(BERT) are effectively used to capture the contextual information present in a tweet. But it is not effectively capturing the global structural information of tweets. A Graph Convolutional Network's(GCN) power lies in its ability to capture global information. In this study, we present a novel hybrid model called VocabGCN-BERT by combining the GCN made from a vocabulary graph of tweets and the pre-trained BERT model. A powerful representation for classifying tweets is created by combining local contextual information acquired from BERT with global structural information acquired from VocabGCN. The results of the experiments demonstrate that the proposed VocabGCN-BERT performs better than the currently available state-of-art models based on GCN on seven publicly available datasets by a margin of +1.66% to +4.79% in weighted average F1 score and +1.45% to +4.34% in accuracy.","PeriodicalId":425541,"journal":{"name":"2022 OITS International Conference on Information Technology (OCIT)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"VocabGCN-BERT: A Hybrid Model to Classify Disaster Related Tweets\",\"authors\":\"Nayan Ranjan Paul, Deepak Sahoo, R. Balabantaray\",\"doi\":\"10.1109/OCIT56763.2022.00021\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"When it comes to classifying tweets about disasters, Deep Neural Network-based models have shown great potential over conventional machine learning models. In particular, Bidirectional Encoder Representations from Transformers(BERT) are effectively used to capture the contextual information present in a tweet. But it is not effectively capturing the global structural information of tweets. A Graph Convolutional Network's(GCN) power lies in its ability to capture global information. In this study, we present a novel hybrid model called VocabGCN-BERT by combining the GCN made from a vocabulary graph of tweets and the pre-trained BERT model. A powerful representation for classifying tweets is created by combining local contextual information acquired from BERT with global structural information acquired from VocabGCN. The results of the experiments demonstrate that the proposed VocabGCN-BERT performs better than the currently available state-of-art models based on GCN on seven publicly available datasets by a margin of +1.66% to +4.79% in weighted average F1 score and +1.45% to +4.34% in accuracy.\",\"PeriodicalId\":425541,\"journal\":{\"name\":\"2022 OITS International Conference on Information Technology (OCIT)\",\"volume\":\"35 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 OITS International Conference on Information Technology (OCIT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/OCIT56763.2022.00021\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 OITS International Conference on Information Technology (OCIT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/OCIT56763.2022.00021","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
VocabGCN-BERT: A Hybrid Model to Classify Disaster Related Tweets
When it comes to classifying tweets about disasters, Deep Neural Network-based models have shown great potential over conventional machine learning models. In particular, Bidirectional Encoder Representations from Transformers(BERT) are effectively used to capture the contextual information present in a tweet. But it is not effectively capturing the global structural information of tweets. A Graph Convolutional Network's(GCN) power lies in its ability to capture global information. In this study, we present a novel hybrid model called VocabGCN-BERT by combining the GCN made from a vocabulary graph of tweets and the pre-trained BERT model. A powerful representation for classifying tweets is created by combining local contextual information acquired from BERT with global structural information acquired from VocabGCN. The results of the experiments demonstrate that the proposed VocabGCN-BERT performs better than the currently available state-of-art models based on GCN on seven publicly available datasets by a margin of +1.66% to +4.79% in weighted average F1 score and +1.45% to +4.34% in accuracy.