Sentiment analysis on COVID tweets using COVID-Twitter-BERT with auxiliary sentence approach

Hung-Yeh Lin, Teng-Sheng Moh
{"title":"Sentiment analysis on COVID tweets using COVID-Twitter-BERT with auxiliary sentence approach","authors":"Hung-Yeh Lin, Teng-Sheng Moh","doi":"10.1145/3409334.3452074","DOIUrl":null,"url":null,"abstract":"Sentiment analysis is a fascinating area as a natural language understanding benchmark to evaluate customers' feedback and needs. Moreover, sentiment analysis can be applied to understand the people's reactions to public events such as the presidential elections and disease pandemics. Recent works in sentiment analysis on COVID-19 present a domain-targeted Bidirectional Encoder Representations from Transformer (BERT) language model, COVID-Twitter BERT (CT-BERT). However, there is little improvement in text classification using a BERT-based language model directly. Therefore, an auxiliary approach using BERT was proposed. This method converts single-sentence classification into pair-sentence classification, which solves the performance issue of BERT in text classification tasks. In this paper, we combine a pre-trained BERT model from COVID-related tweets and the auxiliary-sentence method to achieve better classification performance on COVID tweets sentiment analysis. We show that converting single-sentence classification into pair-sentence classification extends the dataset and obtains higher accuracies and F1 scores. However, we expect a domain-specific language model would perform better than a general language model. In our results, we show that the performance of CT-BERT does not necessarily outperform BERT specifically in understanding sentiments.","PeriodicalId":148741,"journal":{"name":"Proceedings of the 2021 ACM Southeast Conference","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2021 ACM Southeast Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3409334.3452074","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6

Abstract

Sentiment analysis is a fascinating area as a natural language understanding benchmark to evaluate customers' feedback and needs. Moreover, sentiment analysis can be applied to understand the people's reactions to public events such as the presidential elections and disease pandemics. Recent works in sentiment analysis on COVID-19 present a domain-targeted Bidirectional Encoder Representations from Transformer (BERT) language model, COVID-Twitter BERT (CT-BERT). However, there is little improvement in text classification using a BERT-based language model directly. Therefore, an auxiliary approach using BERT was proposed. This method converts single-sentence classification into pair-sentence classification, which solves the performance issue of BERT in text classification tasks. In this paper, we combine a pre-trained BERT model from COVID-related tweets and the auxiliary-sentence method to achieve better classification performance on COVID tweets sentiment analysis. We show that converting single-sentence classification into pair-sentence classification extends the dataset and obtains higher accuracies and F1 scores. However, we expect a domain-specific language model would perform better than a general language model. In our results, we show that the performance of CT-BERT does not necessarily outperform BERT specifically in understanding sentiments.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
使用COVID- twitter - bert辅助句方法对COVID推文进行情感分析
情感分析是一个很有吸引力的领域,它可以作为自然语言理解的基准来评估客户的反馈和需求。此外,还可以运用情绪分析来了解国民对总统选举、传染病等公共事件的反应。最近关于COVID-19情绪分析的工作提出了一种面向领域的双向编码器表示转换器(BERT)语言模型,即COVID-Twitter BERT (CT-BERT)。然而,直接使用基于bert的语言模型在文本分类方面几乎没有改进。因此,提出了一种基于BERT的辅助方法。该方法将单句分类转化为对句分类,解决了BERT在文本分类任务中的性能问题。在本文中,我们将来自COVID相关推文的预训练BERT模型与辅助句方法相结合,以获得更好的COVID推文情感分析分类性能。我们表明,将单句分类转换为成对分类扩展了数据集,并获得了更高的准确率和F1分数。然而,我们期望特定于领域的语言模型比一般的语言模型表现得更好。在我们的结果中,我们表明CT-BERT的表现不一定优于BERT,特别是在理解情绪方面。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Application of back-translation: a transfer learning approach to identify ambiguous software requirements A survey of wireless network simulation and/or emulation software for use in higher education Implementing a network intrusion detection system using semi-supervised support vector machine and random forest Performance evaluation of a widely used implementation of the MQTT protocol with large payloads in normal operation and under a DoS attack Benefits of combining dimensional attention and working memory for partially observable reinforcement learning problems
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1