EMFSA: Emoji-based multifeature fusion sentiment analysis.

IF 2.9 3区 综合性期刊 Q1 MULTIDISCIPLINARY SCIENCES PLoS ONE Pub Date : 2024-09-19 eCollection Date: 2024-01-01 DOI:10.1371/journal.pone.0310715
Hongmei Tang, Wenzhong Tang, Dixiongxiao Zhu, Shuai Wang, Yanyang Wang, Lihong Wang
{"title":"EMFSA: Emoji-based multifeature fusion sentiment analysis.","authors":"Hongmei Tang, Wenzhong Tang, Dixiongxiao Zhu, Shuai Wang, Yanyang Wang, Lihong Wang","doi":"10.1371/journal.pone.0310715","DOIUrl":null,"url":null,"abstract":"<p><p>Short texts on social platforms often suffer from insufficient emotional semantic expressions, sparse features, and polysemy. To enhance the accuracy achieved by sentiment analysis for short texts, this paper proposes an emoji-based multifeature fusion sentiment analysis model (EMFSA). The model mines the sentiments of emojis, topics, and text features. Initially, a pretraining method for feature extraction is employed to enhance the semantic expressions of emotions in text by extracting contextual semantic information from emojis. Following this, a sentiment- and emoji-masked language model is designed to prioritize the masking of emojis and words with implicit sentiments, focusing on learning the emotional semantics contained in text. Additionally, we proposed a multifeature fusion method based on a cross-attention mechanism by determining the importance of each word in a text from a topic perspective. Next, this method is integrated with the original semantic information of emojis and the enhanced text features, attaining improved sentiment representation accuracy for short texts. Comparative experiments conducted with the state-of-the-art baseline methods on three public datasets demonstrate that the proposed model achieves accuracy improvements of 2.3%, 10.9%, and 2.7%, respectively, validating its effectiveness.</p>","PeriodicalId":20189,"journal":{"name":"PLoS ONE","volume":null,"pages":null},"PeriodicalIF":2.9000,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11412488/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"PLoS ONE","FirstCategoryId":"103","ListUrlMain":"https://doi.org/10.1371/journal.pone.0310715","RegionNum":3,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/1/1 0:00:00","PubModel":"eCollection","JCR":"Q1","JCRName":"MULTIDISCIPLINARY SCIENCES","Score":null,"Total":0}
引用次数: 0

Abstract

Short texts on social platforms often suffer from insufficient emotional semantic expressions, sparse features, and polysemy. To enhance the accuracy achieved by sentiment analysis for short texts, this paper proposes an emoji-based multifeature fusion sentiment analysis model (EMFSA). The model mines the sentiments of emojis, topics, and text features. Initially, a pretraining method for feature extraction is employed to enhance the semantic expressions of emotions in text by extracting contextual semantic information from emojis. Following this, a sentiment- and emoji-masked language model is designed to prioritize the masking of emojis and words with implicit sentiments, focusing on learning the emotional semantics contained in text. Additionally, we proposed a multifeature fusion method based on a cross-attention mechanism by determining the importance of each word in a text from a topic perspective. Next, this method is integrated with the original semantic information of emojis and the enhanced text features, attaining improved sentiment representation accuracy for short texts. Comparative experiments conducted with the state-of-the-art baseline methods on three public datasets demonstrate that the proposed model achieves accuracy improvements of 2.3%, 10.9%, and 2.7%, respectively, validating its effectiveness.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
EMFSA:基于表情符号的多特征融合情感分析。
社交平台上的短文往往存在情感语义表达不足、特征稀疏和多义性等问题。为了提高短文情感分析的准确性,本文提出了一种基于表情符号的多特征融合情感分析模型(EMFSA)。该模型挖掘了表情符号、话题和文本特征的情感。首先,采用特征提取预训练方法,通过从表情符号中提取上下文语义信息来增强文本中的情感语义表达。随后,我们设计了一个情感和表情符号屏蔽语言模型,优先屏蔽表情符号和带有隐含情感的词语,重点学习文本中包含的情感语义。此外,我们还提出了一种基于交叉关注机制的多特征融合方法,从主题角度确定文本中每个词的重要性。接下来,我们将这种方法与表情符号的原始语义信息和增强的文本特征相融合,从而提高了短文的情感表征准确率。在三个公共数据集上与最先进的基线方法进行的对比实验表明,所提出的模型的准确率分别提高了 2.3%、10.9% 和 2.7%,验证了其有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
PLoS ONE
PLoS ONE 生物-生物学
CiteScore
6.20
自引率
5.40%
发文量
14242
审稿时长
3.7 months
期刊介绍: PLOS ONE is an international, peer-reviewed, open-access, online publication. PLOS ONE welcomes reports on primary research from any scientific discipline. It provides: * Open-access—freely accessible online, authors retain copyright * Fast publication times * Peer review by expert, practicing researchers * Post-publication tools to indicate quality and impact * Community-based dialogue on articles * Worldwide media coverage
期刊最新文献
"What does this mean for our future?" uncertainty management in mothers' narratives about the diagnosis and birth of their child with Down syndrome. A multimodal understanding of the role of sound and music in gendered toy marketing. Evaluation of the Vaginal Panel Realtime PCR kit (Vircell, SL) for diagnosing vaginitis: A comparative study with routinely used diagnostics. Fault-tolerant partition resolvability of cycle with chord. Few-shot learning for inference in medical imaging with subspace feature representations.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1