Detecting fake review intentions in the review context: A multimodal deep learning approach

IF 6.3 3区 管理学 Q1 BUSINESS Electronic Commerce Research and Applications Pub Date : 2025-03-01 Epub Date: 2025-02-05 DOI:10.1016/j.elerap.2025.101485
Jingrui Hou , Zhihang Tan , Shitou Zhang , Qibiao Hu , Ping Wang
{"title":"Detecting fake review intentions in the review context: A multimodal deep learning approach","authors":"Jingrui Hou ,&nbsp;Zhihang Tan ,&nbsp;Shitou Zhang ,&nbsp;Qibiao Hu ,&nbsp;Ping Wang","doi":"10.1016/j.elerap.2025.101485","DOIUrl":null,"url":null,"abstract":"<div><div>The proliferation of fake reviews on the internet has had significant repercussions for both consumers and businesses. However, existing research predominantly employs a binary classification approach to ascertain review authenticity, often neglecting the rich multimodal context information and nuanced intentions embedded within them. To bridge this gap, our study introduces a novel task, Fake Review Intention Detection in Review Context (FRIDRC), which aims to detect fake review intentions by leveraging both textual and visual information, and constructs a dataset comprising both manually and AI-generated fake reviews. Additionally, we develop a predictive framework encompassing modules for multimodal representation and modality fusion. These modules, while independent, are synergistic and effectively tackle the challenge of discerning fake review intentions. Our framework demonstrates outstanding performance, achieving an average F1 score exceeding 0.97 and a Macro F1 score surpassing 0.96 in this task and outperforming advanced pre-trained models. This research not only presents an effective methodology for accurately identifying and addressing fake review intentions but also underscores the efficacy of leveraging multimodal review context information in fake review detection. The dataset and code implementation are publicly available for further research.</div></div>","PeriodicalId":50541,"journal":{"name":"Electronic Commerce Research and Applications","volume":"70 ","pages":"Article 101485"},"PeriodicalIF":6.3000,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Electronic Commerce Research and Applications","FirstCategoryId":"91","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1567422325000109","RegionNum":3,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/2/5 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"BUSINESS","Score":null,"Total":0}
引用次数: 0

Abstract

The proliferation of fake reviews on the internet has had significant repercussions for both consumers and businesses. However, existing research predominantly employs a binary classification approach to ascertain review authenticity, often neglecting the rich multimodal context information and nuanced intentions embedded within them. To bridge this gap, our study introduces a novel task, Fake Review Intention Detection in Review Context (FRIDRC), which aims to detect fake review intentions by leveraging both textual and visual information, and constructs a dataset comprising both manually and AI-generated fake reviews. Additionally, we develop a predictive framework encompassing modules for multimodal representation and modality fusion. These modules, while independent, are synergistic and effectively tackle the challenge of discerning fake review intentions. Our framework demonstrates outstanding performance, achieving an average F1 score exceeding 0.97 and a Macro F1 score surpassing 0.96 in this task and outperforming advanced pre-trained models. This research not only presents an effective methodology for accurately identifying and addressing fake review intentions but also underscores the efficacy of leveraging multimodal review context information in fake review detection. The dataset and code implementation are publicly available for further research.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
在评论语境中检测虚假评论意图:一种多模态深度学习方法
互联网上虚假评论的泛滥对消费者和企业都产生了重大影响。然而,现有的研究主要采用二元分类方法来确定评论的真实性,往往忽略了丰富的多模态上下文信息和其中隐含的微妙意图。为了弥补这一差距,我们的研究引入了一项新任务,即评论上下文中的虚假评论意图检测(FRIDRC),其目的是通过利用文本和视觉信息来检测虚假评论意图,并构建一个包含手动和人工智能生成的虚假评论的数据集。此外,我们还开发了一个包含多模态表示和模态融合模块的预测框架。这些模块虽然是独立的,但它们是协同的,并有效地解决了识别虚假审查意图的挑战。我们的框架表现出出色的性能,在这项任务中,平均F1得分超过0.97,宏观F1得分超过0.96,优于先进的预训练模型。本研究不仅提出了准确识别和处理虚假评论意图的有效方法,而且强调了在虚假评论检测中利用多模态评论上下文信息的有效性。数据集和代码实现是公开的,可供进一步研究。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Electronic Commerce Research and Applications
Electronic Commerce Research and Applications 工程技术-计算机:跨学科应用
CiteScore
10.10
自引率
8.30%
发文量
97
审稿时长
63 days
期刊介绍: Electronic Commerce Research and Applications aims to create and disseminate enduring knowledge for the fast-changing e-commerce environment. A major dilemma in e-commerce research is how to achieve a balance between the currency and the life span of knowledge. Electronic Commerce Research and Applications will contribute to the establishment of a research community to create the knowledge, technology, theory, and applications for the development of electronic commerce. This is targeted at the intersection of technological potential and business aims.
期刊最新文献
Editorial Board Key decision criteria for integrating non-fungible tokens (NFTs) into e-commerce platforms ROQuA: A RAG-based opinion question-answering framework for e-commerce reviews What to recommend is when to recommend: Modeling multi-granularity repurchase cycles Coping with e-commerce filter bubbles: proactive versus reactive strategies
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1