首页 > 最新文献

Online Social Networks and Media最新文献

英文 中文
Generative exaggeration in LLM social agents: Consistency, bias, and toxicity 法学硕士社会代理人的生成夸张:一致性、偏见和毒性
IF 2.9 Q1 Social Sciences Pub Date : 2025-12-15 DOI: 10.1016/j.osnem.2025.100344
Jacopo Nudo , Mario Edoardo Pandolfo , Edoardo Loru , Mattia Samory , Matteo Cinelli , Walter Quattrociocchi
We investigate how Large Language Models (LLMs) behave when simulating political discourse on social media. Leveraging 21 million interactions on X during the 2024 U.S. presidential election, we construct LLM agents based on 1186 real users, prompting them to reply to politically salient tweets under controlled conditions. Agents are initialized either with minimal ideological cues (Zero Shot) or recent tweet history (Few Shot), allowing one-to-one comparisons with human replies. We evaluate three model families — Gemini, Mistral, and DeepSeek — across linguistic style, ideological consistency, and toxicity. We find that richer contextualization improves internal consistency but also amplifies polarization, stylized signals, and harmful language. We observe an emergent distortion that we call “generation exaggeration”: a systematic amplification of salient traits beyond empirical baselines. Our analysis shows that LLMs do not emulate users, they reconstruct them. Their outputs, indeed, reflect internal optimization dynamics more than observed behavior, introducing structural biases that compromise their reliability as social proxies. This challenges their use in content moderation, deliberative simulations, and policy modeling.
我们研究了大型语言模型(llm)在模拟社交媒体上的政治话语时的行为。利用2024年美国总统大选期间X上的2100万次互动,我们基于1186名真实用户构建了LLM代理,促使他们在受控条件下回复具有政治意义的推文。通过最小的意识形态线索(Zero Shot)或最近的tweet历史(Few Shot)初始化代理,允许与人类回复进行一对一的比较。我们从语言风格、意识形态一致性和毒性方面评估了三个模范家庭——双子座、西北风和DeepSeek。我们发现,丰富的语境化提高了内部一致性,但也放大了两极分化、程式化信号和有害语言。我们观察到一种新兴的扭曲现象,我们称之为“世代夸张”:一种超越经验基线的显著特征的系统性放大。我们的分析表明,法学硕士并不模仿用户,而是重构用户。事实上,它们的输出更多地反映了内部优化动态,而不是观察到的行为,这引入了结构性偏差,损害了它们作为社会代理的可靠性。这对它们在内容审核、审慎模拟和策略建模中的使用提出了挑战。
{"title":"Generative exaggeration in LLM social agents: Consistency, bias, and toxicity","authors":"Jacopo Nudo ,&nbsp;Mario Edoardo Pandolfo ,&nbsp;Edoardo Loru ,&nbsp;Mattia Samory ,&nbsp;Matteo Cinelli ,&nbsp;Walter Quattrociocchi","doi":"10.1016/j.osnem.2025.100344","DOIUrl":"10.1016/j.osnem.2025.100344","url":null,"abstract":"<div><div>We investigate how Large Language Models (LLMs) behave when simulating political discourse on social media. Leveraging 21 million interactions on X during the 2024 U.S. presidential election, we construct LLM agents based on 1186 real users, prompting them to reply to politically salient tweets under controlled conditions. Agents are initialized either with minimal ideological cues (Zero Shot) or recent tweet history (Few Shot), allowing one-to-one comparisons with human replies. We evaluate three model families — Gemini, Mistral, and DeepSeek — across linguistic style, ideological consistency, and toxicity. We find that richer contextualization improves internal consistency but also amplifies polarization, stylized signals, and harmful language. We observe an emergent distortion that we call “generation exaggeration”: a systematic amplification of salient traits beyond empirical baselines. Our analysis shows that LLMs do not emulate users, they reconstruct them. Their outputs, indeed, reflect internal optimization dynamics more than observed behavior, introducing structural biases that compromise their reliability as social proxies. This challenges their use in content moderation, deliberative simulations, and policy modeling.</div></div>","PeriodicalId":52228,"journal":{"name":"Online Social Networks and Media","volume":"51 ","pages":"Article 100344"},"PeriodicalIF":2.9,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145790242","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A longitudinal analysis of misinformation, polarization and toxicity on Bluesky after its public launch 蓝天上市后的误传、极化和毒性纵向分析
IF 2.9 Q1 Social Sciences Pub Date : 2025-11-27 DOI: 10.1016/j.osnem.2025.100342
Gianluca Nogara , Erfan Samieyan Sahneh , Matthew R. DeVerna , Nick Liu , Luca Luceri , Filippo Menczer , Francesco Pierri , Silvia Giordano
Bluesky is a decentralized, Twitter-like social media platform that has rapidly gained popularity. Following an invite-only phase, it officially opened to the public on February 6th, 2024, leading to a significant expansion of its user base. In this paper, we present a longitudinal analysis of user activity in the two months surrounding its public launch, examining how the platform evolved due to this rapid growth. Our analysis reveals that Bluesky exhibits an activity distribution comparable to more established social platforms, yet it features a higher volume of original content relative to reshared posts and maintains low toxicity levels. We further investigate the political leanings of its user base, misinformation dynamics, and engagement in harmful conversations. Our findings indicate that Bluesky users predominantly lean left politically and tend to share high-credibility sources. After the platform’s public launch, an influx of new users — particularly those posting in English and Japanese — contributed to a surge in activity. Among them, several accounts displayed suspicious behaviors, such as mass-following users and sharing content from low-credibility news sources. Some of these accounts have already been flagged as spam or suspended, suggesting that Bluesky’s moderation efforts have been effective.
Bluesky是一个去中心化的、类似twitter的社交媒体平台,已经迅速流行起来。在只接受邀请的阶段之后,它于2024年2月6日正式向公众开放,导致其用户群大幅扩大。在本文中,我们对其公开发行前后两个月的用户活动进行了纵向分析,考察了该平台是如何在这种快速增长中发展的。我们的分析显示,Bluesky的活动分布与更成熟的社交平台相当,但它的原创内容相对于转发帖子的数量更多,并且保持较低的毒性水平。我们将进一步调查其用户群的政治倾向、错误信息动态以及参与有害对话。我们的研究结果表明,蓝天用户主要在政治上偏左,并倾向于分享高可信度的消息来源。该平台公开发布后,新用户的涌入——尤其是那些用英语和日语发帖的用户——推动了活动的激增。其中,多个账号表现出了大量关注用户、分享低可信度新闻来源内容等可疑行为。其中一些账户已经被标记为垃圾邮件或被暂停,这表明Bluesky的审核工作是有效的。
{"title":"A longitudinal analysis of misinformation, polarization and toxicity on Bluesky after its public launch","authors":"Gianluca Nogara ,&nbsp;Erfan Samieyan Sahneh ,&nbsp;Matthew R. DeVerna ,&nbsp;Nick Liu ,&nbsp;Luca Luceri ,&nbsp;Filippo Menczer ,&nbsp;Francesco Pierri ,&nbsp;Silvia Giordano","doi":"10.1016/j.osnem.2025.100342","DOIUrl":"10.1016/j.osnem.2025.100342","url":null,"abstract":"<div><div>Bluesky is a decentralized, Twitter-like social media platform that has rapidly gained popularity. Following an invite-only phase, it officially opened to the public on February 6th, 2024, leading to a significant expansion of its user base. In this paper, we present a longitudinal analysis of user activity in the two months surrounding its public launch, examining how the platform evolved due to this rapid growth. Our analysis reveals that Bluesky exhibits an activity distribution comparable to more established social platforms, yet it features a higher volume of original content relative to reshared posts and maintains low toxicity levels. We further investigate the political leanings of its user base, misinformation dynamics, and engagement in harmful conversations. Our findings indicate that Bluesky users predominantly lean left politically and tend to share high-credibility sources. After the platform’s public launch, an influx of new users — particularly those posting in English and Japanese — contributed to a surge in activity. Among them, several accounts displayed suspicious behaviors, such as mass-following users and sharing content from low-credibility news sources. Some of these accounts have already been flagged as spam or suspended, suggesting that Bluesky’s moderation efforts have been effective.</div></div>","PeriodicalId":52228,"journal":{"name":"Online Social Networks and Media","volume":"51 ","pages":"Article 100342"},"PeriodicalIF":2.9,"publicationDate":"2025-11-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145618325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic user trust assessment in online social networks using liquid neural networks 基于液体神经网络的在线社交网络动态用户信任评估
IF 2.9 Q1 Social Sciences Pub Date : 2025-11-21 DOI: 10.1016/j.osnem.2025.100343
Youssef Gamha, Lotfi Ben Romdhane
Trust in online social networks (OSNs) is inherently dynamic, shaped by evolving user interactions, contextual shifts and behavioral changes. Traditional static trust models struggle to adapt to these fluid dynamics, limiting their applicability in real-time environments. This paper proposes DUTrust, a novel dynamic trust assessment framework that leverages Liquid Neural Networks (LNNs) to continuously update trust scores based on temporal, relational and context-sensitive factors. The DUTrust model integrates multiple facets of user behavior including user profile, interaction patterns, shared interests, network influence, and reciprocity into a unified model. This holistic data consolidation is a significant contribution, as it facilitates adaptive trust computation through LNNs’ real-time temporal reasoning. Experiments on real-world Twitter datasets demonstrate DUTrust’s effectiveness in predicting trustworthiness with high accuracy and adaptability to evolving user behavior.
在线社交网络(OSNs)中的信任本质上是动态的,受不断发展的用户交互、环境变化和行为变化的影响。传统的静态信任模型难以适应这些流体动力学,限制了它们在实时环境中的适用性。本文提出了一种新的动态信任评估框架DUTrust,该框架利用液态神经网络(LNNs)基于时间、关系和上下文敏感因素不断更新信任分数。DUTrust模型将用户行为的多个方面,包括用户概况、交互模式、共享兴趣、网络影响和互惠性,集成到一个统一的模型中。这种整体数据整合是重要的贡献,因为它通过LNNs的实时时间推理促进了自适应信任计算。在真实Twitter数据集上的实验证明了DUTrust在预测可信度方面的有效性,具有很高的准确性和对不断变化的用户行为的适应性。
{"title":"Dynamic user trust assessment in online social networks using liquid neural networks","authors":"Youssef Gamha,&nbsp;Lotfi Ben Romdhane","doi":"10.1016/j.osnem.2025.100343","DOIUrl":"10.1016/j.osnem.2025.100343","url":null,"abstract":"<div><div>Trust in online social networks (OSNs) is inherently dynamic, shaped by evolving user interactions, contextual shifts and behavioral changes. Traditional static trust models struggle to adapt to these fluid dynamics, limiting their applicability in real-time environments. This paper proposes DUTrust, a novel dynamic trust assessment framework that leverages Liquid Neural Networks (LNNs) to continuously update trust scores based on temporal, relational and context-sensitive factors. The DUTrust model integrates multiple facets of user behavior including user profile, interaction patterns, shared interests, network influence, and reciprocity into a unified model. This holistic data consolidation is a significant contribution, as it facilitates adaptive trust computation through LNNs’ real-time temporal reasoning. Experiments on real-world Twitter datasets demonstrate DUTrust’s effectiveness in predicting trustworthiness with high accuracy and adaptability to evolving user behavior.</div></div>","PeriodicalId":52228,"journal":{"name":"Online Social Networks and Media","volume":"51 ","pages":"Article 100343"},"PeriodicalIF":2.9,"publicationDate":"2025-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145555037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Identifying coordination in online social networks through anomalous sharing behaviour 通过异常分享行为识别在线社交网络中的协调
IF 2.9 Q1 Social Sciences Pub Date : 2025-11-14 DOI: 10.1016/j.osnem.2025.100341
Ahmad Zareie, Mehmet E. Bakir, Mark A. Greenwood, Kalina Bontcheva, Carolina Scarton
The proliferation of coordinated campaigns on Online Social Networks (OSNs) has raised increasing concerns over the last decade. These campaigns typically involve organised efforts by multiple accounts to manipulate public discourse or amplify particular narratives, and may include disinformation, astroturfing, or other influence operations. Therefore, identifying coordinated accounts and detecting the content they promote has become a critical challenge in OSN analysis. Existing methods for coordination detection focus mainly on the idea that accounts repeatedly sharing similar content are coordinated accounts. Since these methods ignore how this sharing behaviour differs from that of non-coordinated (regular) accounts, they may misidentify highly active accounts as coordinated accounts. To fill this gap, this paper proposes a novel method to detect coordination by looking for anomalies in accounts’ sharing behaviour. This method takes into account the extent to which the sharing behaviour of coordinated accounts diverges from that of regular accounts. Experimental results indicate that our approach is superior to the compared baselines for detecting coordination despite not requiring training or threshold optimisation.
在过去的十年中,在线社交网络(OSNs)上的协调运动的扩散引起了越来越多的关注。这些活动通常涉及多个账户有组织的努力,以操纵公共话语或放大特定叙述,并可能包括虚假信息、造谣或其他影响行动。因此,识别协调账号并检测其推广的内容成为OSN分析中的关键挑战。现有的协调检测方法主要集中在重复分享相似内容的账户是协调账户的想法上。由于这些方法忽略了这种共享行为与非协调(常规)帐户的不同之处,因此它们可能会将高度活跃的帐户误认为是协调帐户。为了填补这一空白,本文提出了一种通过寻找账户共享行为中的异常来检测协调的新方法。这种方法考虑到协调帐户的共享行为与正常帐户的共享行为的差异程度。实验结果表明,尽管不需要训练或阈值优化,我们的方法在检测协调方面优于比较基线。
{"title":"Identifying coordination in online social networks through anomalous sharing behaviour","authors":"Ahmad Zareie,&nbsp;Mehmet E. Bakir,&nbsp;Mark A. Greenwood,&nbsp;Kalina Bontcheva,&nbsp;Carolina Scarton","doi":"10.1016/j.osnem.2025.100341","DOIUrl":"10.1016/j.osnem.2025.100341","url":null,"abstract":"<div><div>The proliferation of coordinated campaigns on Online Social Networks (OSNs) has raised increasing concerns over the last decade. These campaigns typically involve organised efforts by multiple accounts to manipulate public discourse or amplify particular narratives, and may include disinformation, astroturfing, or other influence operations. Therefore, identifying coordinated accounts and detecting the content they promote has become a critical challenge in OSN analysis. Existing methods for coordination detection focus mainly on the idea that accounts repeatedly sharing similar content are coordinated accounts. Since these methods ignore how this sharing behaviour differs from that of non-coordinated (regular) accounts, they may misidentify highly active accounts as coordinated accounts. To fill this gap, this paper proposes a novel method to detect coordination by looking for anomalies in accounts’ sharing behaviour. This method takes into account the extent to which the sharing behaviour of coordinated accounts diverges from that of regular accounts. Experimental results indicate that our approach is superior to the compared baselines for detecting coordination despite not requiring training or threshold optimisation.</div></div>","PeriodicalId":52228,"journal":{"name":"Online Social Networks and Media","volume":"50 ","pages":"Article 100341"},"PeriodicalIF":2.9,"publicationDate":"2025-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145519975","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Misinformation mitigation in online social networks using continual learning with graph neural networks 利用图神经网络持续学习缓解在线社交网络中的错误信息
IF 2.9 Q1 Social Sciences Pub Date : 2025-11-04 DOI: 10.1016/j.osnem.2025.100340
Hichem Merini , Adil Imad Eddine Hosni , Kadda Beghdad Bey , Vincenzo Lomonaco , Marco Podda , Islem Baira
In today’s digital landscape, online social networks (OSNs) facilitate rapid information dissemination. However, they also serve as conduits for misinformation, leading to severe real-world consequences such as public panic, social unrest, and the erosion of institutional trust. Existing rumor influence minimization strategies predominantly rely on static models or specific diffusion mechanisms, restricting their ability to dynamically adapt to the evolving nature of misinformation. To address this gap, this paper proposes a novel misinformation influence mitigation framework that integrates Graph Neural Networks (GNNs) with continual learning and employs a Node Blocking strategy as its intervention approach. The framework comprises three key components: (1) a Dataset Generator, (2) a GNN Model Trainer, and (3) an Influential Node Identifier. Given the scarcity of real-world data on misinformation propagation, the first component simulates misinformation diffusion processes within social networks, leveraging the Human Individual and Social Behavior (HISB) model as a case study. The second component employs GNNs to learn from these synthetic datasets and predict the most influential nodes susceptible to misinformation. Subsequently, these nodes are strategically targeted and blocked to minimize further misinformation spread. Finally, the continual learning mechanism ensures the model dynamically adapts to evolving network structures and propagation patterns. Beyond evaluating the Human Individual and Social Behavior (HISB) propagation model, we empirically demonstrate that our framework is propagation-model agnostic by reproducing the pipeline under Independent Cascade and Linear Threshold with consistent gains over baselines. Finally, we introduce a truth-aware intervention rule that gates and weights actions by an external veracity score at detection time, selecting most influential nodes. This addition ensures interventions are enacted only when content is likely false, aligning the method with responsible deployment. Experimental evaluations conducted on multiple benchmark datasets demonstrate the superiority of the proposed node blocking framework over state-of-the-art methods. Our results indicate a statistically significant reduction in misinformation spread, with non-parametric statistical tests yielding p-values below 0.001 (p<0.001), confirming the robustness of our approach. This work presents a scalable and adaptable solution for misinformation containment, contributing to the development of more reliable and trustworthy online information ecosystems.
在当今的数字环境中,在线社交网络(OSNs)促进了信息的快速传播。然而,它们也成为错误信息的渠道,导致严重的现实后果,如公众恐慌、社会动荡和机构信任的侵蚀。现有的谣言影响最小化策略主要依赖于静态模型或特定的扩散机制,限制了它们动态适应错误信息演变性质的能力。为了解决这一差距,本文提出了一种新的错误信息影响缓解框架,该框架将图神经网络(gnn)与持续学习相结合,并采用节点阻塞策略作为其干预方法。该框架包括三个关键组件:(1)数据集生成器,(2)GNN模型训练器,以及(3)有影响力的节点标识符。考虑到关于错误信息传播的真实世界数据的稀缺性,第一部分模拟了社会网络中的错误信息传播过程,利用人类个体和社会行为(HISB)模型作为案例研究。第二个部分使用gnn从这些合成数据集中学习,并预测最容易受到错误信息影响的节点。随后,这些节点被战略性地锁定并封锁,以尽量减少进一步的错误信息传播。最后,持续学习机制保证了模型能够动态适应不断变化的网络结构和传播模式。除了评估人类个体和社会行为(HISB)传播模型之外,我们通过在独立级联和线性阈值下再现管道,并在基线上获得一致的收益,经验证明我们的框架是传播模型不可知的。最后,我们引入了一个真相感知干预规则,该规则通过检测时的外部准确性评分来对动作进行门和权,选择最具影响力的节点。这确保了只有在内容可能是错误的情况下才实施干预,使方法与负责任的部署保持一致。在多个基准数据集上进行的实验评估表明,所提出的节点阻塞框架优于最先进的方法。我们的结果表明,错误信息的传播在统计上显著减少,非参数统计检验的p值低于0.001 (p<0.001),证实了我们方法的稳健性。这项工作为遏制错误信息提供了一种可扩展和适应性强的解决方案,有助于开发更可靠和值得信赖的在线信息生态系统。
{"title":"Misinformation mitigation in online social networks using continual learning with graph neural networks","authors":"Hichem Merini ,&nbsp;Adil Imad Eddine Hosni ,&nbsp;Kadda Beghdad Bey ,&nbsp;Vincenzo Lomonaco ,&nbsp;Marco Podda ,&nbsp;Islem Baira","doi":"10.1016/j.osnem.2025.100340","DOIUrl":"10.1016/j.osnem.2025.100340","url":null,"abstract":"<div><div>In today’s digital landscape, online social networks (OSNs) facilitate rapid information dissemination. However, they also serve as conduits for misinformation, leading to severe real-world consequences such as public panic, social unrest, and the erosion of institutional trust. Existing rumor influence minimization strategies predominantly rely on static models or specific diffusion mechanisms, restricting their ability to dynamically adapt to the evolving nature of misinformation. To address this gap, this paper proposes a novel misinformation influence mitigation framework that integrates Graph Neural Networks (GNNs) with continual learning and employs a Node Blocking strategy as its intervention approach. The framework comprises three key components: (1) a Dataset Generator, (2) a GNN Model Trainer, and (3) an Influential Node Identifier. Given the scarcity of real-world data on misinformation propagation, the first component simulates misinformation diffusion processes within social networks, leveraging the Human Individual and Social Behavior (HISB) model as a case study. The second component employs GNNs to learn from these synthetic datasets and predict the most influential nodes susceptible to misinformation. Subsequently, these nodes are strategically targeted and blocked to minimize further misinformation spread. Finally, the continual learning mechanism ensures the model dynamically adapts to evolving network structures and propagation patterns. Beyond evaluating the Human Individual and Social Behavior (HISB) propagation model, we empirically demonstrate that our framework is propagation-model agnostic by reproducing the pipeline under Independent Cascade and Linear Threshold with consistent gains over baselines. Finally, we introduce a truth-aware intervention rule that gates and weights actions by an external veracity score at detection time, selecting most influential nodes. This addition ensures interventions are enacted only when content is likely false, aligning the method with responsible deployment. Experimental evaluations conducted on multiple benchmark datasets demonstrate the superiority of the proposed node blocking framework over state-of-the-art methods. Our results indicate a statistically significant reduction in misinformation spread, with non-parametric statistical tests yielding <span><math><mi>p</mi></math></span>-values below 0.001 (p<span><math><mo>&lt;</mo></math></span>0.001), confirming the robustness of our approach. This work presents a scalable and adaptable solution for misinformation containment, contributing to the development of more reliable and trustworthy online information ecosystems.</div></div>","PeriodicalId":52228,"journal":{"name":"Online Social Networks and Media","volume":"50 ","pages":"Article 100340"},"PeriodicalIF":2.9,"publicationDate":"2025-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145466715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
How does depression talk on social media? Modeling depression language with relevance-based statistical language models 抑郁症是如何在社交媒体上传播的?用基于相关性的统计语言模型建模抑郁语言
IF 2.9 Q1 Social Sciences Pub Date : 2025-10-22 DOI: 10.1016/j.osnem.2025.100339
Eliseo Bao , Anxo Perez , David Otero , Javier Parapar
Many individuals with mental health problems turn to the internet and social media for information and support. The text generated on these platforms serves as a valuable resource for identifying mental health risks, driving interdisciplinary research to develop models for mental health analysis and prediction. In this paper, we model depression-related language using relevance-based statistical language models to create lexicons that characterize linguistic patterns associated with depression. We also propose a ranking method that leverages these lexicons to prioritize users exhibiting stronger signs of depressive language on social media. Our models integrate clinical markers from established depression questionnaires, particularly the Beck Depression Inventory-II (BDI-II), enhancing explainability, generalization, and performance. Experiments across multiple social media datasets show that incorporating clinical knowledge improves user ranking and generalizes effectively across platforms. Additionally, we refine existing depression lexicons by applying weights estimated from our models, achieving better performance in generating depression-related queries. A comparative analysis of our models highlights differences in language use between control users and those with depression, aligning with prior psycholinguistic findings. This work advances the understanding of depression-related language through statistical modeling, paving the way for scalable social media interventions to identify at-risk individuals.
许多有心理健康问题的人转向互联网和社交媒体寻求信息和支持。在这些平台上生成的文本是识别心理健康风险的宝贵资源,推动跨学科研究开发心理健康分析和预测模型。在本文中,我们使用基于相关性的统计语言模型对抑郁症相关语言进行建模,以创建表征与抑郁症相关的语言模式的词汇。我们还提出了一种排序方法,利用这些词汇来优先考虑在社交媒体上表现出更强烈抑郁语言迹象的用户。我们的模型整合了来自已建立的抑郁症问卷的临床标记,特别是贝克抑郁量表ii (BDI-II),增强了可解释性、泛化性和性能。跨多个社交媒体数据集的实验表明,结合临床知识可以提高用户排名,并有效地跨平台推广。此外,我们通过应用从我们的模型中估计的权重来改进现有的抑郁症词汇,从而在生成抑郁症相关查询方面获得更好的性能。我们的模型对比分析强调了控制组使用者和抑郁症患者在语言使用上的差异,这与先前的心理语言学研究结果一致。这项工作通过统计建模促进了对抑郁症相关语言的理解,为可扩展的社交媒体干预措施铺平了道路,以识别有风险的个体。
{"title":"How does depression talk on social media? Modeling depression language with relevance-based statistical language models","authors":"Eliseo Bao ,&nbsp;Anxo Perez ,&nbsp;David Otero ,&nbsp;Javier Parapar","doi":"10.1016/j.osnem.2025.100339","DOIUrl":"10.1016/j.osnem.2025.100339","url":null,"abstract":"<div><div>Many individuals with mental health problems turn to the internet and social media for information and support. The text generated on these platforms serves as a valuable resource for identifying mental health risks, driving interdisciplinary research to develop models for mental health analysis and prediction. In this paper, we model depression-related language using relevance-based statistical language models to create lexicons that characterize linguistic patterns associated with depression. We also propose a ranking method that leverages these lexicons to prioritize users exhibiting stronger signs of depressive language on social media. Our models integrate clinical markers from established depression questionnaires, particularly the Beck Depression Inventory-II (BDI-II), enhancing explainability, generalization, and performance. Experiments across multiple social media datasets show that incorporating clinical knowledge improves user ranking and generalizes effectively across platforms. Additionally, we refine existing depression lexicons by applying weights estimated from our models, achieving better performance in generating depression-related queries. A comparative analysis of our models highlights differences in language use between control users and those with depression, aligning with prior psycholinguistic findings. This work advances the understanding of depression-related language through statistical modeling, paving the way for scalable social media interventions to identify at-risk individuals.</div></div>","PeriodicalId":52228,"journal":{"name":"Online Social Networks and Media","volume":"50 ","pages":"Article 100339"},"PeriodicalIF":2.9,"publicationDate":"2025-10-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145363065","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Online expressions, offline struggles: Using social media to identify depression-related symptoms 线上表达,线下挣扎:利用社交媒体识别抑郁相关症状
IF 2.9 Q1 Social Sciences Pub Date : 2025-10-14 DOI: 10.1016/j.osnem.2025.100338
Mario Ezra Aragón , Adrián Pastor López-Monroy , Manuel Montes-y-Gómez , David E. Losada
With their growing popularity, social media platforms have become valuable tools for researchers and health professionals, offering new opportunities to identify linguistic patterns associated with mental health. In this study, we analyze depression-related symptoms using user-generated posts on social media and the Beck Depression Inventory (BDI). Using posts from individuals who have self-reported a depression diagnosis, we train and evaluate sentence classification models to assess their ability to detect BDI symptoms. Specifically, we conduct binary classification experiments to identify the presence of depression-related symptoms and additional tests to categorize sentences into specific BDI symptom types. We also perform a comprehensive symptom-level analysis to examine how depressive symptoms are expressed linguistically, linking social media data with a clinically validated framework. In addition, we analyze symptom distributions between users with and without depression and across platforms, providing insight into how symptoms manifest in diverse online contexts. Furthermore, we incorporate a data augmentation strategy that leverages Large Language Models to generate clinically grounded synthetic examples and evaluate their effectiveness against human-generated data. Our findings indicate that users with depression exhibit a significantly higher prevalence of certain BDI symptoms – particularly Suicidal Thoughts, Crying, Self-Dislike, and Changes in Sleeping Pattern – while control users predominantly express milder categories such as Sadness or Pessimism. Synthetic data improves the detection of underrepresented symptoms and enhances model robustness, although human-generated data better captures subtle linguistic nuances. Specialized models outperform general ones, but specific symptom categories remain challenging, underscoring the need for more interpretable and clinically grounded detection frameworks.
随着社交媒体平台的日益普及,社交媒体平台已成为研究人员和卫生专业人员的宝贵工具,为识别与心理健康相关的语言模式提供了新的机会。在这项研究中,我们使用用户在社交媒体上发布的帖子和贝克抑郁量表(BDI)来分析抑郁相关症状。使用来自自我报告抑郁诊断的个人的帖子,我们训练和评估句子分类模型,以评估它们检测BDI症状的能力。具体来说,我们进行了二元分类实验来识别抑郁相关症状的存在,并进行了额外的测试来将句子分类为特定的BDI症状类型。我们还进行了全面的症状水平分析,以检查抑郁症状是如何在语言上表达的,将社交媒体数据与临床验证的框架联系起来。此外,我们还分析了患有和不患有抑郁症的用户之间以及跨平台的症状分布,从而深入了解了症状在不同在线环境中的表现。此外,我们结合了一个数据增强策略,利用大型语言模型来生成临床基础的合成示例,并评估它们对人类生成数据的有效性。我们的研究结果表明,患有抑郁症的用户在某些BDI症状(尤其是自杀念头、哭泣、自我厌恶和睡眠模式改变)方面表现出明显更高的患病率,而对照组用户主要表现出较温和的类别,如悲伤或悲观。尽管人工生成的数据可以更好地捕捉细微的语言差异,但合成数据可以改进对未充分代表的症状的检测并增强模型的鲁棒性。专业模型优于一般模型,但具体症状类别仍然具有挑战性,强调需要更多可解释和临床基础的检测框架。
{"title":"Online expressions, offline struggles: Using social media to identify depression-related symptoms","authors":"Mario Ezra Aragón ,&nbsp;Adrián Pastor López-Monroy ,&nbsp;Manuel Montes-y-Gómez ,&nbsp;David E. Losada","doi":"10.1016/j.osnem.2025.100338","DOIUrl":"10.1016/j.osnem.2025.100338","url":null,"abstract":"<div><div>With their growing popularity, social media platforms have become valuable tools for researchers and health professionals, offering new opportunities to identify linguistic patterns associated with mental health. In this study, we analyze depression-related symptoms using user-generated posts on social media and the Beck Depression Inventory (BDI). Using posts from individuals who have self-reported a depression diagnosis, we train and evaluate sentence classification models to assess their ability to detect BDI symptoms. Specifically, we conduct binary classification experiments to identify the presence of depression-related symptoms and additional tests to categorize sentences into specific BDI symptom types. We also perform a comprehensive symptom-level analysis to examine how depressive symptoms are expressed linguistically, linking social media data with a clinically validated framework. In addition, we analyze symptom distributions between users with and without depression and across platforms, providing insight into how symptoms manifest in diverse online contexts. Furthermore, we incorporate a data augmentation strategy that leverages Large Language Models to generate clinically grounded synthetic examples and evaluate their effectiveness against human-generated data. Our findings indicate that users with depression exhibit a significantly higher prevalence of certain BDI symptoms – particularly Suicidal Thoughts, Crying, Self-Dislike, and Changes in Sleeping Pattern – while control users predominantly express milder categories such as Sadness or Pessimism. Synthetic data improves the detection of underrepresented symptoms and enhances model robustness, although human-generated data better captures subtle linguistic nuances. Specialized models outperform general ones, but specific symptom categories remain challenging, underscoring the need for more interpretable and clinically grounded detection frameworks.</div></div>","PeriodicalId":52228,"journal":{"name":"Online Social Networks and Media","volume":"50 ","pages":"Article 100338"},"PeriodicalIF":2.9,"publicationDate":"2025-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145325050","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Web3 vs Fediverse: A comparative analysis of DeSo and Mastodon as decentralised social media ecosystems Web3 vs Fediverse: DeSo和乳齿象作为分散的社会媒体生态系统的比较分析
IF 2.9 Q1 Social Sciences Pub Date : 2025-10-11 DOI: 10.1016/j.osnem.2025.100337
Terence Zhang , Aniket Mahanti , Ranesh Naha
The rise of centralised social networks has consolidated power among a few major technology companies, raising critical concerns about privacy, censorship, and transparency. In response, decentralised alternatives, including Web3 platforms like Decentralised Social (DeSo) and Fediverse platforms such as Mastodon, have gained increasing attention. While prior research has explored individual aspects of decentralised networks, comparisons between Fediverse and Web3 platforms remain limited, and the unique dynamics of Web3 networks like DeSo are not well understood. This study provides the first in-depth study of DeSo, characterising user behaviour, discourse, and economic activities, and compares these with Mastodon and memo.cash. We collected over 3.1M posts from 13K users on DeSo and Mastodon, along with 11M DeSo on-chain transactions via public APIs. Our analysis reveals that while DeSo and Mastodon share similarities in passive content engagement, they differ in their use of URLs, hashtags, and community focus. DeSo is primarily oriented around Decentralised Finance (DeFi) topics, whereas Mastodon hosts diverse discussions with an emphasis on news and politics. Despite DeSo’s decentralised social graph, its transaction graph remains centralised, underscoring the need for further decentralisation in Web3 platforms. Additionally, while wealth inequality exists on DeSo, low transaction fees promote user participation irrespective of financial status. These findings provide new insights into the evolving landscape of decentralised social networks and highlight critical areas for future research and platform development.
集中式社交网络的兴起巩固了少数几家大型科技公司的权力,引发了对隐私、审查和透明度的严重担忧。作为回应,包括Web3平台(如DeSo)和Fediverse平台(如Mastodon)在内的去中心化替代方案获得了越来越多的关注。虽然之前的研究已经探索了去中心化网络的各个方面,但Fediverse和Web3平台之间的比较仍然有限,而且像DeSo这样的Web3网络的独特动态也没有得到很好的理解。本研究首次对DeSo进行了深入研究,描述了用户行为、话语和经济活动的特征,并将其与Mastodon和memo.cash进行了比较。我们在DeSo和Mastodon上收集了1.3万名用户的310万篇帖子,以及通过公共api进行的1100万笔DeSo链上交易。我们的分析显示,尽管DeSo和Mastodon在被动内容参与方面有相似之处,但它们在使用url、标签和社区焦点方面有所不同。DeSo主要围绕去中心化金融(DeFi)主题,而Mastodon则举办以新闻和政治为重点的各种讨论。尽管DeSo的社交图谱是去中心化的,但其交易图谱仍然是中心化的,这凸显了Web3平台进一步去中心化的必要性。此外,尽管DeSo上存在财富不平等,但无论财务状况如何,低廉的交易费用都促进了用户的参与。这些发现为去中心化社交网络的发展前景提供了新的见解,并强调了未来研究和平台开发的关键领域。
{"title":"Web3 vs Fediverse: A comparative analysis of DeSo and Mastodon as decentralised social media ecosystems","authors":"Terence Zhang ,&nbsp;Aniket Mahanti ,&nbsp;Ranesh Naha","doi":"10.1016/j.osnem.2025.100337","DOIUrl":"10.1016/j.osnem.2025.100337","url":null,"abstract":"<div><div>The rise of centralised social networks has consolidated power among a few major technology companies, raising critical concerns about privacy, censorship, and transparency. In response, decentralised alternatives, including Web3 platforms like Decentralised Social (DeSo) and Fediverse platforms such as Mastodon, have gained increasing attention. While prior research has explored individual aspects of decentralised networks, comparisons between Fediverse and Web3 platforms remain limited, and the unique dynamics of Web3 networks like DeSo are not well understood. This study provides the first in-depth study of DeSo, characterising user behaviour, discourse, and economic activities, and compares these with Mastodon and <span>memo.cash</span>. We collected over 3.1M posts from 13K users on DeSo and Mastodon, along with 11M DeSo on-chain transactions via public APIs. Our analysis reveals that while DeSo and Mastodon share similarities in passive content engagement, they differ in their use of URLs, hashtags, and community focus. DeSo is primarily oriented around Decentralised Finance (DeFi) topics, whereas Mastodon hosts diverse discussions with an emphasis on news and politics. Despite DeSo’s decentralised social graph, its transaction graph remains centralised, underscoring the need for further decentralisation in Web3 platforms. Additionally, while wealth inequality exists on DeSo, low transaction fees promote user participation irrespective of financial status. These findings provide new insights into the evolving landscape of decentralised social networks and highlight critical areas for future research and platform development.</div></div>","PeriodicalId":52228,"journal":{"name":"Online Social Networks and Media","volume":"50 ","pages":"Article 100337"},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145268666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predicting, evaluating, and explaining top misinformation spreaders via archetypal user behavior 通过原型用户行为预测、评估和解释最主要的错误信息传播者
IF 2.9 Q1 Social Sciences Pub Date : 2025-10-03 DOI: 10.1016/j.osnem.2025.100336
Enrico Verdolotti , Luca Luceri , Silvia Giordano
The spread of misinformation on social networks poses a significant challenge to online communities and society at large. Not all users contribute equally to this phenomenon: a small number of highly effective individuals can exert outsized influence, amplifying false narratives and contributing to significant societal harm. This paper seeks to mitigate the spread of misinformation by enabling proactive interventions, identifying and ranking users according to key behavioral indicators associated with harmful content dissemination. We examine three user archetypes — amplifiers, super-spreaders, and coordinated accounts — each characterized by distinct behavioral patterns in the dissemination of misinformation. These are not mutually exclusive, and individual users may exhibit characteristics of multiple archetypes. We develop and evaluate several user ranking models, each aligned with a specific archetype, and find that super-spreader traits consistently dominate the top ranks among the most influential misinformation spreaders. As we move down the ranking, however, the interplay of multiple archetypes becomes more prominent. Additionally, we demonstrate the critical role of temporal dynamics in predictive performance, and introduce methods that reduce data requirements by minimizing the observation window needed for accurate forecasting. Finally, we demonstrate the utility and benefits of explainable AI (XAI) techniques, integrating multiple archetypal traits into a unified model to enhance interpretability and offer deeper insight into the key factors driving misinformation propagation. Our findings provide actionable tools for identifying potentially harmful users and guiding content moderation strategies, enabling platforms to monitor accounts of concern more effectively.
错误信息在社交网络上的传播对在线社区和整个社会构成了重大挑战。并不是所有的用户都对这种现象做出了同样的贡献:少数高效的个人可以施加巨大的影响力,放大虚假的叙述,造成重大的社会危害。本文旨在通过主动干预、根据与有害内容传播相关的关键行为指标识别和对用户进行排名来减轻错误信息的传播。我们研究了三种用户原型——放大者、超级传播者和协调账户——每一种都以传播错误信息的不同行为模式为特征。这些不是相互排斥的,单个用户可能表现出多个原型的特征。我们开发并评估了几个用户排名模型,每个模型都与一个特定的原型相一致,并发现超级传播者的特征始终在最具影响力的错误信息传播者中占据主导地位。然而,随着排名的下降,多种原型的相互作用变得更加突出。此外,我们展示了时间动态在预测性能中的关键作用,并介绍了通过最小化准确预测所需的观测窗口来减少数据需求的方法。最后,我们展示了可解释人工智能(XAI)技术的效用和好处,将多个原型特征集成到一个统一的模型中,以增强可解释性,并更深入地了解驱动错误信息传播的关键因素。我们的研究结果为识别潜在有害用户和指导内容审核策略提供了可操作的工具,使平台能够更有效地监控相关账户。
{"title":"Predicting, evaluating, and explaining top misinformation spreaders via archetypal user behavior","authors":"Enrico Verdolotti ,&nbsp;Luca Luceri ,&nbsp;Silvia Giordano","doi":"10.1016/j.osnem.2025.100336","DOIUrl":"10.1016/j.osnem.2025.100336","url":null,"abstract":"<div><div>The spread of misinformation on social networks poses a significant challenge to online communities and society at large. Not all users contribute equally to this phenomenon: a small number of highly effective individuals can exert outsized influence, amplifying false narratives and contributing to significant societal harm. This paper seeks to mitigate the spread of misinformation by enabling proactive interventions, identifying and ranking users according to key behavioral indicators associated with harmful content dissemination. We examine three user archetypes — <em>amplifiers</em>, <em>super-spreaders</em>, and <em>coordinated accounts</em> — each characterized by distinct behavioral patterns in the dissemination of misinformation. These are not mutually exclusive, and individual users may exhibit characteristics of multiple archetypes. We develop and evaluate several user ranking models, each aligned with a specific archetype, and find that <em>super-spreader</em> traits consistently dominate the top ranks among the most influential misinformation spreaders. As we move down the ranking, however, the interplay of multiple archetypes becomes more prominent. Additionally, we demonstrate the critical role of temporal dynamics in predictive performance, and introduce methods that reduce data requirements by minimizing the observation window needed for accurate forecasting. Finally, we demonstrate the utility and benefits of explainable AI (XAI) techniques, integrating multiple archetypal traits into a unified model to enhance interpretability and offer deeper insight into the key factors driving misinformation propagation. Our findings provide actionable tools for identifying potentially harmful users and guiding content moderation strategies, enabling platforms to monitor accounts of concern more effectively.</div></div>","PeriodicalId":52228,"journal":{"name":"Online Social Networks and Media","volume":"50 ","pages":"Article 100336"},"PeriodicalIF":2.9,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145222482","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The perils of stealthy data poisoning attacks in misogynistic content moderation 在歧视女性的内容审核中,秘密数据中毒攻击的危险
IF 2.9 Q1 Social Sciences Pub Date : 2025-09-23 DOI: 10.1016/j.osnem.2025.100334
Syrine Enneifer, Federica Baccini, Federico Siciliano, Irene Amerini, Fabrizio Silvestri
Moderating harmful content, such as misogynistic language, is essential to ensure safety and well-being in online spaces. To this end, text classification models have been used to detect toxic content, especially in communities that are known to promote violence and radicalization in the real world, such as the incel movement. However, these models remain vulnerable to targeted data poisoning attacks. In this work, we present a realistic targeted poisoning strategy in which an adversary aims at misclassifying specific misogynistic comments in order to evade detection. While prior approaches craft poisoned samples with explicit trigger phrases, our method relies exclusively on existing training data. In particular, we repurpose the concept of opponents, training points that negatively influence the prediction of a target test point, to identify poisoned points to be added to the training set, either in their original form or in a paraphrased variant. The effectiveness of the attack is then measured on several aspects: success rate, number of poisoned samples required, and preservation of the overall model performance. Our results on two different datasets show that only a small fraction of malicious inputs are possibly sufficient to undermine classification of a target sample, while leaving the model performance on non-target points virtually unaffected, revealing the stealthy nature of the attack. Finally, we show that the attack can be transferred across different models, thus highlighting its practical relevance in real-world scenarios. Overall, our work raises awareness on the vulnerability of powerful machine learning models to data poisoning attacks, and will possibly encourage the development of efficient defense and mitigation techniques to strengthen the security of automated moderation systems.
节制有害内容,如歧视女性的语言,对于确保网络空间的安全和福祉至关重要。为此,文本分类模型已被用于检测有毒内容,特别是在已知在现实世界中促进暴力和激进化的社区中,例如incel运动。然而,这些模型仍然容易受到有针对性的数据中毒攻击。在这项工作中,我们提出了一种现实的针对性中毒策略,其中对手旨在错误分类特定的厌女评论以逃避检测。虽然之前的方法使用明确的触发短语来制作有毒样本,但我们的方法完全依赖于现有的训练数据。特别是,我们重新定义了对手的概念,即对目标测试点的预测产生负面影响的训练点,以识别要添加到训练集中的有毒点,无论是以原始形式还是以改写的形式。然后从几个方面来衡量攻击的有效性:成功率、所需的中毒样本数量和整体模型性能的保存。我们在两个不同数据集上的结果表明,只有一小部分恶意输入可能足以破坏目标样本的分类,同时使模型在非目标点上的性能几乎不受影响,揭示了攻击的隐蔽性。最后,我们展示了攻击可以在不同的模型之间转移,从而突出了其在现实场景中的实际相关性。总的来说,我们的工作提高了人们对强大的机器学习模型对数据中毒攻击的脆弱性的认识,并可能鼓励开发有效的防御和缓解技术,以加强自动审核系统的安全性。
{"title":"The perils of stealthy data poisoning attacks in misogynistic content moderation","authors":"Syrine Enneifer,&nbsp;Federica Baccini,&nbsp;Federico Siciliano,&nbsp;Irene Amerini,&nbsp;Fabrizio Silvestri","doi":"10.1016/j.osnem.2025.100334","DOIUrl":"10.1016/j.osnem.2025.100334","url":null,"abstract":"<div><div>Moderating harmful content, such as misogynistic language, is essential to ensure safety and well-being in online spaces. To this end, text classification models have been used to detect toxic content, especially in communities that are known to promote violence and radicalization in the real world, such as the <em>incel</em> movement. However, these models remain vulnerable to targeted data poisoning attacks. In this work, we present a realistic targeted poisoning strategy in which an adversary aims at misclassifying specific misogynistic comments in order to evade detection. While prior approaches craft poisoned samples with explicit trigger phrases, our method relies exclusively on existing training data. In particular, we repurpose the concept of <em>opponents</em>, training points that negatively influence the prediction of a target test point, to identify poisoned points to be added to the training set, either in their original form or in a paraphrased variant. The effectiveness of the attack is then measured on several aspects: success rate, number of poisoned samples required, and preservation of the overall model performance. Our results on two different datasets show that only a small fraction of malicious inputs are possibly sufficient to undermine classification of a target sample, while leaving the model performance on non-target points virtually unaffected, revealing the stealthy nature of the attack. Finally, we show that the attack can be transferred across different models, thus highlighting its practical relevance in real-world scenarios. Overall, our work raises awareness on the vulnerability of powerful machine learning models to data poisoning attacks, and will possibly encourage the development of efficient defense and mitigation techniques to strengthen the security of automated moderation systems.</div></div>","PeriodicalId":52228,"journal":{"name":"Online Social Networks and Media","volume":"50 ","pages":"Article 100334"},"PeriodicalIF":2.9,"publicationDate":"2025-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145109455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Online Social Networks and Media
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1