{"title":"Safeguarding Decentralized Social Media: LLM Agents for Automating Community Rule Compliance","authors":"Lucio La Cava, Andrea Tagarelli","doi":"arxiv-2409.08963","DOIUrl":null,"url":null,"abstract":"Ensuring content compliance with community guidelines is crucial for\nmaintaining healthy online social environments. However, traditional\nhuman-based compliance checking struggles with scaling due to the increasing\nvolume of user-generated content and a limited number of moderators. Recent\nadvancements in Natural Language Understanding demonstrated by Large Language\nModels unlock new opportunities for automated content compliance verification.\nThis work evaluates six AI-agents built on Open-LLMs for automated rule\ncompliance checking in Decentralized Social Networks, a challenging environment\ndue to heterogeneous community scopes and rules. Analyzing over 50,000 posts\nfrom hundreds of Mastodon servers, we find that AI-agents effectively detect\nnon-compliant content, grasp linguistic subtleties, and adapt to diverse\ncommunity contexts. Most agents also show high inter-rater reliability and\nconsistency in score justification and suggestions for compliance. Human-based\nevaluation with domain experts confirmed the agents' reliability and\nusefulness, rendering them promising tools for semi-automated or\nhuman-in-the-loop content moderation systems.","PeriodicalId":501043,"journal":{"name":"arXiv - PHYS - Physics and Society","volume":"3 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - PHYS - Physics and Society","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.08963","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Ensuring content compliance with community guidelines is crucial for
maintaining healthy online social environments. However, traditional
human-based compliance checking struggles with scaling due to the increasing
volume of user-generated content and a limited number of moderators. Recent
advancements in Natural Language Understanding demonstrated by Large Language
Models unlock new opportunities for automated content compliance verification.
This work evaluates six AI-agents built on Open-LLMs for automated rule
compliance checking in Decentralized Social Networks, a challenging environment
due to heterogeneous community scopes and rules. Analyzing over 50,000 posts
from hundreds of Mastodon servers, we find that AI-agents effectively detect
non-compliant content, grasp linguistic subtleties, and adapt to diverse
community contexts. Most agents also show high inter-rater reliability and
consistency in score justification and suggestions for compliance. Human-based
evaluation with domain experts confirmed the agents' reliability and
usefulness, rendering them promising tools for semi-automated or
human-in-the-loop content moderation systems.