{"title":"Zero-Shot Machine-Generated Text Detection Using Mixture of Large Language Models","authors":"Matthieu Dubois, François Yvon, Pablo Piantanida","doi":"arxiv-2409.07615","DOIUrl":null,"url":null,"abstract":"The dissemination of Large Language Models (LLMs), trained at scale, and\nendowed with powerful text-generating abilities has vastly increased the\nthreats posed by generative AI technologies by reducing the cost of producing\nharmful, toxic, faked or forged content. In response, various proposals have\nbeen made to automatically discriminate artificially generated from\nhuman-written texts, typically framing the problem as a classification problem.\nMost approaches evaluate an input document by a well-chosen detector LLM,\nassuming that low-perplexity scores reliably signal machine-made content. As\nusing one single detector can induce brittleness of performance, we instead\nconsider several and derive a new, theoretically grounded approach to combine\ntheir respective strengths. Our experiments, using a variety of generator LLMs,\nsuggest that our method effectively increases the robustness of detection.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"70 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computation and Language","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07615","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The dissemination of Large Language Models (LLMs), trained at scale, and
endowed with powerful text-generating abilities has vastly increased the
threats posed by generative AI technologies by reducing the cost of producing
harmful, toxic, faked or forged content. In response, various proposals have
been made to automatically discriminate artificially generated from
human-written texts, typically framing the problem as a classification problem.
Most approaches evaluate an input document by a well-chosen detector LLM,
assuming that low-perplexity scores reliably signal machine-made content. As
using one single detector can induce brittleness of performance, we instead
consider several and derive a new, theoretically grounded approach to combine
their respective strengths. Our experiments, using a variety of generator LLMs,
suggest that our method effectively increases the robustness of detection.