The silence of the LLMs: Cross-lingual analysis of guardrail-related political bias and false information prevalence in ChatGPT, Google Bard (Gemini), and Bing Chat

IF 7.6 2区 管理学 Q1 INFORMATION SCIENCE & LIBRARY SCIENCE Telematics and Informatics Pub Date : 2024-11-20 DOI:10.1016/j.tele.2024.102211
Aleksandra Urman , Mykola Makhortykh
{"title":"The silence of the LLMs: Cross-lingual analysis of guardrail-related political bias and false information prevalence in ChatGPT, Google Bard (Gemini), and Bing Chat","authors":"Aleksandra Urman ,&nbsp;Mykola Makhortykh","doi":"10.1016/j.tele.2024.102211","DOIUrl":null,"url":null,"abstract":"<div><div>This article presents a comparative analysis of political bias in the outputs of three Large Language Model (LLM)-based chatbots – ChatGPT (GPT3.5, GPT4, GPT4o), Bing Chat, and Bard/Gemini – in response to political queries concerning the authoritarian regime in Russia. We investigate whether safeguards implemented in these chatbots contribute to the censorship of information that is viewed as harmful by the regime, in particular information about Vladimir Putin and the Russian war against Ukraine, and whether these safeguards enable the generation of false claims, in particular in relation to the regime’s internal and external opponents. To detect whether LLM safeguards reiterate political bias, the article compares the outputs of prompts focusing on Putin’s regime and the ones dealing with the Russian opposition and the US and Ukrainian politicians. It also examines whether the degree of bias varies depending on the language of the prompt and compares outputs concerning political personalities and issues across three languages: Russian, Ukrainian, and English. The results reveal significant disparities in how individual chatbots withhold politics-related information or produce false claims in relation to it. Notably, Bard consistently refused to respond to queries about Vladimir Putin in Russian, even when the relevant information was accessible via Google Search, and generally followed the censorship guidelines that, according to Yandex-related data leaks, were issued by the Russian authorities. A subsequent evaluation of Gemini showed that the chatbot restricts political information beyond what was officially confirmed by Google. In terms of false claims, we find substantial variation across languages with Ukrainian and Russian prompts generating false information more often and Bard being more prone to produce false claims in relation to Russian regime opponents (e.g., Navalny or Zelenskyy) than other chatbots. We also found that while GPT4 and GPT4o generate less factually incorrect information, both models still make mistakes, with their prevalence being higher in Russian and Ukrainian than in English. This research aims to stimulate further dialogue and research on developing safeguards against the misuse of LLMs outside of democratic environments.</div></div>","PeriodicalId":48257,"journal":{"name":"Telematics and Informatics","volume":"96 ","pages":"Article 102211"},"PeriodicalIF":7.6000,"publicationDate":"2024-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Telematics and Informatics","FirstCategoryId":"91","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0736585324001151","RegionNum":2,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"INFORMATION SCIENCE & LIBRARY SCIENCE","Score":null,"Total":0}
引用次数: 0

Abstract

This article presents a comparative analysis of political bias in the outputs of three Large Language Model (LLM)-based chatbots – ChatGPT (GPT3.5, GPT4, GPT4o), Bing Chat, and Bard/Gemini – in response to political queries concerning the authoritarian regime in Russia. We investigate whether safeguards implemented in these chatbots contribute to the censorship of information that is viewed as harmful by the regime, in particular information about Vladimir Putin and the Russian war against Ukraine, and whether these safeguards enable the generation of false claims, in particular in relation to the regime’s internal and external opponents. To detect whether LLM safeguards reiterate political bias, the article compares the outputs of prompts focusing on Putin’s regime and the ones dealing with the Russian opposition and the US and Ukrainian politicians. It also examines whether the degree of bias varies depending on the language of the prompt and compares outputs concerning political personalities and issues across three languages: Russian, Ukrainian, and English. The results reveal significant disparities in how individual chatbots withhold politics-related information or produce false claims in relation to it. Notably, Bard consistently refused to respond to queries about Vladimir Putin in Russian, even when the relevant information was accessible via Google Search, and generally followed the censorship guidelines that, according to Yandex-related data leaks, were issued by the Russian authorities. A subsequent evaluation of Gemini showed that the chatbot restricts political information beyond what was officially confirmed by Google. In terms of false claims, we find substantial variation across languages with Ukrainian and Russian prompts generating false information more often and Bard being more prone to produce false claims in relation to Russian regime opponents (e.g., Navalny or Zelenskyy) than other chatbots. We also found that while GPT4 and GPT4o generate less factually incorrect information, both models still make mistakes, with their prevalence being higher in Russian and Ukrainian than in English. This research aims to stimulate further dialogue and research on developing safeguards against the misuse of LLMs outside of democratic environments.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
法学硕士的沉默:在ChatGPT, b谷歌Bard (Gemini)和Bing Chat中护栏相关的政治偏见和虚假信息流行的跨语言分析
本文对三种基于大型语言模型(LLM)的聊天机器人——ChatGPT (GPT3.5、GPT4、gpt40)、必应聊天和Bard/Gemini——在回答有关俄罗斯独裁政权的政治问题时的输出中的政治偏见进行了比较分析。我们调查在这些聊天机器人中实施的保障措施是否有助于审查被政权视为有害的信息,特别是关于弗拉基米尔·普京和俄罗斯对乌克兰的战争的信息,以及这些保障措施是否能够产生虚假声明,特别是与政权的内部和外部对手有关。为了检测法学硕士保障是否重申了政治偏见,本文比较了关注普京政权的提示和处理俄罗斯反对派、美国和乌克兰政客的提示的输出。它还检查了偏见的程度是否因提示语的不同而不同,并比较了三种语言(俄语、乌克兰语和英语)关于政治人物和问题的输出。结果显示,在个人聊天机器人如何隐瞒与政治相关的信息或产生与之相关的虚假声明方面,存在巨大差异。值得注意的是,巴德一直拒绝用俄语回答有关弗拉基米尔·普京(Vladimir Putin)的问题,即使相关信息可以通过谷歌搜索(谷歌Search)获得,而且总体上遵守审查准则,根据与yandex相关的数据泄露,这些准则是由俄罗斯当局发布的。随后对Gemini的评估显示,该聊天机器人限制的政治信息超出了b谷歌官方确认的范围。在虚假声明方面,我们发现不同语言之间存在很大差异,乌克兰语和俄语提示更频繁地产生虚假信息,而巴德比其他聊天机器人更容易产生与俄罗斯政权对手(例如纳瓦尔尼或泽伦斯基)有关的虚假声明。我们还发现,虽然GPT4和gpt40生成的事实错误信息较少,但这两种模型仍然存在错误,俄语和乌克兰语的错误发生率高于英语。本研究旨在促进进一步的对话和研究,以制定防止在民主环境之外滥用法学硕士的保障措施。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Telematics and Informatics
Telematics and Informatics INFORMATION SCIENCE & LIBRARY SCIENCE-
CiteScore
17.00
自引率
4.70%
发文量
104
审稿时长
24 days
期刊介绍: Telematics and Informatics is an interdisciplinary journal that publishes cutting-edge theoretical and methodological research exploring the social, economic, geographic, political, and cultural impacts of digital technologies. It covers various application areas, such as smart cities, sensors, information fusion, digital society, IoT, cyber-physical technologies, privacy, knowledge management, distributed work, emergency response, mobile communications, health informatics, social media's psychosocial effects, ICT for sustainable development, blockchain, e-commerce, and e-government.
期刊最新文献
Towards harmonization of EU and UN methodologies for the assessment of eGovernment The silence of the LLMs: Cross-lingual analysis of guardrail-related political bias and false information prevalence in ChatGPT, Google Bard (Gemini), and Bing Chat How to fight loneliness: The importance of networked friendship quality and self-esteem differs with age Overcoming social and ethical challenges of AI through organizational resilience. A PLS-SEM approach Understanding online shopping fraud among Chinese elderly: Extending routine activity theory in the online context
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1