{"title":"Bending the Automation Bias Curve: A Study of Human and AI-Based Decision Making in National Security Contexts","authors":"Michael C Horowitz, Lauren Kahn","doi":"10.1093/isq/sqae020","DOIUrl":null,"url":null,"abstract":"Uses of artificial intelligence (AI) are growing around the world. What will influence AI adoption in the international security realm? Research on automation bias suggests that humans can often be overconfident in AI, whereas research on algorithm aversion shows that, as the stakes of a decision rise, humans become more cautious about trusting algorithms. We theorize about the relationship between background knowledge about AI, trust in AI, and how these interact with other factors to influence the probability of automation bias in the international security context. We test these in a preregistered task identification experiment across a representative sample of 9,000 adults in nine countries with varying levels of AI industries. The results strongly support the theory, especially concerning AI background knowledge. A version of the Dunning–Kruger effect appears to be at play, whereby those with the lowest level of experience with AI are slightly more likely to be algorithm-averse, then automation bias occurs at lower levels of knowledge before leveling off as a respondent’s AI background reaches the highest levels. Additional results show effects from the task’s difficulty, overall AI trust, and whether a human or AI decision aid is described as highly competent or less competent.","PeriodicalId":48313,"journal":{"name":"International Studies Quarterly","volume":"46 1","pages":""},"PeriodicalIF":2.4000,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Studies Quarterly","FirstCategoryId":"90","ListUrlMain":"https://doi.org/10.1093/isq/sqae020","RegionNum":1,"RegionCategory":"社会学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"INTERNATIONAL RELATIONS","Score":null,"Total":0}
引用次数: 0
Abstract
Uses of artificial intelligence (AI) are growing around the world. What will influence AI adoption in the international security realm? Research on automation bias suggests that humans can often be overconfident in AI, whereas research on algorithm aversion shows that, as the stakes of a decision rise, humans become more cautious about trusting algorithms. We theorize about the relationship between background knowledge about AI, trust in AI, and how these interact with other factors to influence the probability of automation bias in the international security context. We test these in a preregistered task identification experiment across a representative sample of 9,000 adults in nine countries with varying levels of AI industries. The results strongly support the theory, especially concerning AI background knowledge. A version of the Dunning–Kruger effect appears to be at play, whereby those with the lowest level of experience with AI are slightly more likely to be algorithm-averse, then automation bias occurs at lower levels of knowledge before leveling off as a respondent’s AI background reaches the highest levels. Additional results show effects from the task’s difficulty, overall AI trust, and whether a human or AI decision aid is described as highly competent or less competent.
期刊介绍:
International Studies Quarterly, the official journal of the International Studies Association, seeks to acquaint a broad audience of readers with the best work being done in the variety of intellectual traditions included under the rubric of international studies. Therefore, the editors welcome all submissions addressing this community"s theoretical, empirical, and normative concerns. First preference will continue to be given to articles that address and contribute to important disciplinary and interdisciplinary questions and controversies.