Military Artificial Intelligence and the Principle of Distinction: A State Responsibility Perspective

IF 0.7 Q2 Social Sciences ISRAEL LAW REVIEW Pub Date : 2022-12-13 DOI:10.1017/S0021223722000188
Magdalena Pacholska
{"title":"Military Artificial Intelligence and the Principle of Distinction: A State Responsibility Perspective","authors":"Magdalena Pacholska","doi":"10.1017/S0021223722000188","DOIUrl":null,"url":null,"abstract":"Abstract Military artificial intelligence (AI)-enabled technology might still be in the relatively fledgling stages but the debate on how to regulate its use is already in full swing. Much of the discussion revolves around autonomous weapons systems (AWS) and the ‘responsibility gap’ they would ostensibly produce. This contribution argues that while some military AI technologies may indeed cause a range of conceptual hurdles in the realm of individual responsibility, they do not raise any unique issues under the law of state responsibility. The following analysis considers the latter regime and maps out crucial junctions in applying it to potential violations of the cornerstone of international humanitarian law (IHL) – the principle of distinction – resulting from the use of AI-enabled military technologies. It reveals that any challenges in ascribing responsibility in cases involving AWS would not be caused by the incorporation of AI, but stem from pre-existing systemic shortcomings of IHL and the unclear reverberations of mistakes thereunder. The article reiterates that state responsibility for the effects of AWS deployment is always retained through the commander's ultimate responsibility to authorise weapon deployment in accordance with IHL. It is proposed, however, that should the so-called fully autonomous weapon systems – that is, machine learning-based lethal systems that are capable of changing their own rules of operation beyond a predetermined framework – ever be fielded, it might be fairer to attribute their conduct to the fielding state, by conceptualising them as state agents, and treat them akin to state organs.","PeriodicalId":44911,"journal":{"name":"ISRAEL LAW REVIEW","volume":null,"pages":null},"PeriodicalIF":0.7000,"publicationDate":"2022-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISRAEL LAW REVIEW","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1017/S0021223722000188","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"Social Sciences","Score":null,"Total":0}
引用次数: 0

Abstract

Abstract Military artificial intelligence (AI)-enabled technology might still be in the relatively fledgling stages but the debate on how to regulate its use is already in full swing. Much of the discussion revolves around autonomous weapons systems (AWS) and the ‘responsibility gap’ they would ostensibly produce. This contribution argues that while some military AI technologies may indeed cause a range of conceptual hurdles in the realm of individual responsibility, they do not raise any unique issues under the law of state responsibility. The following analysis considers the latter regime and maps out crucial junctions in applying it to potential violations of the cornerstone of international humanitarian law (IHL) – the principle of distinction – resulting from the use of AI-enabled military technologies. It reveals that any challenges in ascribing responsibility in cases involving AWS would not be caused by the incorporation of AI, but stem from pre-existing systemic shortcomings of IHL and the unclear reverberations of mistakes thereunder. The article reiterates that state responsibility for the effects of AWS deployment is always retained through the commander's ultimate responsibility to authorise weapon deployment in accordance with IHL. It is proposed, however, that should the so-called fully autonomous weapon systems – that is, machine learning-based lethal systems that are capable of changing their own rules of operation beyond a predetermined framework – ever be fielded, it might be fairer to attribute their conduct to the fielding state, by conceptualising them as state agents, and treat them akin to state organs.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
军事人工智能与区别原则:国家责任视角
军事人工智能(AI)技术可能仍处于相对羽翼未丰的阶段,但关于如何规范其使用的辩论已经全面展开。大部分讨论都围绕着自主武器系统(AWS)及其表面上可能产生的“责任鸿沟”展开。这篇文章认为,虽然一些军用人工智能技术确实可能在个人责任领域造成一系列概念障碍,但在国家责任法下,它们并没有引起任何独特的问题。以下分析考虑了后一种制度,并指出了将其应用于因使用人工智能军事技术而可能违反国际人道法基石(区分原则)的关键节点。它表明,在涉及人工智能的案件中,责任归属方面的任何挑战都不是由人工智能的纳入引起的,而是源于国际人道法先前存在的系统性缺陷以及其中错误的不明确影响。该条重申,根据国际人道法,指挥官负有授权部署武器的最终责任,因此始终保留国家对部署空战部队的影响的责任。然而,有人建议,如果所谓的完全自主武器系统——即基于机器学习的致命系统,能够在预定框架之外改变自己的操作规则——被部署,那么通过将其概念化为国家代理人,并将其视为类似于国家机关,将其行为归因于部署国可能更公平。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
1.00
自引率
14.30%
发文量
19
期刊最新文献
Politics and Justice at the International Criminal Court The Effect of Russia's Invasion of Ukraine on Non-Human Animals: International Humanitarian Law Perspectives The Conduct of Hostilities, Attack Effects, and Criminal Accountability Charging Aggression as a Crime against Humanity? Revisiting the Proposal after Russia's Invasion of Ukraine How Misuse of Emergency Powers Dismantled the Rule of Law in Hungary
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1