Carsten Orwat, Jascha Bareis, Anja Folberth, Jutta Jahnel, Christian Wadephul
{"title":"Normative Challenges of Risk Regulation of Artificial Intelligence","authors":"Carsten Orwat, Jascha Bareis, Anja Folberth, Jutta Jahnel, Christian Wadephul","doi":"10.1007/s11569-024-00454-9","DOIUrl":null,"url":null,"abstract":"<p>Approaches aimed at regulating artificial intelligence (AI) include a particular form of risk regulation, i.e. a risk-based approach. The most prominent example is the European Union’s Artificial Intelligence Act (AI Act). This article addresses the challenges for adequate risk regulation that arise primarily from the specific type of risks involved, i.e. risks to the protection of fundamental rights and fundamental societal values. This is mainly due to the normative ambiguity of such rights and societal values when attempts are made to select, interpret, specify or operationalise them for the purposes of risk assessments and risk mitigation. This is exemplified by (1) human dignity, (2) informational self-determination, data protection and privacy, (3) anti-discrimination, fairness and justice, and (4) the common good. Normative ambiguities require normative choices, which are assigned to different actors under the regime of the AI Act. Particularly critical normative choices include selecting normative concepts by which to operationalise and specify risks, aggregating and quantifying risks (including the use of metrics), balancing value conflicts, setting levels of acceptable risks, and standardisation. To ensure that these normative choices do not lack democratic legitimacy and to avoid legal uncertainty, further political processes and scientific debates are suggested.</p>","PeriodicalId":18802,"journal":{"name":"Nanoethics","volume":"18 1","pages":""},"PeriodicalIF":1.1000,"publicationDate":"2024-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Nanoethics","FirstCategoryId":"98","ListUrlMain":"https://doi.org/10.1007/s11569-024-00454-9","RegionNum":4,"RegionCategory":"哲学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ETHICS","Score":null,"Total":0}
引用次数: 0
Abstract
Approaches aimed at regulating artificial intelligence (AI) include a particular form of risk regulation, i.e. a risk-based approach. The most prominent example is the European Union’s Artificial Intelligence Act (AI Act). This article addresses the challenges for adequate risk regulation that arise primarily from the specific type of risks involved, i.e. risks to the protection of fundamental rights and fundamental societal values. This is mainly due to the normative ambiguity of such rights and societal values when attempts are made to select, interpret, specify or operationalise them for the purposes of risk assessments and risk mitigation. This is exemplified by (1) human dignity, (2) informational self-determination, data protection and privacy, (3) anti-discrimination, fairness and justice, and (4) the common good. Normative ambiguities require normative choices, which are assigned to different actors under the regime of the AI Act. Particularly critical normative choices include selecting normative concepts by which to operationalise and specify risks, aggregating and quantifying risks (including the use of metrics), balancing value conflicts, setting levels of acceptable risks, and standardisation. To ensure that these normative choices do not lack democratic legitimacy and to avoid legal uncertainty, further political processes and scientific debates are suggested.
期刊介绍:
NanoEthics: Ethics for Technologies that Converge at the Nanoscale will focus on the philosophically and scientifically rigorous examination of the ethical and societal considerations and the public and policy concerns inherent in nanotechnology research and development. These issues include both individual and societal problems, and include individual health, wellbeing and human enhancement, human integrity and autonomy, distribution of the costs and benefits, threats to culture and tradition and to political and economic stability. Additionally there are meta-issues including the neutrality or otherwise of technology, designing technology in a value-sensitive way, and the control of scientific research.