首页 > 最新文献

Journal of responsible technology最新文献

英文 中文
Ten simple guidelines for decolonising algorithmic systems 非殖民化算法系统的十条简单准则
Pub Date : 2025-07-15 DOI: 10.1016/j.jrt.2025.100125
Dion R.J. O’Neale , Daniel Wilson , Paul T. Brown , Pascarn Dickinson , Manakore Rikus-Graham , Asia Ropeti
As the scope and prevalence of algorithmic systems and artificial intelligence for decision making expand, there is a growing understanding of the need for approaches to help with anticipating adverse consequences and to support the development and deployment of algorithmic systems that are socially responsible and ethically aware. This has led to increasing interest in "decolonising" algorithmic systems as a method of managing and mitigating harms and biases from algorithms and for supporting social benefits from algorithmic decision making for Indigenous peoples.
This article presents ten simple guidelines for giving practical effect to foundational Māori (the Indigenous people of Aotearoa New Zealand) principles in the design, deployment, and operation of algorithmic systems. The guidelines are based on previously established literature regarding ethical use of Māori data. Where possible we have related these guidelines and recommendations to other development practices, for example, to open-source software.
While not intended to be exhaustive or extensive, we hope that these guidelines are able to facilitate and encourage those who work with Māori data in algorithmic systems to engage with processes and practices that support culturally appropriate and ethical approaches for algorithmic systems.
随着用于决策的算法系统和人工智能的范围和流行程度的扩大,人们越来越认识到需要一些方法来帮助预测不利后果,并支持对社会负责和有道德意识的算法系统的开发和部署。这导致人们对“非殖民化”算法系统越来越感兴趣,将其作为一种管理和减轻算法带来的危害和偏见的方法,并支持土著人民从算法决策中获得社会效益。本文提出了十个简单的指导方针,在设计、部署和操作算法系统时,为基础的Māori(新西兰Aotearoa土著人)原则提供实际效果。该指南基于先前建立的关于Māori数据伦理使用的文献。在可能的情况下,我们将这些指导方针和建议与其他开发实践联系起来,例如,与开源软件联系起来。虽然不打算详尽或广泛,但我们希望这些指南能够促进和鼓励那些在算法系统中使用Māori数据的人参与支持文化上适当和道德的算法系统方法的流程和实践。
{"title":"Ten simple guidelines for decolonising algorithmic systems","authors":"Dion R.J. O’Neale ,&nbsp;Daniel Wilson ,&nbsp;Paul T. Brown ,&nbsp;Pascarn Dickinson ,&nbsp;Manakore Rikus-Graham ,&nbsp;Asia Ropeti","doi":"10.1016/j.jrt.2025.100125","DOIUrl":"10.1016/j.jrt.2025.100125","url":null,"abstract":"<div><div>As the scope and prevalence of algorithmic systems and artificial intelligence for decision making expand, there is a growing understanding of the need for approaches to help with anticipating adverse consequences and to support the development and deployment of algorithmic systems that are socially responsible and ethically aware. This has led to increasing interest in \"decolonising\" algorithmic systems as a method of managing and mitigating harms and biases from algorithms and for supporting social benefits from algorithmic decision making for Indigenous peoples.</div><div>This article presents ten simple guidelines for giving practical effect to foundational Māori (the Indigenous people of Aotearoa New Zealand) principles in the design, deployment, and operation of algorithmic systems. The guidelines are based on previously established literature regarding ethical use of Māori data. Where possible we have related these guidelines and recommendations to other development practices, for example, to open-source software.</div><div>While not intended to be exhaustive or extensive, we hope that these guidelines are able to facilitate and encourage those who work with Māori data in algorithmic systems to engage with processes and practices that support culturally appropriate and ethical approaches for algorithmic systems.</div></div>","PeriodicalId":73937,"journal":{"name":"Journal of responsible technology","volume":"23 ","pages":"Article 100125"},"PeriodicalIF":0.0,"publicationDate":"2025-07-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144662246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Participatory research in low resource settings - Endeavours in epistemic justice at the Banyan, India 参与性研究在低资源设置-在榕树认识正义的努力,印度
Pub Date : 2025-06-24 DOI: 10.1016/j.jrt.2025.100123
Mrinalini Ravi , Swarna Tyagi , Vandana Gopikumar , Emma Emily de Wit , Joske Bunders , Deborah Padgett , Barbara Regeer
Involving persons with lived experience in knowledge generation through participatory research (PR) has become increasingly important to challenge power structures in knowledge production and research. In the case of persons with lived experiences of mental illness, participatory research has gained popularity since the early 70 s, but there is little empirical work from countries like India on how PR can be implemented in psychiatric settings.
This study focuses on exploring the way persons with lived experiences of mental illness can be engaged as peer researchers in a service utilisation audit of The Banyan’s inpatient, outpatient and inclusive living facilities. The audit was an attempt by The Banyan to co-opt clients as peer-researchers, thereby enhancing participatory approaches to care planning and provision. Notes and transcripts of research process activities (three meetings for training purposes), 180 interviews as part of the audit, as well as follow up Focus Group Discussions (n = 4) conducted with 18 peer researchers, were used to document their experiences and gather feedback on the training and research process.
We foundthat, reflected against the lack of formal education in the past, the opportunity and support received to be part of a research endeavour, elicited a sense of pride, relief, and liberation in peer researchers. Additionally, actualising the role of an academic and researcher, and not just being passive responders to people in positions of intellectual and systemic power, engendered a sense of responsibility and accountability to peer researchers, and to the mental health system. Thirdly, supporting persons with experiences of mental illness in participatory research activities, especially in the context of low resource settings, requires specific consideration of practical conditions and adjustments needed to avoid tokenism. Finally, both peer- and staff researchers spoke about persisting hierarchies between them which deserve attention.
We conclude that participatory research has a significant scope amongst clients from disadvantaged communities in low-resource settings. Respondents repeatedly expressed an urgency for persons with lived experience to contribute to mental health pedagogy, and, in so doing, disrupt archaic treatment approaches.. Experiences from this enquiry also call for a rethink on how training in research can be developed for individuals without formal education and with cognitive difficulties, with the help of auditory support systemssuch that key concepts are available and accessible, and long-term memory becomes less of a deterrent in the pursuit of knowledge and truth.
通过参与式研究(PR)让有经验的人参与知识生成,对于挑战知识生产和研究中的权力结构变得越来越重要。就有精神疾病经历的人而言,参与性研究自70年代初开始流行,但印度等国家很少有关于如何在精神疾病环境中实施PR的实证研究。本研究的重点是探索有精神疾病生活经历的人可以作为同行研究人员参与榕树医院住院、门诊和包容性生活设施的服务利用审计。这次审计是The Banyan的一次尝试,旨在让客户成为同行研究人员,从而加强护理计划和提供的参与式方法。研究过程活动的笔记和抄本(三次培训会议)、作为审计一部分的180次访谈以及与18名同行研究人员进行的后续焦点小组讨论(n = 4)被用来记录他们的经验并收集关于培训和研究过程的反馈。我们发现,与过去缺乏正规教育相比,参与研究工作的机会和支持让同行研究人员感到自豪、宽慰和解放。此外,实现学术和研究人员的角色,而不仅仅是对智力和系统权力职位上的人的被动反应,产生了对同行研究人员和精神卫生系统的责任感和责任感。第三,在参与性研究活动中支持有精神疾病经历的人,特别是在资源匮乏的情况下,需要具体考虑实际情况和必要的调整,以避免表面现象。最后,同事和员工研究人员都谈到了他们之间持续存在的等级制度,这值得关注。我们的结论是,参与式研究在资源匮乏的弱势社区的客户中有很大的应用范围。答复者一再表示,迫切需要有实际经验的人为精神卫生教育学做出贡献,并在这样做时打破陈旧的治疗方法。这项调查的经验也要求我们重新思考,在听觉支持系统的帮助下,如何为没有受过正规教育和有认知困难的个人开发研究培训,这样关键概念就可以获得和理解,长期记忆在追求知识和真理的过程中就不再是一种阻碍。
{"title":"Participatory research in low resource settings - Endeavours in epistemic justice at the Banyan, India","authors":"Mrinalini Ravi ,&nbsp;Swarna Tyagi ,&nbsp;Vandana Gopikumar ,&nbsp;Emma Emily de Wit ,&nbsp;Joske Bunders ,&nbsp;Deborah Padgett ,&nbsp;Barbara Regeer","doi":"10.1016/j.jrt.2025.100123","DOIUrl":"10.1016/j.jrt.2025.100123","url":null,"abstract":"<div><div>Involving persons with lived experience in knowledge generation through participatory research (PR) has become increasingly important to challenge power structures in knowledge production and research. In the case of persons with lived experiences of mental illness, participatory research has gained popularity since the early 70 s, but there is little empirical work from countries like India on how PR can be implemented in psychiatric settings.</div><div>This study focuses on exploring the way persons with lived experiences of mental illness can be engaged as peer researchers in a service utilisation audit of The Banyan’s inpatient, outpatient and inclusive living facilities. The audit was an attempt by The Banyan to co-opt clients as peer-researchers, thereby enhancing participatory approaches to care planning and provision. Notes and transcripts of research process activities (three meetings for training purposes), 180 interviews as part of the audit, as well as follow up Focus Group Discussions (<em>n</em> = 4) conducted with 18 peer researchers, were used to document their experiences and gather feedback on the training and research process.</div><div>We foundthat, reflected against the lack of formal education in the past, the opportunity and support received to be part of a research endeavour, elicited a sense of pride, relief, and liberation in peer researchers. Additionally, actualising the role of an academic and researcher, and not just being passive responders to people in positions of intellectual and systemic power, engendered a sense of responsibility and accountability to peer researchers, and to the mental health system. Thirdly, supporting persons with experiences of mental illness in participatory research activities, especially in the context of low resource settings, requires specific consideration of practical conditions and adjustments needed to avoid tokenism. Finally, both peer- and staff researchers spoke about persisting hierarchies between them which deserve attention.</div><div>We conclude that participatory research has a significant scope amongst clients from disadvantaged communities in low-resource settings. Respondents repeatedly expressed an urgency for persons with lived experience to contribute to mental health pedagogy, and, in so doing, disrupt archaic treatment approaches.. Experiences from this enquiry also call for a rethink on how training in research can be developed for individuals without formal education and with cognitive difficulties, with the help of auditory support systemssuch that key concepts are available and accessible, and long-term memory becomes less of a deterrent in the pursuit of knowledge and truth.</div></div>","PeriodicalId":73937,"journal":{"name":"Journal of responsible technology","volume":"23 ","pages":"Article 100123"},"PeriodicalIF":0.0,"publicationDate":"2025-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144679387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A capability approach to ethical development and internal auditing of AI technology 人工智能技术伦理发展和内部审计的能力方法
Pub Date : 2025-06-01 DOI: 10.1016/j.jrt.2025.100121
Mark Graves , Emanuele Ratti
Responsible artificial intelligence (AI) requires integrating ethical awareness into the full process of designing and developing AI, including ethics-based auditing of AI technology. We claim the Capability Approach (CA) of Sen and Nussbaum grounds AI ethics in essential human freedoms and can increase awareness of the moral dimension in the technical decision making of developers and data scientists constructing data-centric AI systems. Our use of CA focuses awareness on the ethical impact that day-to-day technical decisions have on the freedom of data subjects to make choices and live meaningful lives according to their own values. For internal auditing of AI technology development, we design and develop a light-weight ethical auditing tool (LEAT) that uses simple natural language processing (NLP) techniques to search design and development documents for relevant ethical characterizations. We describe how CA guides our design, demonstrate LEAT on both principle- and capabilities-based use cases, and characterize its limitations.
负责任的人工智能(AI)需要将道德意识融入人工智能设计和开发的全过程,包括对人工智能技术进行基于道德的审计。我们声称,Sen和Nussbaum的能力方法(CA)将人工智能伦理建立在基本的人类自由之上,可以提高开发人员和数据科学家在构建以数据为中心的人工智能系统的技术决策中的道德维度意识。我们使用CA的重点是关注日常技术决策对数据主体根据自己的价值观做出选择和过有意义的生活的自由产生的道德影响。对于人工智能技术开发的内部审计,我们设计并开发了一个轻量级的道德审计工具(LEAT),该工具使用简单的自然语言处理(NLP)技术来搜索设计和开发文档以获取相关的道德特征。我们描述了CA如何指导我们的设计,在基于原则和基于功能的用例上演示了LEAT,并描述了它的局限性。
{"title":"A capability approach to ethical development and internal auditing of AI technology","authors":"Mark Graves ,&nbsp;Emanuele Ratti","doi":"10.1016/j.jrt.2025.100121","DOIUrl":"10.1016/j.jrt.2025.100121","url":null,"abstract":"<div><div>Responsible artificial intelligence (AI) requires integrating ethical awareness into the full process of designing and developing AI, including ethics-based auditing of AI technology. We claim the Capability Approach (CA) of Sen and Nussbaum grounds AI ethics in essential human freedoms and can increase awareness of the moral dimension in the technical decision making of developers and data scientists constructing data-centric AI systems. Our use of CA focuses awareness on the ethical impact that day-to-day technical decisions have on the freedom of data subjects to make choices and live meaningful lives according to their own values. For internal auditing of AI technology development, we design and develop a light-weight ethical auditing tool (LEAT) that uses simple natural language processing (NLP) techniques to search design and development documents for relevant ethical characterizations. We describe how CA guides our design, demonstrate LEAT on both principle- and capabilities-based use cases, and characterize its limitations.</div></div>","PeriodicalId":73937,"journal":{"name":"Journal of responsible technology","volume":"22 ","pages":"Article 100121"},"PeriodicalIF":0.0,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144243259","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A robot with human values: assessing value-sensitive design in an agri-food context 具有人类价值观的机器人:评估农业食品环境中的价值敏感设计
Pub Date : 2025-04-27 DOI: 10.1016/j.jrt.2025.100120
Else Giesbers , Kelly Rijswijk , Mark Ryan , Mashiat Hossain , Aneesh Chauhan
Value Sensitive Design (VSD) aims to take societal values on board in the design of innovative technologies. While a lot has been written on VSD and the added value of using it for technology development, limited literature is available on its application to the agri-food sector. This article describes a VSD case study on an agri-food robotic system and reflects on the insights into the added value of using VSD. This paper concludes that while VSD contributes to broadening the perspective of technical researchers about non-technical requirements, its application in this case is constrained by five factors related to the nature of the VSD approach: i) lack of clarity on dealing with conflicting values; ii) the ideal timing of VSD is unclear; iii) VSD lacks effectiveness when technology is outsourced; iv) VSD does not account for time and context specificness of values; and v) the operationalisation of values in VSD.
价值敏感设计(VSD)旨在将社会价值纳入创新技术的设计中。虽然关于VSD及其用于技术开发的附加价值的文章很多,但关于其在农业食品部门的应用的文献有限。本文描述了一个农业食品机器人系统的VSD案例研究,并反思了使用VSD的附加值的见解。本文的结论是,虽然VSD有助于拓宽技术研究人员对非技术需求的视角,但它在这种情况下的应用受到与VSD方法性质相关的五个因素的限制:1)在处理冲突价值方面缺乏清晰度;ii) VSD的理想时机尚不明确;iii)当技术外包时,VSD缺乏有效性;iv) VSD没有考虑值的时间和上下文特殊性;v) VSD值的操作。
{"title":"A robot with human values: assessing value-sensitive design in an agri-food context","authors":"Else Giesbers ,&nbsp;Kelly Rijswijk ,&nbsp;Mark Ryan ,&nbsp;Mashiat Hossain ,&nbsp;Aneesh Chauhan","doi":"10.1016/j.jrt.2025.100120","DOIUrl":"10.1016/j.jrt.2025.100120","url":null,"abstract":"<div><div>Value Sensitive Design (VSD) aims to take societal values on board in the design of innovative technologies. While a lot has been written on VSD and the added value of using it for technology development, limited literature is available on its application to the agri-food sector. This article describes a VSD case study on an agri-food robotic system and reflects on the insights into the added value of using VSD. This paper concludes that while VSD contributes to broadening the perspective of technical researchers about non-technical requirements, its application in this case is constrained by five factors related to the nature of the VSD approach: i) lack of clarity on dealing with conflicting values; ii) the ideal timing of VSD is unclear; iii) VSD lacks effectiveness when technology is outsourced; iv) VSD does not account for time and context specificness of values; and v) the operationalisation of values in VSD.</div></div>","PeriodicalId":73937,"journal":{"name":"Journal of responsible technology","volume":"22 ","pages":"Article 100120"},"PeriodicalIF":0.0,"publicationDate":"2025-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143898947","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Decision-making on an AI-supported youth mental health app: A multilogue among ethicists, social scientists, AI-researchers, biomedical engineers, young experiential experts, and psychiatrists 关于人工智能支持的青少年心理健康应用程序的决策:伦理学家、社会科学家、人工智能研究人员、生物医学工程师、年轻经验专家和精神科医生之间的多语对话
Pub Date : 2025-04-18 DOI: 10.1016/j.jrt.2025.100119
Dorothee Horstkötter , Mariël Kanne , Simona Karbouniaris , Noussair Lazrak , Maria Bulgheroni , Ella Sheltawy , Laura Giani , Margherita La Gamba , Esmeralda Ruiz Pujadas , Marina Camacho , Finty Royle , Irene Baggetto , Sinan Gülöksüz , Bart Rutten , Jim van Os
This article explores the decision-making processes in the ongoing development of an AI-supported youth mental health app. Document analysis reveals decisions taken during the grant proposal and funding phase and reflects upon reasons why AI is incorporated in innovative youth mental health care. An innovative multilogue among the transdisciplinary team of researchers, covering ethicists, social scientists, AI-experts, biomedical engineers, young experts by experience, and psychiatrists points out which decisions are taken how. This covers i) the role of a biomedical and exposomic understanding of psychiatry as compared to a phenomenological and experiential perspective, ii) the impact and limits of AI-co-creation by young experts by experience and mental health experts, and iii) the different perspectives regarding the impact of AI on autonomy, empowerment and human relationships. The multilogue does not merely highlight different steps taken during human decision-making in AI-development, it also raises awareness about the many complexities, and sometimes contradictions, when engaging in transdisciplinary work, and it points towards ethical challenges of digitalized youth mental health care.
本文探讨了人工智能支持的青少年心理健康应用程序在持续开发过程中的决策过程。文件分析揭示了在拨款提案和筹资阶段所做的决策,并反思了将人工智能纳入创新型青少年心理健康护理的原因。跨学科研究团队(包括伦理学家、社会科学家、人工智能专家、生物医学工程师、经验丰富的青年专家和精神病学家)之间的创新性多方对话指出了如何做出决定。这包括:i) 生物医学和阐释学对精神病学的理解与现象学和经验学观点的比较;ii) 经验丰富的青年专家和心理健康专家对人工智能共同创造的影响和局限;iii) 人工智能对自主权、赋权和人际关系影响的不同观点。多方对话不仅强调了在人工智能开发过程中人类决策所采取的不同步骤,还提高了人们对跨学科工作中的许多复杂性(有时甚至是矛盾)的认识,并指出了数字化青年心理健康护理所面临的伦理挑战。
{"title":"Decision-making on an AI-supported youth mental health app: A multilogue among ethicists, social scientists, AI-researchers, biomedical engineers, young experiential experts, and psychiatrists","authors":"Dorothee Horstkötter ,&nbsp;Mariël Kanne ,&nbsp;Simona Karbouniaris ,&nbsp;Noussair Lazrak ,&nbsp;Maria Bulgheroni ,&nbsp;Ella Sheltawy ,&nbsp;Laura Giani ,&nbsp;Margherita La Gamba ,&nbsp;Esmeralda Ruiz Pujadas ,&nbsp;Marina Camacho ,&nbsp;Finty Royle ,&nbsp;Irene Baggetto ,&nbsp;Sinan Gülöksüz ,&nbsp;Bart Rutten ,&nbsp;Jim van Os","doi":"10.1016/j.jrt.2025.100119","DOIUrl":"10.1016/j.jrt.2025.100119","url":null,"abstract":"<div><div>This article explores the decision-making processes in the ongoing development of an AI-supported youth mental health app. Document analysis reveals decisions taken during the grant proposal and funding phase and reflects upon reasons <em>why</em> AI is incorporated in innovative youth mental health care. An innovative multilogue among the transdisciplinary team of researchers, covering ethicists, social scientists, AI-experts, biomedical engineers, young experts by experience, and psychiatrists points out <em>which</em> decisions are taken <em>how</em>. This covers i) the role of a biomedical and exposomic understanding of psychiatry as compared to a phenomenological and experiential perspective, ii) the impact and limits of AI-co-creation by young experts by experience and mental health experts, and iii) the different perspectives regarding the impact of AI on autonomy, empowerment and human relationships. The multilogue does not merely highlight different steps taken during human decision-making in AI-development, it also raises awareness about the many complexities, and sometimes contradictions, when engaging in transdisciplinary work, and it points towards ethical challenges of digitalized youth mental health care.</div></div>","PeriodicalId":73937,"journal":{"name":"Journal of responsible technology","volume":"22 ","pages":"Article 100119"},"PeriodicalIF":0.0,"publicationDate":"2025-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143867928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Responsible AI innovation in the public sector: Lessons from and recommendations for facilitating Fundamental Rights and Algorithms Impact Assessments 公共部门负责任的人工智能创新:促进基本权利和算法影响评估的经验教训和建议
Pub Date : 2025-04-03 DOI: 10.1016/j.jrt.2025.100118
I.M. Muis, J. Straatman, B.A. Kamphorst
Since the initial development of the Fundamental Rights and Algorithms Impact Assessment (FRAIA) in 2021, there has been an increasing interest from public sector organizations to gain experience with performing a FRAIA in contexts of developing, procuring, and deploying AI systems. In this contribution, we share observations from fifteen FRAIA trajectories performed in the field within the Dutch public sector context. Based on our experiences facilitating these trajectories, we offer a set of recommendations directed at practitioners with the aim of helping organizations make the best use of FRAIA and similar impact assessment instruments. We conclude by calling for the development of an informal FRAIA community in which practical handholds and advice can be shared to promote responsible AI innovation by ensuring that the human decision making around AI and other algorithms is well informed and well documented with respect to the protection of fundamental rights.
自2021年基本权利和算法影响评估(FRAIA)的初步开发以来,公共部门组织越来越有兴趣在开发、采购和部署人工智能系统的背景下获得执行FRAIA的经验。在这篇文章中,我们分享了在荷兰公共部门背景下实地执行的15个FRAIA轨迹的观察结果。基于我们促进这些轨迹的经验,我们提供了一组针对从业者的建议,目的是帮助组织最好地利用FRAIA和类似的影响评估工具。最后,我们呼吁建立一个非正式的FRAIA社区,在这个社区中,可以分享实践经验和建议,通过确保围绕人工智能和其他算法的人类决策在保护基本权利方面得到充分的了解和充分的记录,促进负责任的人工智能创新。
{"title":"Responsible AI innovation in the public sector: Lessons from and recommendations for facilitating Fundamental Rights and Algorithms Impact Assessments","authors":"I.M. Muis,&nbsp;J. Straatman,&nbsp;B.A. Kamphorst","doi":"10.1016/j.jrt.2025.100118","DOIUrl":"10.1016/j.jrt.2025.100118","url":null,"abstract":"<div><div>Since the initial development of the Fundamental Rights and Algorithms Impact Assessment (FRAIA) in 2021, there has been an increasing interest from public sector organizations to gain experience with performing a FRAIA in contexts of developing, procuring, and deploying AI systems. In this contribution, we share observations from fifteen FRAIA trajectories performed in the field within the Dutch public sector context. Based on our experiences facilitating these trajectories, we offer a set of recommendations directed at practitioners with the aim of helping organizations make the best use of FRAIA and similar impact assessment instruments. We conclude by calling for the development of an informal FRAIA community in which practical handholds and advice can be shared to promote responsible AI innovation by ensuring that the human decision making around AI and other algorithms is well informed and well documented with respect to the protection of fundamental rights.</div></div>","PeriodicalId":73937,"journal":{"name":"Journal of responsible technology","volume":"22 ","pages":"Article 100118"},"PeriodicalIF":0.0,"publicationDate":"2025-04-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143800374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Piloting a maturity model for responsible artificial intelligence: A portuguese case study 负责任人工智能成熟度模型的试点:一个葡萄牙案例研究
Pub Date : 2025-04-02 DOI: 10.1016/j.jrt.2025.100117
Rui Miguel Frazão Dias Ferreira , António GRILO , Maria MAIA
Recently, frameworks and guidelines aiming to assist trustworthiness in organizations and assess ethical issues related to the development and use of Artificial Intelligence (AI) have been translated into self-assessment checklists and other instruments. However, such tools can be very time consuming to apply. Aiming to develop a more practical tool, an Industry-Wide Maturity Model for Responsible AI was piloted in 3 companies and 2 research centres, in Portugal. Results show that organizations are aware of requirements (44 %) to deploy a responsible AI approach and have a reactive response to its implementation, as they are willing to integrate other requirements (33 %) into their business processes. The proposed Model was welcomed and showed openness from companies to consistently use it, since it helped to identify gaps and needs when it comes to foster a more trustworthy approach to the development and deployment of AI.
最近,旨在帮助组织可信度并评估与人工智能(AI)的开发和使用相关的道德问题的框架和指南已被转化为自我评估清单和其他工具。然而,应用这些工具可能非常耗时。为了开发更实用的工具,负责任人工智能的全行业成熟度模型在葡萄牙的3家公司和2个研究中心进行了试点。结果显示,组织意识到部署负责任的人工智能方法的需求(44%),并对其实施做出反应,因为他们愿意将其他需求(33%)集成到他们的业务流程中。拟议的模型受到了公司的欢迎,并显示出持续使用它的开放性,因为它有助于确定差距和需求,从而促进更值得信赖的人工智能开发和部署方法。
{"title":"Piloting a maturity model for responsible artificial intelligence: A portuguese case study","authors":"Rui Miguel Frazão Dias Ferreira ,&nbsp;António GRILO ,&nbsp;Maria MAIA","doi":"10.1016/j.jrt.2025.100117","DOIUrl":"10.1016/j.jrt.2025.100117","url":null,"abstract":"<div><div>Recently, frameworks and guidelines aiming to assist trustworthiness in organizations and assess ethical issues related to the development and use of Artificial Intelligence (AI) have been translated into self-assessment checklists and other instruments. However, such tools can be very time consuming to apply. Aiming to develop a more practical tool, an Industry-Wide Maturity Model for Responsible AI was piloted in 3 companies and 2 research centres, in Portugal. Results show that organizations are aware of requirements (44 %) to deploy a responsible AI approach and have a reactive response to its implementation, as they are willing to integrate other requirements (33 %) into their business processes. The proposed Model was welcomed and showed openness from companies to consistently use it, since it helped to identify gaps and needs when it comes to foster a more trustworthy approach to the development and deployment of AI.</div></div>","PeriodicalId":73937,"journal":{"name":"Journal of responsible technology","volume":"22 ","pages":"Article 100117"},"PeriodicalIF":0.0,"publicationDate":"2025-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143865054","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The ethics of bioinspired animal-robot interaction: A relational meta-ethical approach 受生物启发的动物与机器人互动的伦理:一种关系的元伦理方法
Pub Date : 2025-03-22 DOI: 10.1016/j.jrt.2025.100116
Marco Tamborini
In this article, I focus on a specific aspect of biorobotics: biohybrid interaction between bioinspired robots and animals. My goal is to analyze the ethical and epistemic implications of this practice, starting with a central question: Is it ethically permissible to have a bioinspired robot that mimics and reproduces the behaviors and/or morphology of an animal interact with a particular population, even if the animals do not know that the object they are interacting with is a robot and not a conspecific? My answer to the ethical question is that the interaction between animals and bioinspired robots is ethically acceptable if the animal actively participates in the language game (sense Coeckelbergh) established with the robot. I proceed as follows: First, I define the field of biorobotics and describe its four macro-categories. Second, I present concrete examples of interactive biorobotics, showing two emblematic cases in which the relationship between bioinspired robots and animals plays a central role. Third, I address one key issue—among many—in applied ethics regarding my ethical question. Fourth, I explore the ethical question on a metaethical level, making use of the theories of David Gunkel and Mark Coeckelbergh, as well as the linguistic approach and ethics of the late Ludwig Wittgenstein. Last, I argue that from a meta-ethical approach the original ethical question turns out to be misplaced. The ethical boundary lies not in the distinction between a real or fake relationship between the robot and the organism, but in the degree of mutual participation and understanding between the entities involved.
在本文中,我将重点讨论生物机器人技术的一个具体方面:生物启发机器人与动物之间的生物混合互动。我的目标是从一个核心问题入手,分析这种做法的伦理和认识论意义:让模仿和复制动物行为和/或形态的生物启发机器人与特定种群互动,即使动物不知道与之互动的对象是机器人而非同种动物,这在伦理上是否允许?我对这个伦理问题的回答是,如果动物积极参与与机器人建立的语言游戏(意义上的 Coeckelbergh),那么动物与生物启发机器人之间的互动在伦理上是可以接受的。我的论述如下:首先,我定义了生物机器人领域,并描述了其四个宏观类别。其次,我将介绍互动式生物机器人技术的具体实例,展示生物启发机器人与动物之间的关系在其中发挥核心作用的两个典型案例。第三,我讨论了应用伦理学中有关我的伦理问题的诸多关键问题之一。第四,我利用大卫-冈克尔(David Gunkel)和马克-科克尔伯格(Mark Coeckelbergh)的理论,以及已故路德维希-维特根斯坦(Ludwig Wittgenstein)的语言学方法和伦理学,从元伦理学的层面探讨伦理问题。最后,我认为,从元伦理学的角度来看,最初的伦理问题是错位的。伦理界限不在于机器人与生物体之间是真关系还是假关系,而在于相关实体之间相互参与和理解的程度。
{"title":"The ethics of bioinspired animal-robot interaction: A relational meta-ethical approach","authors":"Marco Tamborini","doi":"10.1016/j.jrt.2025.100116","DOIUrl":"10.1016/j.jrt.2025.100116","url":null,"abstract":"<div><div>In this article, I focus on a specific aspect of biorobotics: biohybrid interaction between bioinspired robots and animals. My goal is to analyze the ethical and epistemic implications of this practice, starting with a central question<em>:</em> Is it ethically permissible to have a bioinspired robot that mimics and reproduces the behaviors and/or morphology of an animal interact with a particular population, even if the animals do not know that the object they are interacting with is a robot and not a conspecific? My answer to the ethical question is that the interaction between animals and bioinspired robots is ethically acceptable if the animal actively participates in the language game (sense Coeckelbergh) established with the robot. I proceed as follows: First, I define the field of biorobotics and describe its four macro-categories. Second, I present concrete examples of interactive biorobotics, showing two emblematic cases in which the relationship between bioinspired robots and animals plays a central role. Third, I address one key issue—among many—in applied ethics regarding my ethical question. Fourth, I explore the ethical question on a metaethical level, making use of the theories of David Gunkel and Mark Coeckelbergh, as well as the linguistic approach and ethics of the late Ludwig Wittgenstein. Last, I argue that from a meta-ethical approach the original ethical question turns out to be misplaced. The ethical boundary lies not in the distinction between a real or fake relationship between the robot and the organism, but in the degree of mutual participation and understanding between the entities involved.</div></div>","PeriodicalId":73937,"journal":{"name":"Journal of responsible technology","volume":"22 ","pages":"Article 100116"},"PeriodicalIF":0.0,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143704339","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reflexivity and AI start-ups: A collective virtue for dynamic teams 反身性和人工智能初创企业:动态团队的集体美德
Pub Date : 2025-03-14 DOI: 10.1016/j.jrt.2025.100115
Marco Innocenti
This paper investigates the ethical challenges faced by AI-driven start-ups, where the rapid pace of innovation and limited resources often preclude team members from fully understanding the product under development or its societal implications. We propose the concept of “swarm moral reflexivity”, where ethical reflection emerges collectively from the interactions of individuals focused on their specific tasks. Drawing on Swarm Intelligence theories and Alasdair MacIntyre's framework of moral deliberation, this approach enables teams to engage with ethical issues through daily encounters with conflicting responsibilities, rather than relying on top-down value systems or comprehensive ethical oversight. Our model suggests that decentralised, collective moral awareness can effectively support Responsible Innovation in AI start-ups, ensuring that ethical concerns are recognised and addressed throughout the development process, even in fast-paced and resource-constrained environments.
本文调查了人工智能驱动的初创企业所面临的道德挑战,在这些初创企业中,快速的创新步伐和有限的资源往往使团队成员无法充分理解正在开发的产品或其社会影响。我们提出了“群体道德反身性”的概念,即伦理反思从专注于特定任务的个体的相互作用中集体出现。利用群体智能理论和Alasdair MacIntyre的道德审议框架,这种方法使团队能够通过日常遇到的相互冲突的责任来处理道德问题,而不是依赖于自上而下的价值体系或全面的道德监督。我们的模型表明,分散的、集体的道德意识可以有效地支持人工智能初创企业的负责任创新,确保在整个开发过程中,即使在快节奏和资源受限的环境中,也能认识和解决道德问题。
{"title":"Reflexivity and AI start-ups: A collective virtue for dynamic teams","authors":"Marco Innocenti","doi":"10.1016/j.jrt.2025.100115","DOIUrl":"10.1016/j.jrt.2025.100115","url":null,"abstract":"<div><div>This paper investigates the ethical challenges faced by AI-driven start-ups, where the rapid pace of innovation and limited resources often preclude team members from fully understanding the product under development or its societal implications. We propose the concept of “swarm moral reflexivity”, where ethical reflection emerges collectively from the interactions of individuals focused on their specific tasks. Drawing on Swarm Intelligence theories and Alasdair MacIntyre's framework of moral deliberation, this approach enables teams to engage with ethical issues through daily encounters with conflicting responsibilities, rather than relying on top-down value systems or comprehensive ethical oversight. Our model suggests that decentralised, collective moral awareness can effectively support Responsible Innovation in AI start-ups, ensuring that ethical concerns are recognised and addressed throughout the development process, even in fast-paced and resource-constrained environments.</div></div>","PeriodicalId":73937,"journal":{"name":"Journal of responsible technology","volume":"22 ","pages":"Article 100115"},"PeriodicalIF":0.0,"publicationDate":"2025-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143685401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Normative conflict resolution through human–autonomous agent interaction 通过人类自主代理交互规范冲突解决
Pub Date : 2025-03-01 DOI: 10.1016/j.jrt.2025.100114
Beverley Townsend , Katie J. Parnell , Sinem Getir Yaman , Gabriel Nemirovsky , Radu Calinescu
We have become increasingly reliant on the decision-making capabilities of autonomous agents. These decisions are often executed under non-ideal conditions, offer significant moral risk, and directly affect human well-being. Such decisions may involve the choice to optimise one value over another: promoting safety over human autonomy, or ensuring accuracy over fairness, for example. All too often decision-making of this kind requires a level of normative evaluation involving ethically defensible moral choices and value judgements, compromises, and trade-offs. Guided by normative principles such decisions inform the possible courses of action the agent may take and may even change a set of established actionable courses.
This paper seeks to map the decision-making processes in normative choice scenarios wherein autonomous agents are intrinsically linked to the decision process. A care-robot is used to illustrate how a normative choice - underpinned by normative principles - arises, where the agent must ‘choose’ an actionable path involving the administration of critical or non-critical medication. Critically, the choice is dependent upon the trade-off involving two normative principles: respect for human autonomy and the prevention of harm. An additional dimension is presented, that of the inclusion of the urgency of the medication to be administered, which further informs and changes the course of action to be followed.
We offer a means to map decision-making involving a normative choice within a decision ladder using stakeholder input, and, using defeasibility, we show how specification rules with defeaters can be written to operationalise such choice.
我们越来越依赖自主代理的决策能力。这些决定通常是在非理想条件下执行的,会带来重大的道德风险,并直接影响人类的福祉。这样的决策可能涉及选择优化一种价值而不是另一种价值:例如,促进安全而不是人类自主,或确保准确性而不是公平性。这类决策往往需要一定程度的规范评估,包括道德上可辩护的道德选择和价值判断、妥协和权衡。在规范原则的指导下,这些决定告知代理人可能采取的行动路线,甚至可能改变一套既定的可行路线。本文试图映射规范性选择场景中的决策过程,其中自主代理与决策过程有着内在的联系。护理机器人被用来说明如何在规范原则的基础上产生规范选择,其中代理必须“选择”一条可操作的路径,涉及关键或非关键药物的管理。关键的是,这种选择取决于涉及两个规范原则的权衡:尊重人类自主权和防止伤害。提出了一个额外的方面,即纳入所要施用药物的紧迫性,这进一步通知并改变了所要遵循的行动方针。我们提供了一种方法,利用利益相关者的输入,在决策阶梯中映射涉及规范性选择的决策,并且,使用可撤销性,我们展示了如何编写具有撤销者的规范规则来实现此类选择。
{"title":"Normative conflict resolution through human–autonomous agent interaction","authors":"Beverley Townsend ,&nbsp;Katie J. Parnell ,&nbsp;Sinem Getir Yaman ,&nbsp;Gabriel Nemirovsky ,&nbsp;Radu Calinescu","doi":"10.1016/j.jrt.2025.100114","DOIUrl":"10.1016/j.jrt.2025.100114","url":null,"abstract":"<div><div>We have become increasingly reliant on the decision-making capabilities of autonomous agents. These decisions are often executed under non-ideal conditions, offer significant moral risk, and directly affect human well-being. Such decisions may involve the choice to optimise one value over another: promoting safety over human autonomy, or ensuring accuracy over fairness, for example. All too often decision-making of this kind requires a level of normative evaluation involving ethically defensible moral choices and value judgements, compromises, and trade-offs. Guided by normative principles such decisions inform the possible courses of action the agent may take and may even change a set of established actionable courses.</div><div>This paper seeks to map the decision-making processes in normative choice scenarios wherein autonomous agents are intrinsically linked to the decision process. A care-robot is used to illustrate how a normative choice - underpinned by normative principles - arises, where the agent must ‘choose’ an actionable path involving the administration of critical or non-critical medication. Critically, the choice is dependent upon the trade-off involving two normative principles: respect for human autonomy and the prevention of harm. An additional dimension is presented, that of the inclusion of the urgency of the medication to be administered, which further informs and changes the course of action to be followed.</div><div>We offer a means to map decision-making involving a normative choice within a decision ladder using stakeholder input, and, using defeasibility, we show how specification rules with defeaters can be written to operationalise such choice.</div></div>","PeriodicalId":73937,"journal":{"name":"Journal of responsible technology","volume":"21 ","pages":"Article 100114"},"PeriodicalIF":0.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143578378","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of responsible technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1