首页 > 最新文献

AI and ethics最新文献

英文 中文
“ChatGPT says no”: agency, trust, and blame in Twitter discourses after the launch of ChatGPT ChatGPT 说 "不":ChatGPT 推出后 Twitter 讨论中的代理、信任和指责
Pub Date : 2024-01-23 DOI: 10.1007/s43681-023-00414-1
Dan Heaton, Elena Nichele, Jeremie Clos, Joel E. Fischer

ChatGPT, a chatbot using the GPT-n series large language model, has surged in popularity by providing conversation, assistance, and entertainment. This has raised questions about its agency and resulting implications on trust and blame, particularly when concerning its portrayal on social media platforms like Twitter. Understanding trust and blame is crucial for gauging public perception, reliance on, and adoption of AI-driven tools like ChatGPT. To explore ChatGPT’s perceived status as an algorithmic social actor and uncover implications for trust and blame through agency and transitivity, we examined 88,058 tweets about ChatGPT, published in a ‘hype period’ between November 2022 and March 2023, using Corpus Linguistics and Critical Discourse Analysis, underpinned by Social Actor Representation. Notably, ChatGPT was presented in tweets as a social actor on 87% of occasions, using personalisation and agency metaphor to emphasise its role in content creation, information dissemination, and influence. However, a dynamic presentation, oscillating between a creative social actor and an information source, reflected users’ uncertainty regarding its capabilities and, thus, blame attribution occurred. On 13% of occasions, ChatGPT was presented passively through backgrounding and exclusion. Here, the emphasis on ChatGPT’s role in informing and influencing underscores interactors’ reliance on it for information, bearing implications for information dissemination and trust in AI-generated content. Therefore, this study contributes to understanding the perceived social agency of decision-making algorithms and their implications on trust and blame, valuable to AI developers and policymakers and relevant in comprehending and dealing with power dynamics in today’s age of AI.

{"title":"“ChatGPT says no”: agency, trust, and blame in Twitter discourses after the launch of ChatGPT","authors":"Dan Heaton,&nbsp;Elena Nichele,&nbsp;Jeremie Clos,&nbsp;Joel E. Fischer","doi":"10.1007/s43681-023-00414-1","DOIUrl":"10.1007/s43681-023-00414-1","url":null,"abstract":"<div><p>ChatGPT, a chatbot using the GPT-n series large language model, has surged in popularity by providing conversation, assistance, and entertainment. This has raised questions about its agency and resulting implications on trust and blame, particularly when concerning its portrayal on social media platforms like Twitter. Understanding trust and blame is crucial for gauging public perception, reliance on, and adoption of AI-driven tools like ChatGPT. To explore ChatGPT’s perceived status as an algorithmic social actor and uncover implications for trust and blame through agency and transitivity, we examined 88,058 tweets about ChatGPT, published in a ‘hype period’ between November 2022 and March 2023, using Corpus Linguistics and Critical Discourse Analysis, underpinned by Social Actor Representation. Notably, ChatGPT was presented in tweets as a social actor on 87% of occasions, using personalisation and agency metaphor to emphasise its role in content creation, information dissemination, and influence. However, a dynamic presentation, oscillating between a creative social actor and an information source, reflected users’ uncertainty regarding its capabilities and, thus, blame attribution occurred. On 13% of occasions, ChatGPT was presented passively through backgrounding and exclusion. Here, the emphasis on ChatGPT’s role in informing and influencing underscores interactors’ reliance on it for information, bearing implications for information dissemination and trust in AI-generated content. Therefore, this study contributes to understanding the perceived social agency of decision-making algorithms and their implications on trust and blame, valuable to AI developers and policymakers and relevant in comprehending and dealing with power dynamics in today’s age of AI.</p></div>","PeriodicalId":72137,"journal":{"name":"AI and ethics","volume":"5 1","pages":"653 - 675"},"PeriodicalIF":0.0,"publicationDate":"2024-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43681-023-00414-1.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139603612","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Conformal prediction for trustworthy detection of railway signals 用于铁路信号可信检测的共形预测
Pub Date : 2024-01-22 DOI: 10.1007/s43681-023-00400-7
Léo Andéol, Thomas Fel, Florence de Grancey, Luca Mossina

We present an application of conformal prediction, a form of uncertainty quantification with guarantees, to the detection of railway signals. State-of-the-art architectures are tested and the most promising one undergoes the process of conformalization, where a correction is applied to the predicted bounding boxes (i.e., to their height and width) such that they comply with a predefined probability of success. We work with a novel exploratory dataset of images taken from the perspective of a train operator, as a first step to build and validate future trustworthy machine learning models for the detection of railway signals.

我们介绍了保形预测在铁路信号检测中的应用,这是一种有保证的不确定性量化形式。我们对最先进的架构进行了测试,并对最有前途的架构进行了保形化处理,即对预测的边界框(即高度和宽度)进行修正,使其符合预定的成功概率。我们从列车操作员的视角出发,使用了一个新颖的探索性图像数据集,作为建立和验证未来用于检测铁路信号的可信机器学习模型的第一步。
{"title":"Conformal prediction for trustworthy detection of railway signals","authors":"Léo Andéol,&nbsp;Thomas Fel,&nbsp;Florence de Grancey,&nbsp;Luca Mossina","doi":"10.1007/s43681-023-00400-7","DOIUrl":"10.1007/s43681-023-00400-7","url":null,"abstract":"<div><p>We present an application of conformal prediction, a form of uncertainty quantification with guarantees, to the detection of railway signals. State-of-the-art architectures are tested and the most promising one undergoes the process of conformalization, where a correction is applied to the predicted bounding boxes (i.e., to their height and width) such that they comply with a predefined probability of success. We work with a novel exploratory dataset of images taken from the perspective of a train operator, as a first step to build and validate future trustworthy machine learning models for the detection of railway signals.</p></div>","PeriodicalId":72137,"journal":{"name":"AI and ethics","volume":"4 1","pages":"157 - 161"},"PeriodicalIF":0.0,"publicationDate":"2024-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142413036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A-XAI: adversarial machine learning for trustable explainability A-XAI:可信任可解释性的对抗式机器学习
Pub Date : 2024-01-17 DOI: 10.1007/s43681-023-00368-4
Nishita Agrawal, Isha Pendharkar, Jugal Shroff, Jatin Raghuvanshi, Akashdip Neogi, Shruti Patil, Rahee Walambe, Ketan Kotecha

With the recent advancements in the usage of Artificial Intelligence (AI)-based systems in the healthcare and medical domain, it has become necessary to monitor whether these systems make predictions using the correct features or not. For this purpose, many different types of model interpretability and explainability methods are proposed in the literature. However, with the rising number of adversarial attacks against these AI-based systems, it also becomes necessary to make those systems more robust to adversarial attacks and validate the correctness of the generated model explainability. In this work, we first demonstrate how an adversarial attack can affect the model explainability even after robust training. Along with this, we present two different types of attack classifiers: one that can detect whether the given input is benign or adversarial and the other classifier that can identify the type of attack. We also identify the regions affected by the adversarial attack using model explainability. Finally, we demonstrate how the correctness of the generated explainability can be verified using model interpretability methods.

随着最近基于人工智能(AI)的系统在医疗保健和医学领域的应用不断发展,有必要对这些系统是否使用正确的特征进行预测进行监测。为此,文献中提出了许多不同类型的模型可解释性和可解释性方法。然而,随着针对这些基于人工智能的系统的对抗性攻击日益增多,也有必要使这些系统对对抗性攻击更具鲁棒性,并验证生成的模型可解释性的正确性。在这项工作中,我们首先展示了对抗性攻击是如何影响模型的可解释性的,即使是在鲁棒性训练之后。与此同时,我们提出了两种不同类型的攻击分类器:一种可以检测给定输入是良性的还是对抗性的,另一种可以识别攻击类型。我们还利用模型的可解释性确定了受对抗性攻击影响的区域。最后,我们演示了如何使用模型可解释性方法验证生成的可解释性的正确性。
{"title":"A-XAI: adversarial machine learning for trustable explainability","authors":"Nishita Agrawal,&nbsp;Isha Pendharkar,&nbsp;Jugal Shroff,&nbsp;Jatin Raghuvanshi,&nbsp;Akashdip Neogi,&nbsp;Shruti Patil,&nbsp;Rahee Walambe,&nbsp;Ketan Kotecha","doi":"10.1007/s43681-023-00368-4","DOIUrl":"10.1007/s43681-023-00368-4","url":null,"abstract":"<div><p>With the recent advancements in the usage of Artificial Intelligence (AI)-based systems in the healthcare and medical domain, it has become necessary to monitor whether these systems make predictions using the correct features or not. For this purpose, many different types of model interpretability and explainability methods are proposed in the literature. However, with the rising number of adversarial attacks against these AI-based systems, it also becomes necessary to make those systems more robust to adversarial attacks and validate the correctness of the generated model explainability. In this work, we first demonstrate how an adversarial attack can affect the model explainability even after robust training. Along with this, we present two different types of attack classifiers: one that can detect whether the given input is benign or adversarial and the other classifier that can identify the type of attack. We also identify the regions affected by the adversarial attack using model explainability. Finally, we demonstrate how the correctness of the generated explainability can be verified using model interpretability methods.</p></div>","PeriodicalId":72137,"journal":{"name":"AI and ethics","volume":"4 4","pages":"1143 - 1174"},"PeriodicalIF":0.0,"publicationDate":"2024-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139616596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A design perspective on how to tackle gender biases when developing AI-driven systems 从设计角度看如何在开发人工智能驱动系统时解决性别偏见问题
Pub Date : 2024-01-15 DOI: 10.1007/s43681-023-00386-2
Ana Santana González, Lucia Rampino

A growing awareness of bias in artificial intelligence (AI) systems has recently emerged, leading to an increased number of publications discussing ethics in AI. Nevertheless, the specific issue of gender bias remains under-discussed. How can design contribute to preventing the emergence of gender bias in AI-driven systems? To answer this question, we investigated the current state of AI ethical guidelines within the European Union. The results revealed that most guidelines do not acknowledge gender bias but address discrimination. This raised our concerns, as addressing multiple biases simultaneously might not effectively mitigate any of them due to their often-unconscious nature. Furthermore, our results revealed a lack of quantitative evidence supporting the effectiveness of bias prevention implementation methods and solutions. In conclusion, based on our analysis, we propose four recommendations for designing effective guidelines to tackle gender biases in AI. Moreover, we stress the central role of diversity in embedding the gender perspective from the beginning in any design activity.

{"title":"A design perspective on how to tackle gender biases when developing AI-driven systems","authors":"Ana Santana González,&nbsp;Lucia Rampino","doi":"10.1007/s43681-023-00386-2","DOIUrl":"10.1007/s43681-023-00386-2","url":null,"abstract":"<div><p>A growing awareness of bias in artificial intelligence (AI) systems has recently emerged, leading to an increased number of publications discussing ethics in AI. Nevertheless, the specific issue of gender bias remains under-discussed. How can design contribute to preventing the emergence of gender bias in AI-driven systems? To answer this question, we investigated the current state of AI ethical guidelines within the European Union. The results revealed that most guidelines do not acknowledge gender bias but address discrimination. This raised our concerns, as addressing multiple biases simultaneously might not effectively mitigate any of them due to their often-unconscious nature. Furthermore, our results revealed a lack of quantitative evidence supporting the effectiveness of bias prevention implementation methods and solutions. In conclusion, based on our analysis, we propose four recommendations for designing effective guidelines to tackle gender biases in AI. Moreover, we stress the central role of diversity in embedding the gender perspective from the beginning in any design activity.</p></div>","PeriodicalId":72137,"journal":{"name":"AI and ethics","volume":"5 1","pages":"201 - 218"},"PeriodicalIF":0.0,"publicationDate":"2024-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43681-023-00386-2.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139622105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Positive and negative explanation effects in human–agent teams 人类-代理团队中的积极和消极解释效应
Pub Date : 2024-01-10 DOI: 10.1007/s43681-023-00396-0
Bryan Lavender, Sami Abuhaimed, Sandip Sen

Improving agent capabilities and increasing availability of computing platforms and internet connectivity allows more effective and diverse collaboration between human users and automated agents. To increase the viability and effectiveness of human–agent collaborative teams, there is a pressing need for research enabling such teams to maximally leverage relative strengths of human and automated reasoners. We study virtual and ad hoc teams, comprising a human and an agent, collaborating over a few episodes where each episode requires them to complete a set of tasks chosen from given task types. Team members are initially unaware of their partner’s capabilities, and the agent, acting as the task allocator, must adapt the allocation process to maximize team performance. The focus of the current paper is on analyzing how allocation decision explanations can affect both user performance and the human workers’ outlook, including factors, such as motivation and satisfaction. We investigate the effect of explanations provided by the agent allocator to the human on performance and key factors reported by the human teammate on surveys. Survey factors include the effect of explanations on motivation, explanatory power, and understandability, as well as satisfaction with and trust/confidence in the teammate. We evaluated a set of hypotheses on these factors related to positive, negative, and no-explanation scenarios through experiments conducted with MTurk workers.

随着代理能力的提高、计算平台和互联网连接的日益普及,人类用户与自动代理之间的合作变得更加有效和多样化。为了提高人类-代理协作团队的可行性和有效性,迫切需要开展研究,使这些团队能够最大限度地利用人类和自动推理者的相对优势。我们研究了由一名人类和一名代理组成的虚拟临时团队,他们在几个事件中进行合作,每个事件要求他们完成从给定任务类型中选择的一组任务。团队成员最初并不了解其伙伴的能力,而作为任务分配者的代理必须调整任务分配过程,以最大限度地提高团队绩效。本文的重点是分析分配决策的解释如何影响用户的绩效和人类工作者的前景,包括动机和满意度等因素。我们研究了代理分配者向人类提供的解释对人类队友在调查中报告的绩效和关键因素的影响。调查因素包括解释对动机、解释力和可理解性的影响,以及对队友的满意度和信任/信心。我们通过对 MTurk 员工进行实验,评估了与积极、消极和无解释情景相关的这些因素的一系列假设。
{"title":"Positive and negative explanation effects in human–agent teams","authors":"Bryan Lavender,&nbsp;Sami Abuhaimed,&nbsp;Sandip Sen","doi":"10.1007/s43681-023-00396-0","DOIUrl":"10.1007/s43681-023-00396-0","url":null,"abstract":"<div><p>Improving agent capabilities and increasing availability of computing platforms and internet connectivity allows more effective and diverse collaboration between human users and automated agents. To increase the viability and effectiveness of human–agent collaborative teams, there is a pressing need for research enabling such teams to maximally leverage relative strengths of human and automated reasoners. We study virtual and ad hoc teams, comprising a human and an agent, collaborating over a few episodes where each episode requires them to complete a set of tasks chosen from given task types. Team members are initially unaware of their partner’s capabilities, and the agent, acting as the task allocator, must adapt the allocation process to maximize team performance. The focus of the current paper is on analyzing how allocation decision explanations can affect both user performance and the human workers’ outlook, including factors, such as motivation and satisfaction. We investigate the effect of explanations provided by the agent allocator to the human on performance and key factors reported by the human teammate on surveys. Survey factors include the effect of explanations on motivation, explanatory power, and understandability, as well as satisfaction with and trust/confidence in the teammate. We evaluated a set of hypotheses on these factors related to positive, negative, and no-explanation scenarios through experiments conducted with MTurk workers.</p></div>","PeriodicalId":72137,"journal":{"name":"AI and ethics","volume":"4 1","pages":"47 - 56"},"PeriodicalIF":0.0,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139438844","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gender mobility in the labor market with skills-based matching models 劳动力市场中的性别流动性与基于技能的匹配模型
Pub Date : 2024-01-09 DOI: 10.1007/s43681-023-00410-5
Ajaya Adhikari, Steven Vethman, Daan Vos, Marc Lenz, Ioana Cocu, Ioannis Tolios, Cor J. Veenman

Skills-based matching promises mobility of workers between different sectors and occupations in the labor market. In this case, job seekers can look for jobs they do not yet have experience in, but for which they do have relevant skills. Currently, there are multiple occupations with a skewed gender distribution. For skills-based matching, it is unclear if and how a shift in the gender distribution, which we call gender mobility, between occupations will be effected. It is expected that the skills-based matching approach will likely be data-driven, including computational language models and supervised learning methods. This work, first, shows the presence of gender segregation in language model-based skills representation of occupations. Second, we assess the use of these representations in a potential application based on simulated data, and show that the gender segregation is propagated by various data-driven skills-based matching models. These models are based on different language representations (bag of words, word2vec, and BERT), and distance metrics (static and machine learning-based). Accordingly, we show how skills-based matching approaches can be evaluated and compared on matching performance as well as on the risk of gender segregation. Making the gender segregation bias of models more explicit can help in generating healthy trust in the use of these models in practice.

以技能为基础的匹配可以使工人在劳动力市场的不同部门和职业之间流动。在这种情况下,求职者可以寻找他们还没有工作经验,但拥有相关技能的工作。目前,有多种职业的性别分布存在偏差。就技能配对而言,目前还不清楚性别分布的变化(我们称之为性别流动)是否会产生影响以及如何产生影响。预计基于技能的匹配方法很可能是数据驱动的,包括计算语言模型和监督学习方法。这项工作首先显示了基于语言模型的职业技能表征中存在性别隔离。其次,我们以模拟数据为基础,评估了这些表征在潜在应用中的使用情况,并表明性别隔离会通过各种数据驱动的技能匹配模型传播。这些模型基于不同的语言表征(词包、word2vec 和 BERT)和距离度量(静态和基于机器学习)。因此,我们展示了如何评估和比较基于技能的匹配方法的匹配性能以及性别隔离风险。使模型的性别隔离偏差更加明确,有助于在实践中使用这些模型时产生健康的信任。
{"title":"Gender mobility in the labor market with skills-based matching models","authors":"Ajaya Adhikari,&nbsp;Steven Vethman,&nbsp;Daan Vos,&nbsp;Marc Lenz,&nbsp;Ioana Cocu,&nbsp;Ioannis Tolios,&nbsp;Cor J. Veenman","doi":"10.1007/s43681-023-00410-5","DOIUrl":"10.1007/s43681-023-00410-5","url":null,"abstract":"<div><p>Skills-based matching promises mobility of workers between different sectors and occupations in the labor market. In this case, job seekers can look for jobs they do not yet have experience in, but for which they do have relevant skills. Currently, there are multiple occupations with a skewed gender distribution. For skills-based matching, it is unclear if and how a shift in the gender distribution, which we call <i>gender mobility</i>, between occupations will be effected. It is expected that the skills-based matching approach will likely be data-driven, including computational language models and supervised learning methods. This work, first, shows the presence of gender segregation in language model-based skills representation of occupations. Second, we assess the use of these representations in a potential application based on simulated data, and show that the gender segregation is propagated by various data-driven skills-based matching models. These models are based on different language representations (bag of words, word2vec, and BERT), and distance metrics (static and machine learning-based). Accordingly, we show how skills-based matching approaches can be evaluated and compared on matching <i>performance</i> as well as on the risk of <i>gender segregation</i>. Making the gender segregation bias of models more explicit can help in generating healthy trust in the use of these models in practice.</p></div>","PeriodicalId":72137,"journal":{"name":"AI and ethics","volume":"4 1","pages":"163 - 167"},"PeriodicalIF":0.0,"publicationDate":"2024-01-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142410794","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the evaluation of the symbolic knowledge extracted from black boxes 关于评估从黑匣子中提取的符号知识
Pub Date : 2024-01-09 DOI: 10.1007/s43681-023-00406-1
Federico Sabbatini, Roberta Calegari

As opaque decision systems are being increasingly adopted in almost any application field, issues about their lack of transparency and human readability are a concrete concern for end-users. Amongst existing proposals to associate human-interpretable knowledge with accurate predictions provided by opaque models, there are rule extraction techniques, capable of extracting symbolic knowledge out of opaque models. The quantitative assessment of the extracted knowledge’s quality is still an open issue. For this reason, we provide here a first approach to measure the knowledge quality, encompassing several indicators and providing a compact score reflecting readability, completeness and predictive performance associated with a symbolic knowledge representation. We also discuss the main criticalities behind our proposal, related to the readability assessment and evaluation, to push future research efforts towards a more robust score formulation.

随着不透明决策系统在几乎所有应用领域中被越来越多地采用,其缺乏透明度和人类可读性的问题成为终端用户的具体关切。在现有的将人类可读知识与不透明模型提供的准确预测联系起来的建议中,有一些规则提取技术能够从不透明模型中提取符号知识。对提取知识的质量进行定量评估仍是一个未决问题。因此,我们在此提供了第一种衡量知识质量的方法,其中包含多个指标,并提供了一个紧凑的分数,反映了与符号知识表示相关的可读性、完整性和预测性能。我们还讨论了我们的建议背后与可读性评估和评价相关的主要关键点,以推动未来的研究工作朝着更稳健的分数表述方向发展。
{"title":"On the evaluation of the symbolic knowledge extracted from black boxes","authors":"Federico Sabbatini,&nbsp;Roberta Calegari","doi":"10.1007/s43681-023-00406-1","DOIUrl":"10.1007/s43681-023-00406-1","url":null,"abstract":"<div><p>As opaque decision systems are being increasingly adopted in almost any application field, issues about their lack of transparency and human readability are a concrete concern for end-users. Amongst existing proposals to associate human-interpretable knowledge with accurate predictions provided by opaque models, there are rule extraction techniques, capable of extracting symbolic knowledge out of opaque models. The quantitative assessment of the extracted knowledge’s quality is still an open issue. For this reason, we provide here a first approach to measure the knowledge quality, encompassing several indicators and providing a compact score reflecting readability, completeness and predictive performance associated with a symbolic knowledge representation. We also discuss the main criticalities behind our proposal, related to the readability assessment and evaluation, to push future research efforts towards a more robust score formulation.</p></div>","PeriodicalId":72137,"journal":{"name":"AI and ethics","volume":"4 1","pages":"65 - 74"},"PeriodicalIF":0.0,"publicationDate":"2024-01-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139441766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An overview of key trustworthiness attributes and KPIs for trusted ML-based systems engineering 基于 ML 的可信系统工程的关键可信度属性和关键绩效指标概览
Pub Date : 2024-01-08 DOI: 10.1007/s43681-023-00394-2
Juliette Mattioli, Henri Sohier, Agnès Delaborde, Kahina Amokrane-Ferka, Afef Awadid, Zakaria Chihani, Souhaiel Khalfaoui, Gabriel Pedroza

When deployed, machine-learning (ML) adoption depends on its ability to actually deliver the expected service safely, and to meet user expectations in terms of quality and continuity of service. For instance, the users expect that the technology will not do something it is not supposed to do, e.g., performing actions without informing users. Thus, the use of Artificial Intelligence (AI) in safety-critical systems such as in avionics, mobility, defense, and healthcare requires proving their trustworthiness through out its overall lifecycle (from design to deployment). Based on surveys on quality measures, characteristics and sub-characteristics of AI systems, the Confiance.ai program (www.confiance.ai) aims to identify the relevant trustworthiness attributes and their associated key performance indicators (KPI) or their associated methods for assessing the induced level of trust.

机器学习(ML)的应用取决于其是否能够安全地实际提供预期服务,以及是否能够在服务质量和连续性方面满足用户的期望。例如,用户希望技术不会做不该做的事情,如在未通知用户的情况下执行操作。因此,在航空电子、移动、国防和医疗保健等安全关键系统中使用人工智能(AI),需要证明其在整个生命周期(从设计到部署)中的可信度。Confiance.ai 计划 (www.confiance.ai) 基于对人工智能系统的质量测量、特征和子特征的调查,旨在确定相关的可信度属性及其相关的关键性能指标 (KPI),或评估诱导信任度的相关方法。
{"title":"An overview of key trustworthiness attributes and KPIs for trusted ML-based systems engineering","authors":"Juliette Mattioli,&nbsp;Henri Sohier,&nbsp;Agnès Delaborde,&nbsp;Kahina Amokrane-Ferka,&nbsp;Afef Awadid,&nbsp;Zakaria Chihani,&nbsp;Souhaiel Khalfaoui,&nbsp;Gabriel Pedroza","doi":"10.1007/s43681-023-00394-2","DOIUrl":"10.1007/s43681-023-00394-2","url":null,"abstract":"<div><p>When deployed, machine-learning (ML) adoption depends on its ability to actually deliver the expected service safely, and to meet user expectations in terms of quality and continuity of service. For instance, the users expect that the technology will not do something it is not supposed to do, e.g., performing actions without informing users. Thus, the use of Artificial Intelligence (AI) in safety-critical systems such as in avionics, mobility, defense, and healthcare requires proving their trustworthiness through out its overall lifecycle (from design to deployment). Based on surveys on quality measures, characteristics and sub-characteristics of AI systems, the Confiance.ai program (www.confiance.ai) aims to identify the relevant trustworthiness attributes and their associated key performance indicators (KPI) or their associated methods for assessing the induced level of trust.</p></div>","PeriodicalId":72137,"journal":{"name":"AI and ethics","volume":"4 1","pages":"15 - 25"},"PeriodicalIF":0.0,"publicationDate":"2024-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139447649","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Assessing systematic weaknesses of DNNs using counterfactuals 利用反事实评估 DNN 的系统性弱点
Pub Date : 2024-01-08 DOI: 10.1007/s43681-023-00407-0
Sujan Sai Gannamaneni, Michael Mock, Maram Akila

With the advancement of DNNs into safety-critical applications, testing approaches for such models have gained more attention. A current direction is the search for and identification of systematic weaknesses that put safety assumptions based on average performance values at risk. Such weaknesses can take on the form of (semantically coherent) subsets or areas in the input space where a DNN performs systematically worse than its expected average. However, it is non-trivial to attribute the reason for such observed low performances to the specific semantic features that describe the subset. For instance, inhomogeneities within the data w.r.t. other (non-considered) attributes might distort results. However, taking into account all (available) attributes and their interaction is often computationally highly expensive. Inspired by counterfactual explanations, we propose an effective and computationally cheap algorithm to validate the semantic attribution of existing subsets, i.e., to check whether the identified attribute is likely to have caused the degraded performance. We demonstrate this approach on an example from the autonomous driving domain using highly annotated simulated data, where we show for a semantic segmentation model that (i) performance differences among the different pedestrian assets exist, but (ii) only in some cases is the asset type itself the reason for this reduction in the performance.

随着 DNN 在安全关键型应用中的发展,此类模型的测试方法得到了更多关注。目前的一个方向是寻找和识别系统性弱点,这些弱点会使基于平均性能值的安全假设面临风险。这种弱点的形式可以是输入空间中(语义一致的)子集或区域,在这些子集或区域中,DNN 的性能系统性地低于其预期平均值。然而,要将观察到的这种低性能的原因归结于描述该子集的特定语义特征,并非易事。例如,数据中其他(未考虑的)属性的不均匀性可能会扭曲结果。然而,考虑所有(可用的)属性及其交互通常在计算上非常昂贵。在反事实解释的启发下,我们提出了一种有效且计算成本低廉的算法来验证现有子集的语义归属,即检查已识别的属性是否可能导致性能下降。我们利用高度注释的模拟数据,在自动驾驶领域的一个例子中演示了这种方法,我们通过语义分割模型表明:(i) 不同行人资产之间存在性能差异,但 (ii) 只有在某些情况下,资产类型本身才是性能下降的原因。
{"title":"Assessing systematic weaknesses of DNNs using counterfactuals","authors":"Sujan Sai Gannamaneni,&nbsp;Michael Mock,&nbsp;Maram Akila","doi":"10.1007/s43681-023-00407-0","DOIUrl":"10.1007/s43681-023-00407-0","url":null,"abstract":"<div><p>With the advancement of DNNs into safety-critical applications, testing approaches for such models have gained more attention. A current direction is the search for and identification of systematic weaknesses that put safety assumptions based on average performance values at risk. Such weaknesses can take on the form of (semantically coherent) subsets or areas in the input space where a DNN performs systematically worse than its expected average. However, it is non-trivial to attribute the reason for such observed low performances to the specific semantic features that describe the subset. For instance, inhomogeneities within the data w.r.t. other (non-considered) attributes might distort results. However, taking into account all (available) attributes and their interaction is often computationally highly expensive. Inspired by counterfactual explanations, we propose an effective and computationally cheap algorithm to validate the semantic attribution of existing subsets, i.e., to check whether the identified attribute is likely to have caused the degraded performance. We demonstrate this approach on an example from the autonomous driving domain using highly annotated simulated data, where we show for a semantic segmentation model that (i) performance differences among the different pedestrian assets exist, but (ii) only in some cases is the asset type itself the reason for this reduction in the performance.</p></div>","PeriodicalId":72137,"journal":{"name":"AI and ethics","volume":"4 1","pages":"27 - 35"},"PeriodicalIF":0.0,"publicationDate":"2024-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43681-023-00407-0.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142410545","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
QI(^2): an interactive tool for data quality assurance QI $$^2$ 2:数据质量保证互动工具
Pub Date : 2024-01-08 DOI: 10.1007/s43681-023-00390-6
Simon Geerkens, Christian Sieberichs, Alexander Braun, Thomas Waschulzik

The importance of high data quality is increasing with the growing impact and distribution of ML systems and big data. Also, the planned AI Act from the European commission defines challenging legal requirements for data quality especially for the market introduction of safety relevant ML systems. In this paper, we introduce a novel approach that supports the data quality assurance process of multiple data quality aspects. This approach enables the verification of quantitative data quality requirements. The concept and benefits are introduced and explained on small example data sets. How the method is applied is demonstrated on the well-known MNIST data set based an handwritten digits.

随着人工智能系统和大数据的影响和分布日益扩大,高数据质量的重要性也与日俱增。此外,欧盟委员会计划出台的《人工智能法案》对数据质量提出了具有挑战性的法律要求,特别是对安全相关的 ML 系统的市场引入。在本文中,我们介绍了一种新颖的方法,可支持多个数据质量方面的数据质量保证流程。这种方法可以验证定量数据质量要求。本文通过小型示例数据集介绍并解释了该方法的概念和优点。在著名的基于手写数字的 MNIST 数据集上演示了如何应用该方法。
{"title":"QI(^2): an interactive tool for data quality assurance","authors":"Simon Geerkens,&nbsp;Christian Sieberichs,&nbsp;Alexander Braun,&nbsp;Thomas Waschulzik","doi":"10.1007/s43681-023-00390-6","DOIUrl":"10.1007/s43681-023-00390-6","url":null,"abstract":"<div><p>The importance of high data quality is increasing with the growing impact and distribution of ML systems and big data. Also, the planned AI Act from the European commission defines challenging legal requirements for data quality especially for the market introduction of safety relevant ML systems. In this paper, we introduce a novel approach that supports the data quality assurance process of multiple data quality aspects. This approach enables the verification of quantitative data quality requirements. The concept and benefits are introduced and explained on small example data sets. How the method is applied is demonstrated on the well-known MNIST data set based an handwritten digits.</p></div>","PeriodicalId":72137,"journal":{"name":"AI and ethics","volume":"4 1","pages":"141 - 149"},"PeriodicalIF":0.0,"publicationDate":"2024-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43681-023-00390-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139446740","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
AI and ethics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1