首页 > 最新文献

Applied AI letters最新文献

英文 中文
Reframing explanation as an interactive medium: The EQUAS (Explainable QUestion Answering System) project 将解释作为一种互动媒介:EQUAS(可解释问答系统)项目
Pub Date : 2021-11-30 DOI: 10.1002/ail2.60
William Ferguson, Dhruv Batra, Raymond Mooney, Devi Parikh, Antonio Torralba, David Bau, David Diller, Josh Fasching, Jaden Fiotto-Kaufman, Yash Goyal, Jeff Miller, Kerry Moffitt, Alex Montes de Oca, Ramprasaath R. Selvaraju, Ayush Shrivastava, Jialin Wu, Stefan Lee

This letter is a retrospective analysis of our team's research for the Defense Advanced Research Projects Agency Explainable Artificial Intelligence project. Our initial approach was to use salience maps, English sentences, and lists of feature names to explain the behavior of deep-learning-based discriminative systems, with particular focus on visual question answering systems. We found that presenting static explanations along with answers led to limited positive effects. By exploring various combinations of machine and human explanation production and consumption, we evolved a notion of explanation as an interactive process that takes place usually between humans and artificial intelligence systems but sometimes within the software system. We realized that by interacting via explanations people could task and adapt machine learning (ML) agents. We added affordances for editing explanations and modified the ML system to act in accordance with the edits to produce an interpretable interface to the agent. Through this interface, editing an explanation can adapt a system's performance to new, modified purposes. This deep tasking, wherein the agent knows its objective and the explanation for that objective, will be critical to enable higher levels of autonomy.

这封信是对我们团队为国防高级研究计划局可解释人工智能项目所做研究的回顾性分析。我们最初的方法是使用显著性地图、英语句子和特征名称列表来解释基于深度学习的判别系统的行为,特别关注视觉问答系统。我们发现,在给出答案的同时给出静态的解释,其积极效果有限。通过探索机器和人类解释生产和消费的各种组合,我们进化出一种解释的概念,即解释是一种交互过程,通常发生在人类和人工智能系统之间,但有时也发生在软件系统内部。我们意识到,通过解释进行交互,人们可以分配任务并适应机器学习(ML)代理。我们添加了编辑解释的功能,并修改了机器学习系统,使其根据编辑内容采取行动,从而为代理生成可解释的界面。通过这个接口,编辑解释可以使系统的性能适应新的、修改过的目的。在这种深度任务中,智能体知道自己的目标和对目标的解释,这对于实现更高水平的自主性至关重要。
{"title":"Reframing explanation as an interactive medium: The EQUAS (Explainable QUestion Answering System) project","authors":"William Ferguson,&nbsp;Dhruv Batra,&nbsp;Raymond Mooney,&nbsp;Devi Parikh,&nbsp;Antonio Torralba,&nbsp;David Bau,&nbsp;David Diller,&nbsp;Josh Fasching,&nbsp;Jaden Fiotto-Kaufman,&nbsp;Yash Goyal,&nbsp;Jeff Miller,&nbsp;Kerry Moffitt,&nbsp;Alex Montes de Oca,&nbsp;Ramprasaath R. Selvaraju,&nbsp;Ayush Shrivastava,&nbsp;Jialin Wu,&nbsp;Stefan Lee","doi":"10.1002/ail2.60","DOIUrl":"10.1002/ail2.60","url":null,"abstract":"<p>This letter is a retrospective analysis of our team's research for the Defense Advanced Research Projects Agency Explainable Artificial Intelligence project. Our initial approach was to use salience maps, English sentences, and lists of feature names to explain the behavior of deep-learning-based discriminative systems, with particular focus on visual question answering systems. We found that presenting static explanations along with answers led to limited positive effects. By exploring various combinations of machine and human explanation production and consumption, we evolved a notion of explanation as an interactive process that takes place usually between humans and artificial intelligence systems but sometimes within the software system. We realized that by interacting via explanations people could task and adapt machine learning (ML) agents. We added affordances for editing explanations and modified the ML system to act in accordance with the edits to produce an interpretable interface to the agent. Through this interface, editing an explanation can adapt a system's performance to new, modified purposes. This deep tasking, wherein the agent knows its objective and the explanation for that objective, will be critical to enable higher levels of autonomy.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.60","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41941827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards structured NLP interpretation via graph explainers 通过图形解释器实现结构化的NLP解释
Pub Date : 2021-11-26 DOI: 10.1002/ail2.58
Hao Yuan, Fan Yang, Mengnan Du, Shuiwang Ji, Xia Hu

Natural language processing (NLP) models have been increasingly deployed in real-world applications, and interpretation for textual data has also attracted dramatic attention recently. Most existing methods generate feature importance interpretation, which indicate the contribution of each word towards a specific model prediction. Text data typically possess highly structured characteristics and feature importance explanation cannot fully reveal the rich information contained in text. To bridge this gap, we propose to generate structured interpretations for textual data. Specifically, we pre-process the original text using dependency parsing, which could transform the text from sequences into graphs. Then graph neural networks (GNNs) are utilized to classify the transformed graphs. In particular, we explore two kinds of structured interpretation for pre-trained GNNs: edge-level interpretation and subgraph-level interpretation. Experimental results over three text datasets demonstrate that the structured interpretation can better reveal the structured knowledge encoded in the text. The experimental analysis further indicates that the proposed interpretations can faithfully reflect the decision-making process of the GNN model.

自然语言处理(NLP)模型越来越多地应用于现实世界,文本数据的解释也引起了人们的极大关注。大多数现有方法生成特征重要性解释,表明每个词对特定模型预测的贡献。文本数据通常具有高度结构化的特征,特征重要性的解释并不能充分揭示文本所包含的丰富信息。为了弥补这一差距,我们建议为文本数据生成结构化的解释。具体来说,我们使用依赖解析对原始文本进行预处理,这可以将文本从序列转换为图。然后利用图神经网络(gnn)对变换后的图进行分类。特别地,我们探索了预训练gnn的两种结构化解释:边缘级解释和子图级解释。在三个文本数据集上的实验结果表明,结构化解释可以更好地揭示文本中编码的结构化知识。实验分析进一步表明,所提出的解释能够真实地反映GNN模型的决策过程。
{"title":"Towards structured NLP interpretation via graph explainers","authors":"Hao Yuan,&nbsp;Fan Yang,&nbsp;Mengnan Du,&nbsp;Shuiwang Ji,&nbsp;Xia Hu","doi":"10.1002/ail2.58","DOIUrl":"10.1002/ail2.58","url":null,"abstract":"<p>Natural language processing (NLP) models have been increasingly deployed in real-world applications, and interpretation for textual data has also attracted dramatic attention recently. Most existing methods generate feature importance interpretation, which indicate the contribution of each word towards a specific model prediction. Text data typically possess highly structured characteristics and feature importance explanation cannot fully reveal the rich information contained in text. To bridge this gap, we propose to generate structured interpretations for textual data. Specifically, we pre-process the original text using dependency parsing, which could transform the text from sequences into graphs. Then graph neural networks (GNNs) are utilized to classify the transformed graphs. In particular, we explore two kinds of structured interpretation for pre-trained GNNs: edge-level interpretation and subgraph-level interpretation. Experimental results over three text datasets demonstrate that the structured interpretation can better reveal the structured knowledge encoded in the text. The experimental analysis further indicates that the proposed interpretations can faithfully reflect the decision-making process of the GNN model.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.58","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43677736","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Explainable activity recognition in videos: Lessons learned 视频中可解释的活动识别:经验教训
Pub Date : 2021-11-26 DOI: 10.1002/ail2.59
Chiradeep Roy, Mahsan Nourani, Donald R. Honeycutt, Jeremy E. Block, Tahrima Rahman, Eric D. Ragan, Nicholas Ruozzi, Vibhav Gogate

We consider the following activity recognition task: given a video, infer the set of activities being performed in the video and assign each frame to an activity. This task can be solved using modern deep learning architectures based on neural networks or conventional classifiers such as linear models and decision trees. While neural networks exhibit superior predictive performance as compared with decision trees and linear models, they are also uninterpretable and less explainable. We address this accuracy-explanability gap using a novel framework that feeds the output of a deep neural network to an interpretable, tractable probabilistic model called dynamic cutset networks, and performs joint reasoning over the two to answer questions. The neural network helps achieve high accuracy while dynamic cutset networks because of their polytime probabilistic reasoning capabilities make the system more explainable. We demonstrate the efficacy of our approach by using it to build three prototype systems that solve human-machine tasks having varying levels of difficulty using cooking videos as an accessible domain. We describe high-level technical details and key lessons learned in our human subjects evaluations of these systems.

我们考虑以下活动识别任务:给定一个视频,推断视频中正在执行的活动集,并将每一帧分配给一个活动。这个任务可以使用基于神经网络或传统分类器(如线性模型和决策树)的现代深度学习架构来解决。虽然与决策树和线性模型相比,神经网络表现出优越的预测性能,但它们也是不可解释和不可解释的。我们使用一种新的框架来解决这种准确性和可解释性之间的差距,该框架将深度神经网络的输出提供给一个可解释的、可处理的概率模型,称为动态割集网络,并对两者进行联合推理以回答问题。神经网络有助于实现高精度,而动态割集网络由于其多时概率推理能力使系统更具可解释性。我们通过使用它来构建三个原型系统来证明我们的方法的有效性,这些系统可以解决具有不同难度的人机任务,并将烹饪视频作为可访问域。我们描述了在这些系统的人类受试者评估中获得的高级技术细节和关键经验教训。
{"title":"Explainable activity recognition in videos: Lessons learned","authors":"Chiradeep Roy,&nbsp;Mahsan Nourani,&nbsp;Donald R. Honeycutt,&nbsp;Jeremy E. Block,&nbsp;Tahrima Rahman,&nbsp;Eric D. Ragan,&nbsp;Nicholas Ruozzi,&nbsp;Vibhav Gogate","doi":"10.1002/ail2.59","DOIUrl":"10.1002/ail2.59","url":null,"abstract":"<p>We consider the following activity recognition task: given a video, infer the set of activities being performed in the video and assign each frame to an activity. This task can be solved using modern deep learning architectures based on neural networks or conventional classifiers such as linear models and decision trees. While neural networks exhibit superior predictive performance as compared with decision trees and linear models, they are also uninterpretable and less explainable. We address this <i>accuracy-explanability gap</i> using a novel framework that feeds the output of a deep neural network to an interpretable, tractable probabilistic model called dynamic cutset networks, and performs joint reasoning over the two to answer questions. The neural network helps achieve high accuracy while dynamic cutset networks because of their polytime probabilistic reasoning capabilities make the system more explainable. We demonstrate the efficacy of our approach by using it to build three prototype systems that solve human-machine tasks having varying levels of difficulty using cooking videos as an accessible domain. We describe high-level technical details and key lessons learned in our human subjects evaluations of these systems.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.59","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43572819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Toward explainable and advisable model for self-driving cars 为自动驾驶汽车建立一个可解释且可行的模型
Pub Date : 2021-11-23 DOI: 10.1002/ail2.56
Jinkyu Kim, Anna Rohrbach, Zeynep Akata, Suhong Moon, Teruhisa Misu, Yi-Ting Chen, Trevor Darrell, John Canny

Humans learn to drive through both practice and theory, for example, by studying the rules, while most self-driving systems are limited to the former. Being able to incorporate human knowledge of typical causal driving behavior should benefit autonomous systems. We propose a new approach that learns vehicle control with the help of human advice. Specifically, our system learns to summarize its visual observations in natural language, predict an appropriate action response (eg, “I see a pedestrian crossing, so I stop”), and predict the controls, accordingly. Moreover, to enhance the interpretability of our system, we introduce a fine-grained attention mechanism that relies on semantic segmentation and object-centric RoI pooling. We show that our approach of training the autonomous system with human advice, grounded in a rich semantic representation, matches or outperforms prior work in terms of control prediction and explanation generation. Our approach also results in more interpretable visual explanations by visualizing object-centric attention maps. We evaluate our approach on a novel driving dataset with ground-truth human explanations, the Berkeley DeepDrive eXplanation (BDD-X) dataset.

人类通过实践和理论来学习驾驶,例如,通过研究规则,而大多数自动驾驶系统仅限于前者。能够将人类对典型因果驾驶行为的知识结合起来,应该有利于自动驾驶系统。我们提出了一种新的方法,在人类建议的帮助下学习车辆控制。具体来说,我们的系统学会了用自然语言总结它的视觉观察,预测适当的行动反应(例如,“我看到一个人行横道,所以我停下来”),并相应地预测控制。此外,为了增强系统的可解释性,我们引入了一种依赖于语义分割和以对象为中心的RoI池的细粒度注意力机制。我们表明,基于丰富的语义表示,我们用人类建议训练自主系统的方法,在控制预测和解释生成方面匹配或优于先前的工作。通过可视化以对象为中心的注意图,我们的方法也产生了更多可解释的视觉解释。我们在一个新的驾驶数据集上评估了我们的方法,该数据集具有真实的人类解释,即伯克利深度驾驶解释(BDD-X)数据集。
{"title":"Toward explainable and advisable model for self-driving cars","authors":"Jinkyu Kim,&nbsp;Anna Rohrbach,&nbsp;Zeynep Akata,&nbsp;Suhong Moon,&nbsp;Teruhisa Misu,&nbsp;Yi-Ting Chen,&nbsp;Trevor Darrell,&nbsp;John Canny","doi":"10.1002/ail2.56","DOIUrl":"10.1002/ail2.56","url":null,"abstract":"<p>Humans learn to drive through both practice and theory, for example, by studying the rules, while most self-driving systems are limited to the former. Being able to incorporate human knowledge of typical causal driving behavior should benefit autonomous systems. We propose a new approach that learns vehicle control with the help of human advice. Specifically, our system learns to summarize its visual observations in natural language, predict an appropriate action response (eg, “I see a pedestrian crossing, so I stop”), and predict the controls, accordingly. Moreover, to enhance the interpretability of our system, we introduce a fine-grained attention mechanism that relies on semantic segmentation and object-centric RoI pooling. We show that our approach of training the autonomous system with human advice, grounded in a rich semantic representation, matches or outperforms prior work in terms of control prediction and explanation generation. Our approach also results in more interpretable visual explanations by visualizing object-centric attention maps. We evaluate our approach on a novel driving dataset with ground-truth human explanations, the Berkeley DeepDrive eXplanation (BDD-X) dataset.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.56","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46237887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Objective criteria for explanations of machine learning models 解释机器学习模型的客观标准
Pub Date : 2021-11-23 DOI: 10.1002/ail2.57
Chih-Kuan Yeh, Pradeep Ravikumar

Objective criteria to evaluate the performance of machine learning (ML) model explanations are a critical ingredient in bringing greater rigor to the field of explainable artificial intelligence. In this article, we survey three of our proposed criteria that each target different classes of explanations. In the first, targeted at real-valued feature importance explanations, we define a class of “infidelity” measures that capture how well the explanations match the ML models. We show that instances of such infidelity minimizing explanations correspond to many popular recently proposed explanations and, moreover, can be shown to satisfy well-known game-theoretic axiomatic properties. In the second, targeted to feature set explanations, we define a robustness analysis-based criterion and show that deriving explainable feature sets based on the robustness criterion yields more qualitatively impressive explanations. Lastly, for sample explanations, we provide a decomposition-based criterion that allows us to provide very scalable and compelling classes of sample-based explanations.

评估机器学习(ML)模型解释性能的客观标准是提高可解释人工智能领域的严谨性的关键因素。在本文中,我们调查了我们提出的三个标准,每个标准针对不同类别的解释。首先,针对实值特征重要性解释,我们定义了一类“不忠”度量,以捕获解释与ML模型的匹配程度。我们证明了这种不忠最小化解释的实例与最近提出的许多流行解释相对应,而且可以证明满足众所周知的博弈论公理性质。在第二部分中,针对特征集解释,我们定义了一个基于鲁棒性分析的标准,并表明基于鲁棒性标准推导可解释的特征集产生了更多的定性令人印象深刻的解释。最后,对于样本解释,我们提供了一个基于分解的标准,它允许我们提供非常可扩展和引人注目的基于样本的解释类。
{"title":"Objective criteria for explanations of machine learning models","authors":"Chih-Kuan Yeh,&nbsp;Pradeep Ravikumar","doi":"10.1002/ail2.57","DOIUrl":"10.1002/ail2.57","url":null,"abstract":"<p>Objective criteria to evaluate the performance of machine learning (ML) model explanations are a critical ingredient in bringing greater rigor to the field of explainable artificial intelligence. In this article, we survey three of our proposed criteria that each target different classes of explanations. In the first, targeted at real-valued feature importance explanations, we define a class of “infidelity” measures that capture how well the explanations match the ML models. We show that instances of such infidelity minimizing explanations correspond to many popular recently proposed explanations and, moreover, can be shown to satisfy well-known game-theoretic axiomatic properties. In the second, targeted to feature set explanations, we define a robustness analysis-based criterion and show that deriving explainable feature sets based on the robustness criterion yields more qualitatively impressive explanations. Lastly, for sample explanations, we provide a decomposition-based criterion that allows us to provide very scalable and compelling classes of sample-based explanations.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.57","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44018783","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Generating visual explanations with natural language 用自然语言生成视觉解释
Pub Date : 2021-11-22 DOI: 10.1002/ail2.55
Lisa Anne Hendricks, Anna Rohrbach, Bernt Schiele, Trevor Darrell, Zeynep Akata

We generate natural language explanations for a fine-grained visual recognition task. Our explanations fulfill two criteria. First, explanations are class discriminative, meaning they mention attributes in an image which are important to identify a class. Second, explanations are image relevant, meaning they reflect the actual content of an image. Our system, composed of an explanation sampler and phrase-critic model, generates class discriminative and image relevant explanations. In addition, we demonstrate that our explanations can help humans decide whether to accept or reject an AI decision.

我们为细粒度的视觉识别任务生成自然语言解释。我们的解释符合两个标准。首先,解释是类区分的,这意味着它们提到图像中的属性,这些属性对识别类很重要。其次,解释是与图像相关的,这意味着它们反映了图像的实际内容。我们的系统由解释采样器和短语批评模型组成,生成阶级区分和图像相关的解释。此外,我们证明了我们的解释可以帮助人类决定是否接受或拒绝人工智能的决定。
{"title":"Generating visual explanations with natural language","authors":"Lisa Anne Hendricks,&nbsp;Anna Rohrbach,&nbsp;Bernt Schiele,&nbsp;Trevor Darrell,&nbsp;Zeynep Akata","doi":"10.1002/ail2.55","DOIUrl":"10.1002/ail2.55","url":null,"abstract":"<p>We generate natural language explanations for a fine-grained visual recognition task. Our explanations fulfill two criteria. First, explanations are <i>class discriminative</i>, meaning they mention attributes in an image which are important to identify a class. Second, explanations are <i>image relevant</i>, meaning they reflect the actual content of an image. Our system, composed of an explanation sampler and phrase-critic model, generates class discriminative and image relevant explanations. In addition, we demonstrate that our explanations can help humans decide whether to accept or reject an AI decision.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.55","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48200035","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Explaining autonomous drones: An XAI journey 解释自主无人机:XAI之旅
Pub Date : 2021-11-22 DOI: 10.1002/ail2.54
Mark Stefik, Michael Youngblood, Peter Pirolli, Christian Lebiere, Robert Thomson, Robert Price, Lester D. Nelson, Robert Krivacic, Jacob Le, Konstantinos Mitsopoulos, Sterling Somers, Joel Schooler

COGLE (COmmon Ground Learning and Explanation) is an explainable artificial intelligence (XAI) system where autonomous drones deliver supplies to field units in mountainous areas. The mission risks vary with topography, flight decisions, and mission goals. The missions engage a human plus AI team where users determine which of two AI-controlled drones is better for each mission. This article reports on the technical approach and findings of the project and reflects on challenges that complex combinatorial problems present for users, machine learning, user studies, and the context of use for XAI systems. COGLE creates explanations in multiple modalities. Narrative “What” explanations compare what each drone does on a mission and “Why” based on drone competencies determined from experiments using counterfactuals. Visual “Where” explanations highlight risks on maps to help users to interpret flight plans. One branch of the research studied whether the explanations helped users to predict drone performance. In this branch, a model induction user study showed that post-decision explanations had only a small effect in teaching users to determine by themselves which drone is better for a mission. Subsequent reflection suggests that supporting human plus AI decision making with pre-decision explanations is a better context for benefiting from explanations on combinatorial tasks.

COGLE (COmmon Ground Learning and explain)是一种可解释的人工智能(XAI)系统,由无人驾驶飞机向山区野战部队运送物资。任务风险随地形、飞行决策和任务目标而变化。这些任务由人类和人工智能团队组成,用户可以决定两架人工智能控制的无人机中哪一架更适合每个任务。本文报告了该项目的技术方法和发现,并反映了复杂组合问题对用户、机器学习、用户研究和XAI系统使用环境的挑战。COGLE以多种方式创建解释。叙述性的“什么”解释比较了每架无人机在执行任务时所做的事情,以及基于反事实实验确定的无人机能力的“为什么”。可视化的“地点”解释在地图上突出了风险,以帮助用户理解飞行计划。该研究的一个分支是研究这些解释是否有助于用户预测无人机的性能。在这个分支中,一项模型归纳用户研究表明,决策后解释在教导用户自己确定哪种无人机更适合执行任务方面只有很小的作用。随后的反思表明,通过决策前解释来支持人类和人工智能的决策是一个更好的环境,可以从组合任务的解释中受益。
{"title":"Explaining autonomous drones: An XAI journey","authors":"Mark Stefik,&nbsp;Michael Youngblood,&nbsp;Peter Pirolli,&nbsp;Christian Lebiere,&nbsp;Robert Thomson,&nbsp;Robert Price,&nbsp;Lester D. Nelson,&nbsp;Robert Krivacic,&nbsp;Jacob Le,&nbsp;Konstantinos Mitsopoulos,&nbsp;Sterling Somers,&nbsp;Joel Schooler","doi":"10.1002/ail2.54","DOIUrl":"10.1002/ail2.54","url":null,"abstract":"<p>COGLE (<i>CO</i>mmon <i>G</i>round <i>L</i>earning and <i>E</i>xplanation) is an explainable artificial intelligence (XAI) system where autonomous drones deliver supplies to field units in mountainous areas. The mission risks vary with topography, flight decisions, and mission goals. The missions engage a human plus AI team where users determine which of two AI-controlled drones is better for each mission. This article reports on the technical approach and findings of the project and reflects on challenges that complex combinatorial problems present for users, machine learning, user studies, and the context of use for XAI systems. COGLE creates explanations in multiple modalities. Narrative “What” explanations compare what each drone does on a mission and “Why” based on drone competencies determined from experiments using counterfactuals. Visual “Where” explanations highlight risks on maps to help users to interpret flight plans. One branch of the research studied whether the explanations helped users to predict drone performance. In this branch, a model induction user study showed that <i>post-decision explanations</i> had only a small effect in teaching users to determine by themselves which drone is better for a mission. Subsequent reflection suggests that supporting human plus AI decision making with <i>pre-decision explanations</i> is a better context for benefiting from explanations on combinatorial tasks.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.54","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47270045","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Explaining robot policies 解释机器人政策
Pub Date : 2021-11-13 DOI: 10.1002/ail2.52
Olivia Watkins, Sandy Huang, Julius Frost, Kush Bhatia, Eric Weiner, Pieter Abbeel, Trevor Darrell, Bryan Plummer, Kate Saenko, Anca Dragan

In order to interact with a robot or make wise decisions about where and how to deploy it in the real world, humans need to have an accurate mental model of how the robot acts in different situations. We propose to improve users' mental model of a robot by showing them examples of how the robot behaves in informative scenarios. We explore this in two settings. First, we show that when there are many possible environment states, users can more quickly understand the robot's policy if they are shown critical states where taking a particular action is important. Second, we show that when there is a distribution shift between training and test environment distributions, then it is more effective to show exploratory states that the robot does not visit naturally.

为了与机器人互动或做出明智的决定,决定在现实世界中部署机器人的位置和方式,人类需要对机器人在不同情况下的行为有一个准确的心理模型。我们建议通过向用户展示机器人在信息场景中的行为来改善用户对机器人的心理模型。我们在两种情况下对此进行探讨。首先,我们表明,当有许多可能的环境状态时,如果用户看到采取特定行动很重要的关键状态,他们可以更快地理解机器人的策略。其次,我们证明了当训练环境和测试环境之间的分布发生变化时,显示机器人不自然访问的探索状态更有效。
{"title":"Explaining robot policies","authors":"Olivia Watkins,&nbsp;Sandy Huang,&nbsp;Julius Frost,&nbsp;Kush Bhatia,&nbsp;Eric Weiner,&nbsp;Pieter Abbeel,&nbsp;Trevor Darrell,&nbsp;Bryan Plummer,&nbsp;Kate Saenko,&nbsp;Anca Dragan","doi":"10.1002/ail2.52","DOIUrl":"10.1002/ail2.52","url":null,"abstract":"<p>In order to interact with a robot or make wise decisions about where and how to deploy it in the real world, humans need to have an accurate mental model of how the robot acts in different situations. We propose to improve users' mental model of a robot by showing them examples of how the robot behaves in informative scenarios. We explore this in two settings. First, we show that when there are many possible environment states, users can more quickly understand the robot's policy if they are shown <i>critical states</i> where taking a particular action is important. Second, we show that when there is a distribution shift between training and test environment distributions, then it is more effective to show <i>exploratory states</i> that the robot does not visit naturally.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.52","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41445238","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Methods and standards for research on explainable artificial intelligence: Lessons from intelligent tutoring systems 可解释人工智能的研究方法与标准:来自智能辅导系统的经验教训
Pub Date : 2021-11-13 DOI: 10.1002/ail2.53
William J. Clancey, Robert R. Hoffman

The DARPA Explainable Artificial Intelligence (AI) (XAI) Program focused on generating explanations for AI programs that use machine learning techniques. This article highlights progress during the DARPA Program (2017-2021) relative to research since the 1970s in the field of intelligent tutoring systems (ITSs). ITS researchers learned a great deal about explanation that is directly relevant to XAI. We suggest opportunities for future XAI research deriving from ITS methods, and consider the challenges shared by both ITS and XAI in using AI to assist people in solving difficult problems effectively and efficiently.

DARPA可解释人工智能(AI) (XAI)项目专注于为使用机器学习技术的人工智能程序生成解释。本文重点介绍了DARPA计划(2017-2021)期间相对于20世纪70年代以来智能辅导系统(ITSs)领域的研究进展。ITS研究人员学到了很多与XAI直接相关的解释。我们提出了未来人工智能研究从ITS方法衍生出来的机会,并考虑了ITS和人工智能在使用人工智能帮助人们有效和高效地解决难题方面所面临的共同挑战。
{"title":"Methods and standards for research on explainable artificial intelligence: Lessons from intelligent tutoring systems","authors":"William J. Clancey,&nbsp;Robert R. Hoffman","doi":"10.1002/ail2.53","DOIUrl":"https://doi.org/10.1002/ail2.53","url":null,"abstract":"<p>The DARPA Explainable Artificial Intelligence (AI) (XAI) Program focused on generating explanations for AI programs that use machine learning techniques. This article highlights progress during the DARPA Program (2017-2021) relative to research since the 1970s in the field of intelligent tutoring systems (ITSs). ITS researchers learned a great deal about explanation that is directly relevant to XAI. We suggest opportunities for future XAI research deriving from ITS methods, and consider the challenges shared by both ITS and XAI in using AI to assist people in solving difficult problems effectively and efficiently.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.53","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"137515652","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generating and evaluating explanations of attended and error-inducing input regions for VQA models 生成和评估VQA模型的参与和错误诱导输入区域的解释
Pub Date : 2021-11-12 DOI: 10.1002/ail2.51
Arijit Ray, Michael Cogswell, Xiao Lin, Kamran Alipour, Ajay Divakaran, Yi Yao, Giedrius Burachas

Attention maps, a popular heatmap-based explanation method for Visual Question Answering, are supposed to help users understand the model by highlighting portions of the image/question used by the model to infer answers. However, we see that users are often misled by current attention map visualizations that point to relevant regions despite the model producing an incorrect answer. Hence, we propose Error Maps that clarify the error by highlighting image regions where the model is prone to err. Error maps can indicate when a correctly attended region may be processed incorrectly leading to an incorrect answer, and hence, improve users' understanding of those cases. To evaluate our new explanations, we further introduce a metric that simulates users' interpretation of explanations to evaluate their potential helpfulness to understand model correctness. We finally conduct user studies to see that our new explanations help users understand model correctness better than baselines by an expected 30% and that our proxy helpfulness metrics correlate strongly (ρ>0.97) with how well users can predict model correctness.

注意图是一种流行的基于热图的可视化问答解释方法,它通过突出显示模型用来推断答案的图像/问题的部分来帮助用户理解模型。然而,我们发现用户经常被当前的注意力地图可视化所误导,尽管模型产生了错误的答案,但它们指向了相关的区域。因此,我们提出了错误地图,通过突出显示模型容易出错的图像区域来澄清错误。错误映射可以指示正确参与的区域何时可能被错误地处理,从而导致错误的答案,从而提高用户对这些情况的理解。为了评估我们的新解释,我们进一步引入了一个度量,该度量模拟用户对解释的解释,以评估它们对理解模型正确性的潜在帮助。我们最终进行了用户研究,发现我们的新解释帮助用户比基线更好地理解模型正确性,预期高出30%,并且我们的代理帮助度量具有很强的相关性(ρ >0.97)与用户预测模型正确性的程度有关。
{"title":"Generating and evaluating explanations of attended and error-inducing input regions for VQA models","authors":"Arijit Ray,&nbsp;Michael Cogswell,&nbsp;Xiao Lin,&nbsp;Kamran Alipour,&nbsp;Ajay Divakaran,&nbsp;Yi Yao,&nbsp;Giedrius Burachas","doi":"10.1002/ail2.51","DOIUrl":"https://doi.org/10.1002/ail2.51","url":null,"abstract":"<p>Attention maps, a popular heatmap-based explanation method for Visual Question Answering, are supposed to help users understand the model by highlighting portions of the image/question used by the model to infer answers. However, we see that users are often misled by current attention map visualizations that point to relevant regions despite the model producing an incorrect answer. Hence, we propose Error Maps that clarify the error by highlighting image regions where the model is prone to err. Error maps can indicate when a correctly attended region may be processed incorrectly leading to an incorrect answer, and hence, improve users' understanding of those cases. To evaluate our new explanations, we further introduce a metric that simulates users' interpretation of explanations to evaluate their potential helpfulness to understand model correctness. We finally conduct user studies to see that our new explanations help users understand model correctness better than baselines by an expected 30% and that our proxy helpfulness metrics correlate strongly (<math>\u0000 <mi>ρ</mi>\u0000 <mo>&gt;</mo>\u0000 <mn>0.97</mn></math>) with how well users can predict model correctness.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.51","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"137830967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Applied AI letters
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1