首页 > 最新文献

Applied AI letters最新文献

英文 中文
Computer Vision and Machine Learning Techniques for Quantification and Predictive Modeling of Intracellular Anti‐Cancer Drug Delivery by Nanocarriers 计算机视觉和机器学习技术用于纳米载体细胞内抗癌症药物递送的定量和预测建模
Pub Date : 2021-11-10 DOI: 10.1002/ail2.50
S. Goswami, Kshama D. Dhobale, R. Wavhale, B. Goswami, S. Banerjee
{"title":"Computer Vision and Machine Learning Techniques for Quantification and Predictive Modeling of Intracellular\u0000 Anti‐Cancer\u0000 Drug Delivery by Nanocarriers","authors":"S. Goswami, Kshama D. Dhobale, R. Wavhale, B. Goswami, S. Banerjee","doi":"10.1002/ail2.50","DOIUrl":"https://doi.org/10.1002/ail2.50","url":null,"abstract":"","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45346424","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
How level of explanation detail affects human performance in interpretable intelligent systems: A study on explainable fact checking 在可解释的智能系统中,解释细节的水平如何影响人类的表现:一项关于可解释事实核查的研究
Pub Date : 2021-11-08 DOI: 10.1002/ail2.49
Rhema Linder, Sina Mohseni, Fan Yang, Shiva K. Pentyala, Eric D. Ragan, Xia Ben Hu

Explainable artificial intelligence (XAI) systems aim to provide users with information to help them better understand computational models and reason about why outputs were generated. However, there are many different ways an XAI interface might present explanations, which makes designing an appropriate and effective interface an important and challenging task. Our work investigates how different types and amounts of explanatory information affect user ability to utilize explanations to understand system behavior and improve task performance. The presented research employs a system for detecting the truthfulness of news statements. In a controlled experiment, participants were tasked with using the system to assess news statements as well as to learn to predict the output of the AI. Our experiment compares various levels of explanatory information to contribute empirical data about how explanation detail can influence utility. The results show that more explanation information improves participant understanding of AI models, but the benefits come at the cost of time and attention needed to make sense of the explanation.

可解释的人工智能(XAI)系统旨在为用户提供信息,帮助他们更好地理解计算模型和产生输出的原因。然而,XAI界面可能有许多不同的解释方式,这使得设计一个适当而有效的界面成为一项重要而具有挑战性的任务。我们的工作调查了不同类型和数量的解释信息如何影响用户利用解释来理解系统行为和提高任务性能的能力。本研究采用了一种检测新闻陈述真实性的系统。在一项对照实验中,参与者的任务是使用该系统评估新闻声明,并学习预测人工智能的输出。我们的实验比较了不同层次的解释信息,以提供关于解释细节如何影响效用的经验数据。结果表明,更多的解释信息可以提高参与者对人工智能模型的理解,但这些好处是以理解解释所需的时间和注意力为代价的。
{"title":"How level of explanation detail affects human performance in interpretable intelligent systems: A study on explainable fact checking","authors":"Rhema Linder,&nbsp;Sina Mohseni,&nbsp;Fan Yang,&nbsp;Shiva K. Pentyala,&nbsp;Eric D. Ragan,&nbsp;Xia Ben Hu","doi":"10.1002/ail2.49","DOIUrl":"10.1002/ail2.49","url":null,"abstract":"<p>Explainable artificial intelligence (XAI) systems aim to provide users with information to help them better understand computational models and reason about why outputs were generated. However, there are many different ways an XAI interface might present explanations, which makes designing an appropriate and effective interface an important and challenging task. Our work investigates how different types and amounts of explanatory information affect user ability to utilize explanations to understand system behavior and improve task performance. The presented research employs a system for detecting the truthfulness of news statements. In a controlled experiment, participants were tasked with using the system to assess news statements as well as to learn to predict the output of the AI. Our experiment compares various levels of explanatory information to contribute empirical data about how explanation detail can influence utility. The results show that more explanation information improves participant understanding of AI models, but the benefits come at the cost of time and attention needed to make sense of the explanation.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.49","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46870560","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
From heatmaps to structured explanations of image classifiers 从热图到图像分类器的结构化解释
Pub Date : 2021-11-06 DOI: 10.1002/ail2.46
Li Fuxin, Zhongang Qi, Saeed Khorram, Vivswan Shitole, Prasad Tadepalli, Minsuk Kahng, Alan Fern

This paper summarizes our endeavors in the past few years in terms of explaining image classifiers, with the aim of including negative results and insights we have gained. The paper starts with describing the explainable neural network (XNN), which attempts to extract and visualize several high-level concepts purely from the deep network, without relying on human linguistic concepts. This helps users understand network classifications that are less intuitive and substantially improves user performance on a difficult fine-grained classification task of discriminating among different species of seagulls. Realizing that an important missing piece is a reliable heatmap visualization tool, we have developed integrated-gradient optimized saliency (I-GOS) and iGOS++ utilizing integrated gradients to avoid local optima in heatmap generation, which improved the performance across all resolutions. During the development of those visualizations, we realized that for a significant number of images, the classifier has multiple different paths to reach a confident prediction. This has led to our recent development of structured attention graphs, an approach that utilizes beam search to locate multiple coarse heatmaps for a single image, and compactly visualizes a set of heatmaps by capturing how different combinations of image regions impact the confidence of a classifier. Through the research process, we have learned much about insights in building deep network explanations, the existence and frequency of multiple explanations, and various tricks of the trade that make explanations work. In this paper, we attempt to share those insights and opinions with the readers with the hope that some of them will be informative for future researchers on explainable deep learning.

本文总结了我们过去几年在解释图像分类器方面的努力,目的是包括负面结果和我们获得的见解。本文首先描述了可解释神经网络(XNN),它试图纯粹从深度网络中提取和可视化几个高级概念,而不依赖于人类的语言概念。这有助于用户理解不太直观的网络分类,并大大提高用户在区分不同种类海鸥的困难细粒度分类任务上的性能。意识到一个重要的缺失部分是一个可靠的热图可视化工具,我们开发了集成梯度优化显着性(I-GOS)和igos++,利用集成梯度来避免热图生成中的局部最优,从而提高了所有分辨率下的性能。在这些可视化的开发过程中,我们意识到,对于大量的图像,分类器有多条不同的路径来达到一个自信的预测。这导致我们最近开发了结构化注意力图,这种方法利用光束搜索来定位单个图像的多个粗热图,并通过捕获图像区域的不同组合如何影响分类器的置信度来紧凑地可视化一组热图。通过研究过程,我们学到了很多关于构建深度网络解释的见解,多重解释的存在和频率,以及使解释起作用的各种交易技巧。在本文中,我们试图与读者分享这些见解和观点,希望其中一些能够为未来可解释深度学习的研究人员提供信息。
{"title":"From heatmaps to structured explanations of image classifiers","authors":"Li Fuxin,&nbsp;Zhongang Qi,&nbsp;Saeed Khorram,&nbsp;Vivswan Shitole,&nbsp;Prasad Tadepalli,&nbsp;Minsuk Kahng,&nbsp;Alan Fern","doi":"10.1002/ail2.46","DOIUrl":"10.1002/ail2.46","url":null,"abstract":"<p>This paper summarizes our endeavors in the past few years in terms of explaining image classifiers, with the aim of including negative results and insights we have gained. The paper starts with describing the explainable neural network (XNN), which attempts to extract and visualize several high-level concepts purely from the deep network, without relying on human linguistic concepts. This helps users understand network classifications that are less intuitive and substantially improves user performance on a difficult fine-grained classification task of discriminating among different species of seagulls. Realizing that an important missing piece is a reliable heatmap visualization tool, we have developed integrated-gradient optimized saliency (I-GOS) and iGOS++ utilizing integrated gradients to avoid local optima in heatmap generation, which improved the performance across all resolutions. During the development of those visualizations, we realized that for a significant number of images, the classifier has multiple different paths to reach a confident prediction. This has led to our recent development of structured attention graphs, an approach that utilizes beam search to locate multiple coarse heatmaps for a single image, and compactly visualizes a set of heatmaps by capturing how different combinations of image regions impact the confidence of a classifier. Through the research process, we have learned much about insights in building deep network explanations, the existence and frequency of multiple explanations, and various tricks of the trade that make explanations work. In this paper, we attempt to share those insights and opinions with the readers with the hope that some of them will be informative for future researchers on explainable deep learning.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.46","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73608694","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Improving users' mental model with attention-directed counterfactual edits 通过注意力导向的反事实编辑改善用户的心智模型
Pub Date : 2021-11-06 DOI: 10.1002/ail2.47
Kamran Alipour, Arijit Ray, Xiao Lin, Michael Cogswell, Jurgen P. Schulze, Yi Yao, Giedrius T. Burachas

In the domain of visual question answering (VQA), studies have shown improvement in users' mental model of the VQA system when they are exposed to examples of how these systems answer certain image-question (IQ) pairs. In this work, we show that showing controlled counterfactual IQ examples are more effective at improving the mental model of users as compared to simply showing random examples. We compare a generative approach and a retrieval-based approach to show counterfactual examples. We use recent advances in generative adversarial networks to generate counterfactual images by deleting and inpainting certain regions of interest in the image. We then expose users to changes in the VQA system's answer on those altered images. To select the region of interest for inpainting, we experiment with using both human-annotated attention maps and a fully automatic method that uses the VQA system's attention values. Finally, we test the user's mental model by asking them to predict the model's performance on a test counterfactual image. We note an overall improvement in users' accuracy to predict answer change when shown counterfactual explanations. While realistic retrieved counterfactuals obviously are the most effective at improving the mental model, we show that a generative approach can also be equally effective.

在视觉问答(VQA)领域,研究表明,当用户接触到这些系统如何回答某些图像问题(IQ)对的示例时,他们对VQA系统的心理模型有所改善。在这项工作中,我们表明,与简单地展示随机示例相比,展示受控的反事实智商示例在改善用户的心智模型方面更有效。我们比较了生成方法和基于检索的方法来展示反事实的例子。我们使用生成对抗网络的最新进展,通过删除和涂上图像中感兴趣的某些区域来生成反事实图像。然后,我们向用户展示VQA系统对这些改变后的图像的答案的变化。为了选择感兴趣的区域进行绘制,我们尝试使用人工注释的注意力图和使用VQA系统的注意力值的全自动方法。最后,我们通过要求用户预测模型在测试反事实图像上的表现来测试用户的心理模型。我们注意到,当显示反事实解释时,用户预测答案变化的准确性总体上有所提高。虽然现实检索的反事实显然在改进心智模型方面是最有效的,但我们表明生成方法也同样有效。
{"title":"Improving users' mental model with attention-directed counterfactual edits","authors":"Kamran Alipour,&nbsp;Arijit Ray,&nbsp;Xiao Lin,&nbsp;Michael Cogswell,&nbsp;Jurgen P. Schulze,&nbsp;Yi Yao,&nbsp;Giedrius T. Burachas","doi":"10.1002/ail2.47","DOIUrl":"https://doi.org/10.1002/ail2.47","url":null,"abstract":"<p>In the domain of visual question answering (VQA), studies have shown improvement in users' mental model of the VQA system when they are exposed to examples of how these systems answer certain image-question (IQ) pairs. In this work, we show that showing controlled counterfactual IQ examples are more effective at improving the mental model of users as compared to simply showing random examples. We compare a generative approach and a retrieval-based approach to show counterfactual examples. We use recent advances in generative adversarial networks to generate counterfactual images by deleting and inpainting certain regions of interest in the image. We then expose users to changes in the VQA system's answer on those altered images. To select the region of interest for inpainting, we experiment with using both human-annotated attention maps and a fully automatic method that uses the VQA system's attention values. Finally, we test the user's mental model by asking them to predict the model's performance on a test counterfactual image. We note an overall improvement in users' accuracy to predict answer change when shown counterfactual explanations. While realistic retrieved counterfactuals obviously are the most effective at improving the mental model, we show that a generative approach can also be equally effective.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.47","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"137648971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neural response time analysis: Explainable artificial intelligence using only a stopwatch 神经反应时间分析:可解释的人工智能只用一个秒表
Pub Date : 2021-11-06 DOI: 10.1002/ail2.48
J. Eric T. Taylor, Shashank Shekhar, Graham W. Taylor

How would you describe the features that a deep learning model composes if you were restricted to measuring observable behaviours? Explainable artificial intelligence (XAI) methods rely on privileged access to model architecture and parameters that is not always feasible for most users, practitioners and regulators. Inspired by cognitive psychology research on humans, we present a case for measuring response times (RTs) of a forward pass using only the system clock as a technique for XAI. Our method applies to the growing class of models that use input-adaptive dynamic inference and we also extend our approach to standard models that are converted to dynamic inference post hoc. The experimental logic is simple: If the researcher can contrive a stimulus set where variability among input features is tightly controlled, differences in RT for those inputs can be attributed to the way the model composes those features. First, we show that RT is sensitive to difficult, complex features by comparing RTs from ObjectNet and ImageNet. Next, we make specific a priori predictions about RT for abstract features present in the SCEGRAM data set, where object recognition in humans depends on complex intrascene object-object relationships. Finally, we show that RT profiles bear specificity for class identity and therefore the features that define classes. These results cast light on the model's feature space without opening the black box.

如果你被限制在测量可观察的行为,你会如何描述深度学习模型所构成的特征?可解释的人工智能(XAI)方法依赖于对模型架构和参数的特权访问,这对于大多数用户、从业者和监管者来说并不总是可行的。受人类认知心理学研究的启发,我们提出了一个仅使用系统时钟作为XAI技术来测量向前传递的响应时间(RTs)的案例。我们的方法适用于越来越多的使用输入自适应动态推理的模型,我们也将我们的方法扩展到转换为动态推理的标准模型。实验逻辑很简单:如果研究人员可以设计一个刺激集,其中输入特征之间的可变性受到严格控制,那么这些输入的RT差异可以归因于模型组成这些特征的方式。首先,我们通过比较ObjectNet和ImageNet的RT,证明RT对困难、复杂的特征很敏感。接下来,我们对sceggram数据集中存在的抽象特征的RT进行具体的先验预测,其中人类的对象识别依赖于复杂的内部对象-对象关系。最后,我们展示了RT概要文件具有类标识的特异性,因此具有定义类的特性。这些结果揭示了模型的特征空间,而无需打开黑盒。
{"title":"Neural response time analysis: Explainable artificial intelligence using only a stopwatch","authors":"J. Eric T. Taylor,&nbsp;Shashank Shekhar,&nbsp;Graham W. Taylor","doi":"10.1002/ail2.48","DOIUrl":"10.1002/ail2.48","url":null,"abstract":"<p>How would you describe the features that a deep learning model composes if you were restricted to measuring observable behaviours? Explainable artificial intelligence (XAI) methods rely on privileged access to model architecture and parameters that is not always feasible for most users, practitioners and regulators. Inspired by cognitive psychology research on humans, we present a case for measuring response times (RTs) of a forward pass using only the system clock as a technique for XAI. Our method applies to the growing class of models that use input-adaptive dynamic inference and we also extend our approach to standard models that are converted to dynamic inference post hoc. The experimental logic is simple: If the researcher can contrive a stimulus set where variability among input features is tightly controlled, differences in RT for those inputs can be attributed to the way the model composes those features. First, we show that RT is sensitive to difficult, complex features by comparing RTs from ObjectNet and ImageNet. Next, we make specific a priori predictions about RT for abstract features present in the SCEGRAM data set, where object recognition in humans depends on complex intrascene object-object relationships. Finally, we show that RT profiles bear specificity for class identity and therefore the features that define classes. These results cast light on the model's feature space without opening the black box.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.48","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47752665","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Measuring and characterizing generalization in deep reinforcement learning 深度强化学习中泛化的测量和表征
Pub Date : 2021-11-05 DOI: 10.1002/ail2.45
Sam Witty, Jun K. Lee, Emma Tosch, Akanksha Atrey, Kaleigh Clary, Michael L. Littman, David Jensen

Deep reinforcement learning (RL) methods have achieved remarkable performance on challenging control tasks. Observations of the resulting behavior give the impression that the agent has constructed a generalized representation that supports insightful action decisions. We re-examine what is meant by generalization in RL, and propose several definitions based on an agent's performance in on-policy, off-policy, and unreachable states. We propose a set of practical methods for evaluating agents with these definitions of generalization. We demonstrate these techniques on a common benchmark task for deep RL, and we show that the learned networks make poor decisions for states that differ only slightly from on-policy states, even though those states are not selected adversarially. We focus our analyses on the deep Q-networks (DQNs) that kicked off the modern era of deep RL. Taken together, these results call into question the extent to which DQNs learn generalized representations, and suggest that more experimentation and analysis is necessary before claims of representation learning can be supported.

深度强化学习(RL)方法在具有挑战性的控制任务上取得了显著的成绩。对结果行为的观察给人的印象是,代理已经构建了一个支持有洞察力的行动决策的广义表示。我们重新审视了强化学习中泛化的含义,并根据智能体在on-policy、off-policy和不可达状态下的表现提出了几个定义。我们提出了一套实用的方法来评估具有这些泛化定义的智能体。我们在深度强化学习的一个常见基准任务上展示了这些技术,并且我们表明,学习到的网络对与政策状态略有不同的状态做出了糟糕的决策,即使这些状态不是对抗性选择的。综上所述,这些结果对dqn学习广义表征的程度提出了质疑,并表明在支持表征学习的主张之前,需要进行更多的实验和分析。
{"title":"Measuring and characterizing generalization in deep reinforcement learning","authors":"Sam Witty,&nbsp;Jun K. Lee,&nbsp;Emma Tosch,&nbsp;Akanksha Atrey,&nbsp;Kaleigh Clary,&nbsp;Michael L. Littman,&nbsp;David Jensen","doi":"10.1002/ail2.45","DOIUrl":"10.1002/ail2.45","url":null,"abstract":"<p>Deep reinforcement learning (RL) methods have achieved remarkable performance on challenging control tasks. Observations of the resulting behavior give the impression that the agent has constructed a generalized representation that supports insightful action decisions. We re-examine what is meant by generalization in RL, and propose several definitions based on an agent's performance in on-policy, off-policy, and unreachable states. We propose a set of practical methods for evaluating agents with these definitions of generalization. We demonstrate these techniques on a common benchmark task for deep RL, and we show that the learned networks make poor decisions for states that differ only slightly from on-policy states, even though those states are not selected adversarially. We focus our analyses on the deep Q-networks (DQNs) that kicked off the modern era of deep RL. Taken together, these results call into question the extent to which DQNs learn generalized representations, and suggest that more experimentation and analysis is necessary before claims of representation learning can be supported.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.45","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75079317","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 46
Remembering for the right reasons: Explanations reduce catastrophic forgetting 记住正确的原因:解释可以减少灾难性的遗忘
Pub Date : 2021-11-05 DOI: 10.1002/ail2.44
Sayna Ebrahimi, Suzanne Petryk, Akash Gokul, William Gan, Joseph E. Gonzalez, Marcus Rohrbach, Trevor Darrell

The goal of continual learning (CL) is to learn a sequence of tasks without suffering from the phenomenon of catastrophic forgetting. Previous work has shown that leveraging memory in the form of a replay buffer can reduce performance degradation on prior tasks. We hypothesize that forgetting can be further reduced when the model is encouraged to remember the evidence for previously made decisions. As a first step towards exploring this hypothesis, we propose a simple novel training paradigm, called Remembering for the Right Reasons (RRR), that additionally stores visual model explanations for each example in the buffer and ensures the model has “the right reasons” for its predictions by encouraging its explanations to remain consistent with those used to make decisions at training time. Without this constraint, there is a drift in explanations and increase in forgetting as conventional continual learning algorithms learn new tasks. We demonstrate how RRR can be easily added to any memory or regularization-based approach and results in reduced forgetting, and more importantly, improved model explanations. We have evaluated our approach in the standard and few-shot settings and observed a consistent improvement across various CL approaches using different architectures and techniques to generate model explanations and demonstrated our approach showing a promising connection between explainability and continual learning. Our code is available at https://github.com/SaynaEbrahimi/Remembering-for-the-Right-Reasons.

持续学习(CL)的目标是在不遭受灾难性遗忘现象的情况下学习一系列任务。先前的工作表明,以重播缓冲区的形式利用内存可以减少先前任务的性能下降。我们假设,当模型被鼓励记住之前做出的决定的证据时,遗忘可以进一步减少。作为探索这一假设的第一步,我们提出了一个简单的新训练范式,称为“正确原因记忆”(RRR),它额外地在缓冲区中存储每个示例的视觉模型解释,并通过鼓励其解释与训练时用于决策的解释保持一致来确保模型具有“正确的原因”。如果没有这种约束,传统的持续学习算法学习新任务时,解释就会出现偏差,遗忘也会增加。我们展示了如何将RRR轻松地添加到任何基于记忆或正则化的方法中,从而减少遗忘,更重要的是,改进了模型解释。我们已经在标准和少数镜头设置中评估了我们的方法,并观察到使用不同架构和技术生成模型解释的各种CL方法的一致改进,并证明了我们的方法显示了可解释性和持续学习之间的有希望的联系。我们的代码可在https://github.com/SaynaEbrahimi/Remembering-for-the-Right-Reasons上获得。
{"title":"Remembering for the right reasons: Explanations reduce catastrophic forgetting","authors":"Sayna Ebrahimi,&nbsp;Suzanne Petryk,&nbsp;Akash Gokul,&nbsp;William Gan,&nbsp;Joseph E. Gonzalez,&nbsp;Marcus Rohrbach,&nbsp;Trevor Darrell","doi":"10.1002/ail2.44","DOIUrl":"https://doi.org/10.1002/ail2.44","url":null,"abstract":"<p>The goal of continual learning (CL) is to learn a sequence of tasks without suffering from the phenomenon of catastrophic forgetting. Previous work has shown that leveraging memory in the form of a replay buffer can reduce performance degradation on prior tasks. We hypothesize that forgetting can be further reduced when the model is encouraged to remember the <i>evidence</i> for previously made decisions. As a first step towards exploring this hypothesis, we propose a simple novel training paradigm, called Remembering for the Right Reasons (RRR), that additionally stores visual model explanations for each example in the buffer and ensures the model has “the right reasons” for its predictions by encouraging its explanations to remain consistent with those used to make decisions at training time. Without this constraint, there is a drift in explanations and increase in forgetting as conventional continual learning algorithms learn new tasks. We demonstrate how RRR can be easily added to any memory or regularization-based approach and results in reduced forgetting, and more importantly, improved model explanations. We have evaluated our approach in the standard and few-shot settings and observed a consistent improvement across various CL approaches using different architectures and techniques to generate model explanations and demonstrated our approach showing a promising connection between explainability and continual learning. Our code is available at https://github.com/SaynaEbrahimi/Remembering-for-the-Right-Reasons.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.44","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"137488003","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Patching interpretable And-Or-Graph knowledge representation using augmented reality 使用增强现实修补可解释的And-Or-Graph知识表示
Pub Date : 2021-10-20 DOI: 10.1002/ail2.43
Hangxin Liu, Yixin Zhu, Song-Chun Zhu

We present a novel augmented reality (AR) interface to provide effective means to diagnose a robot's erroneous behaviors, endow it with new skills, and patch its knowledge structure represented by an And-Or-Graph (AOG). Specifically, an AOG representation of opening medicine bottles is learned from human demonstration and yields a hierarchical structure that captures the spatiotemporal compositional nature of the given task, which is highly interpretable for the users. Through a series of psychological experiments, we demonstrate that the explanations of a robotic system, inherited from and produced by the AOG, can better foster human trust compared to other forms of explanations. Moreover, by visualizing the knowledge structure and robot states, the AR interface allows human users to intuitively understand what the robot knows, supervise the robot's task planner, and interactively teach the robot with new actions. Together, users can quickly identify the reasons for failures and conveniently patch the current knowledge structure to prevent future errors. This capability demonstrates the interpretability of our knowledge representation and the new forms of interactions afforded by the proposed AR interface.

本文提出了一种新的增强现实(AR)界面,为机器人错误行为诊断、赋予机器人新技能、修补其知识结构提供了有效手段。具体来说,打开药瓶的AOG表示是从人类演示中学习的,并产生一个层次结构,该结构捕获了给定任务的时空组成性质,这对用户来说是高度可解释的。通过一系列的心理学实验,我们证明,与其他形式的解释相比,继承并由AOG产生的机器人系统的解释可以更好地培养人类的信任。此外,通过可视化的知识结构和机器人状态,AR界面可以让人类用户直观地了解机器人知道什么,监督机器人的任务计划,并交互式地教机器人新的动作。用户可以快速识别故障的原因,并方便地修补当前的知识结构,以防止未来的错误。这种能力证明了我们的知识表示的可解释性以及所提出的AR接口提供的新形式的交互。
{"title":"Patching interpretable And-Or-Graph knowledge representation using augmented reality","authors":"Hangxin Liu,&nbsp;Yixin Zhu,&nbsp;Song-Chun Zhu","doi":"10.1002/ail2.43","DOIUrl":"10.1002/ail2.43","url":null,"abstract":"<p>We present a novel augmented reality (AR) interface to provide effective means to diagnose a robot's erroneous behaviors, endow it with new skills, and patch its knowledge structure represented by an And-Or-Graph (AOG). Specifically, an AOG representation of opening medicine bottles is learned from human demonstration and yields a hierarchical structure that captures the spatiotemporal compositional nature of the given task, which is highly interpretable for the users. Through a series of psychological experiments, we demonstrate that the explanations of a robotic system, inherited from and produced by the AOG, can better foster human trust compared to other forms of explanations. Moreover, by visualizing the knowledge structure and robot states, the AR interface allows human users to intuitively understand what the robot knows, supervise the robot's task planner, and interactively teach the robot with new actions. Together, users can quickly identify the reasons for failures and conveniently patch the current knowledge structure to prevent future errors. This capability demonstrates the interpretability of our knowledge representation and the new forms of interactions afforded by the proposed AR interface.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.43","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46548240","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Explainable, interactive content-based image retrieval 可解释的,交互式的基于内容的图像检索
Pub Date : 2021-10-19 DOI: 10.1002/ail2.41
Bhavan Vasu, Brian Hu, Bo Dong, Roddy Collins, Anthony Hoogs

Quantifying the value of explanations in a human-in-the-loop (HITL) system is difficult. Previous methods either measure explanation-specific values that do not correspond to user tasks and needs or poll users on how useful they find the explanations to be. In this work, we quantify how much explanations help the user through a utility-based paradigm that measures change in task performance when using explanations vs not. Our chosen task is content-based image retrieval (CBIR), which has well-established baselines and performance metrics independent of explainability. We extend an existing HITL image retrieval system that incorporates user feedback with similarity-based saliency maps (SBSM) that indicate to the user which parts of the retrieved images are most similar to the query image. The system helps the user understand what it is paying attention to through saliency maps, and the user helps the system understand their goal through saliency-guided relevance feedback. Using the MS-COCO dataset, a standard object detection and segmentation dataset, we conducted extensive, crowd-sourced experiments validating that SBSM improves interactive image retrieval. Although the performance increase is modest in the general case, in more difficult cases such as cluttered scenes, using explanations yields an 6.5% increase in accuracy. To the best of our knowledge, this is the first large-scale user study showing that visual saliency map explanations improve performance on a real-world, interactive task. Our utility-based evaluation paradigm is general and potentially applicable to any task for which explainability can be incorporated.

在人在循环(HITL)系统中,量化解释的价值是困难的。以前的方法要么测量与用户任务和需求不对应的特定于解释的值,要么调查用户对解释的有用程度。在这项工作中,我们通过一种基于效用的范式来量化解释对用户的帮助程度,该范式衡量使用解释与不使用解释时任务性能的变化。我们选择的任务是基于内容的图像检索(CBIR),它具有良好的基线和独立于可解释性的性能指标。我们扩展了现有的HITL图像检索系统,该系统将用户反馈与基于相似性的显著性映射(SBSM)结合在一起,该映射向用户指示检索图像的哪些部分与查询图像最相似。系统通过显著性地图帮助用户理解自己关注的是什么,用户通过显著性引导的相关性反馈帮助系统理解自己的目标。使用MS-COCO数据集(一个标准的目标检测和分割数据集),我们进行了广泛的众包实验,验证了SBSM改进了交互式图像检索。虽然在一般情况下,性能的提高是适度的,但在更困难的情况下,比如混乱的场景,使用解释可以提高6.5%的准确性。据我们所知,这是第一次大规模的用户研究,表明视觉显著性地图解释可以提高现实世界中交互式任务的表现。我们基于效用的评估范式是通用的,并且潜在地适用于任何可解释性可以被纳入的任务。
{"title":"Explainable, interactive content-based image retrieval","authors":"Bhavan Vasu,&nbsp;Brian Hu,&nbsp;Bo Dong,&nbsp;Roddy Collins,&nbsp;Anthony Hoogs","doi":"10.1002/ail2.41","DOIUrl":"10.1002/ail2.41","url":null,"abstract":"<p>Quantifying the value of explanations in a human-in-the-loop (HITL) system is difficult. Previous methods either measure explanation-specific values that do not correspond to user tasks and needs or poll users on how useful they find the explanations to be. In this work, we quantify how much explanations help the user through a utility-based paradigm that measures change in task performance when using explanations vs not. Our chosen task is content-based image retrieval (CBIR), which has well-established baselines and performance metrics independent of explainability. We extend an existing HITL image retrieval system that incorporates user feedback with similarity-based saliency maps (SBSM) that indicate to the user which parts of the retrieved images are most similar to the query image. The system helps the user understand what it is paying attention to through saliency maps, and the user helps the system understand their goal through saliency-guided relevance feedback. Using the MS-COCO dataset, a standard object detection and segmentation dataset, we conducted extensive, crowd-sourced experiments validating that SBSM improves interactive image retrieval. Although the performance increase is modest in the general case, in more difficult cases such as cluttered scenes, using explanations yields an 6.5% increase in accuracy. To the best of our knowledge, this is the first large-scale user study showing that visual saliency map explanations improve performance on a real-world, interactive task. Our utility-based evaluation paradigm is general and potentially applicable to any task for which explainability can be incorporated.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.41","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"102959774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
User-guided global explanations for deep image recognition: A user study 深度图像识别的用户导向全局解释:用户研究
Pub Date : 2021-10-19 DOI: 10.1002/ail2.42
Mandana Hamidi-Haines, Zhongang Qi, Alan Fern, Fuxin Li, Prasad Tadepalli

We study a user-guided approach for producing global explanations of deep networks for image recognition. The global explanations are produced with respect to a test data set and give the overall frequency of different “recognition reasons” across the data. Each reason corresponds to a small number of the most significant human-recognizable visual concepts used by the network. The key challenge is that the visual concepts cannot be predetermined and those concepts will often not correspond to existing vocabulary or have labeled data sets. We address this issue via an interactive-naming interface, which allows users to freely cluster significant image regions in the data into visually similar concepts. Our main contribution is a user study on two visual recognition tasks. The results show that the participants were able to produce a small number of visual concepts sufficient for explanation and that there was significant agreement among the concepts, and hence global explanations, produced by different participants.

我们研究了一种用户导向的方法,用于生成用于图像识别的深度网络的全局解释。全局解释是根据测试数据集产生的,并给出了数据中不同“识别原因”的总体频率。每个原因对应于网络使用的少数最重要的人类可识别的视觉概念。关键的挑战是,视觉概念不能预先确定,这些概念通常不对应于现有的词汇表或有标记的数据集。我们通过交互式命名界面解决了这个问题,该界面允许用户自由地将数据中的重要图像区域聚类到视觉上相似的概念中。我们的主要贡献是对两个视觉识别任务的用户研究。结果表明,参与者能够产生少量足以解释的视觉概念,并且这些概念之间存在显著的一致性,因此不同参与者产生的整体解释。
{"title":"User-guided global explanations for deep image recognition: A user study","authors":"Mandana Hamidi-Haines,&nbsp;Zhongang Qi,&nbsp;Alan Fern,&nbsp;Fuxin Li,&nbsp;Prasad Tadepalli","doi":"10.1002/ail2.42","DOIUrl":"https://doi.org/10.1002/ail2.42","url":null,"abstract":"<p>We study a user-guided approach for producing global explanations of deep networks for image recognition. The global explanations are produced with respect to a test data set and give the overall frequency of different “recognition reasons” across the data. Each reason corresponds to a small number of the most significant human-recognizable visual concepts used by the network. The key challenge is that the visual concepts cannot be predetermined and those concepts will often not correspond to existing vocabulary or have labeled data sets. We address this issue via an interactive-naming interface, which allows users to freely cluster significant image regions in the data into visually similar concepts. Our main contribution is a user study on two visual recognition tasks. The results show that the participants were able to produce a small number of visual concepts sufficient for explanation and that there was significant agreement among the concepts, and hence global explanations, produced by different participants.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.42","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"137863524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Applied AI letters
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1