Towards reconciling usability and usefulness of policy explanations for sequential decision-making systems

Pradyumna Tambwekar, Matthew C. Gombolay
{"title":"Towards reconciling usability and usefulness of policy explanations for sequential decision-making systems","authors":"Pradyumna Tambwekar, Matthew C. Gombolay","doi":"10.3389/frobt.2024.1375490","DOIUrl":null,"url":null,"abstract":"Safefy-critical domains often employ autonomous agents which follow a sequential decision-making setup, whereby the agent follows a policy to dictate the appropriate action at each step. AI-practitioners often employ reinforcement learning algorithms to allow an agent to find the best policy. However, sequential systems often lack clear and immediate signs of wrong actions, with consequences visible only in hindsight, making it difficult to humans to understand system failure. In reinforcement learning, this is referred to as the credit assignment problem. To effectively collaborate with an autonomous system, particularly in a safety-critical setting, explanations should enable a user to better understand the policy of the agent and predict system behavior so that users are cognizant of potential failures and these failures can be diagnosed and mitigated. However, humans are diverse and have innate biases or preferences which may enhance or impair the utility of a policy explanation of a sequential agent. Therefore, in this paper, we designed and conducted human-subjects experiment to identify the factors which influence the perceived usability with the objective usefulness of policy explanations for reinforcement learning agents in a sequential setting. Our study had two factors: the modality of policy explanation shown to the user (Tree, Text, Modified Text, and Programs) and the “first impression” of the agent, i.e., whether the user saw the agent succeed or fail in the introductory calibration video. Our findings characterize a preference-performance tradeoff wherein participants perceived language-based policy explanations to be significantly more useable; however, participants were better able to objectively predict the agent’s behavior when provided an explanation in the form of a decision tree. Our results demonstrate that user-specific factors, such as computer science experience (p < 0.05), and situational factors, such as watching agent crash (p < 0.05), can significantly impact the perception and usefulness of the explanation. This research provides key insights to alleviate prevalent issues regarding innapropriate compliance and reliance, which are exponentially more detrimental in safety-critical settings, providing a path forward for XAI developers for future work on policy-explanations.","PeriodicalId":504612,"journal":{"name":"Frontiers in Robotics and AI","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in Robotics and AI","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3389/frobt.2024.1375490","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Safefy-critical domains often employ autonomous agents which follow a sequential decision-making setup, whereby the agent follows a policy to dictate the appropriate action at each step. AI-practitioners often employ reinforcement learning algorithms to allow an agent to find the best policy. However, sequential systems often lack clear and immediate signs of wrong actions, with consequences visible only in hindsight, making it difficult to humans to understand system failure. In reinforcement learning, this is referred to as the credit assignment problem. To effectively collaborate with an autonomous system, particularly in a safety-critical setting, explanations should enable a user to better understand the policy of the agent and predict system behavior so that users are cognizant of potential failures and these failures can be diagnosed and mitigated. However, humans are diverse and have innate biases or preferences which may enhance or impair the utility of a policy explanation of a sequential agent. Therefore, in this paper, we designed and conducted human-subjects experiment to identify the factors which influence the perceived usability with the objective usefulness of policy explanations for reinforcement learning agents in a sequential setting. Our study had two factors: the modality of policy explanation shown to the user (Tree, Text, Modified Text, and Programs) and the “first impression” of the agent, i.e., whether the user saw the agent succeed or fail in the introductory calibration video. Our findings characterize a preference-performance tradeoff wherein participants perceived language-based policy explanations to be significantly more useable; however, participants were better able to objectively predict the agent’s behavior when provided an explanation in the form of a decision tree. Our results demonstrate that user-specific factors, such as computer science experience (p < 0.05), and situational factors, such as watching agent crash (p < 0.05), can significantly impact the perception and usefulness of the explanation. This research provides key insights to alleviate prevalent issues regarding innapropriate compliance and reliance, which are exponentially more detrimental in safety-critical settings, providing a path forward for XAI developers for future work on policy-explanations.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
努力协调顺序决策系统政策解释的可用性和实用性
安全关键领域通常采用自主代理,这种代理遵循顺序决策设置,即代理在每一步都遵循政策来决定适当的行动。人工智能实践者通常采用强化学习算法,让代理找到最佳策略。然而,顺序系统往往缺乏明确而直接的错误行为迹象,其后果只有在事后才能看到,因此人类很难理解系统的故障。在强化学习中,这被称为信用分配问题。为了有效地与自主系统合作,特别是在对安全至关重要的环境中,解释应使用户能够更好地理解代理的策略并预测系统行为,从而使用户认识到潜在的故障,并对这些故障进行诊断和缓解。然而,人类是多种多样的,他们与生俱来的偏见或偏好可能会增强或削弱顺序代理策略解释的效用。因此,在本文中,我们设计并进行了以人为对象的实验,以确定在顺序环境中影响强化学习代理政策解释的感知可用性和客观有用性的因素。我们的研究有两个因素:向用户展示的策略解释模式(树状、文本、修改文本和程序)和对代理的 "第一印象",即用户在介绍性校准视频中看到代理成功还是失败。我们的研究结果表明,在偏好与性能的权衡中,参与者认为基于语言的策略解释更易于使用;然而,当提供决策树形式的解释时,参与者能更好地客观预测代理的行为。我们的研究结果表明,用户的特定因素,如计算机科学经验(p < 0.05)和情境因素,如观看代理崩溃(p < 0.05),会对解释的感知和实用性产生重大影响。这项研究为缓解普遍存在的不适当遵从和依赖问题提供了重要见解,这些问题在安全关键型环境中具有成倍的危害性,为 XAI 开发人员今后在政策解释方面的工作提供了前进的道路。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Collective predictive coding hypothesis: symbol emergence as decentralized Bayesian inference Adaptive satellite attitude control for varying masses using deep reinforcement learning Towards reconciling usability and usefulness of policy explanations for sequential decision-making systems Semantic learning from keyframe demonstration using object attribute constraints Gaze detection as a social cue to initiate natural human-robot collaboration in an assembly task
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1