{"title":"Aligning Judgment Using Task Context and Explanations to Improve Human-Recommender System Performance","authors":"Divya Srivastava, Karen M. Feigh","doi":"arxiv-2409.10717","DOIUrl":null,"url":null,"abstract":"Recommender systems, while a powerful decision making tool, are often\noperationalized as black box models, such that their AI algorithms are not\naccessible or interpretable by human operators. This in turn can cause\nconfusion and frustration for the operator and result in unsatisfactory\noutcomes. While the field of explainable AI has made remarkable strides in\naddressing this challenge by focusing on interpreting and explaining the\nalgorithms to human operators, there are remaining gaps in the human's\nunderstanding of the recommender system. This paper investigates the relative\nimpact of using context, properties of the decision making task and\nenvironment, to align human and AI algorithm understanding of the state of the\nworld, i.e. judgment, to improve joint human-recommender performance as\ncompared to utilizing post-hoc algorithmic explanations. We conducted an\nempirical, between-subjects experiment in which participants were asked to work\nwith an automated recommender system to complete a decision making task. We\nmanipulated the method of transparency (shared contextual information to\nsupport shared judgment vs algorithmic explanations) and record the human's\nunderstanding of the task, the recommender system, and their overall\nperformance. We found that both techniques yielded equivalent agreement on\nfinal decisions. However, those who saw task context had less tendency to\nover-rely on the recommender system and were able to better pinpoint in what\nconditions the AI erred. Both methods improved participants' confidence in\ntheir own decision making, and increased mental demand equally and frustration\nnegligibly. These results present an alternative approach to improving team\nperformance to post-hoc explanations and illustrate the impact of judgment on\nhuman cognition in working with recommender systems.","PeriodicalId":501541,"journal":{"name":"arXiv - CS - Human-Computer Interaction","volume":"65 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Human-Computer Interaction","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10717","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Recommender systems, while a powerful decision making tool, are often
operationalized as black box models, such that their AI algorithms are not
accessible or interpretable by human operators. This in turn can cause
confusion and frustration for the operator and result in unsatisfactory
outcomes. While the field of explainable AI has made remarkable strides in
addressing this challenge by focusing on interpreting and explaining the
algorithms to human operators, there are remaining gaps in the human's
understanding of the recommender system. This paper investigates the relative
impact of using context, properties of the decision making task and
environment, to align human and AI algorithm understanding of the state of the
world, i.e. judgment, to improve joint human-recommender performance as
compared to utilizing post-hoc algorithmic explanations. We conducted an
empirical, between-subjects experiment in which participants were asked to work
with an automated recommender system to complete a decision making task. We
manipulated the method of transparency (shared contextual information to
support shared judgment vs algorithmic explanations) and record the human's
understanding of the task, the recommender system, and their overall
performance. We found that both techniques yielded equivalent agreement on
final decisions. However, those who saw task context had less tendency to
over-rely on the recommender system and were able to better pinpoint in what
conditions the AI erred. Both methods improved participants' confidence in
their own decision making, and increased mental demand equally and frustration
negligibly. These results present an alternative approach to improving team
performance to post-hoc explanations and illustrate the impact of judgment on
human cognition in working with recommender systems.