{"title":"Simulating Fear as Anticipation of Temporal Differences: An experimental investigation","authors":"L. Dai, J. Broekens","doi":"10.1109/aciiw52867.2021.9666259","DOIUrl":null,"url":null,"abstract":"Humans use emotional expressions to communicate appraisals. Humans also use emotions in evaluating how they are doing compared to their current goals and desires. The Temporal Difference Reinforcement Learning (TDRL) Theory of Emotion proposes a structure for agents to simulate appropriate emotions during the learning process. In previous work, simulations have shown to reproduce plausible emotion dynamics. In this paper we examine the plausibility and intepretability of TDRL-simulated fear, when expressed by the agent. We presented different TDRL-based fear simulation methods to participants ${\\left(n=237\\right)}$ in an online study. Each method used a different action selection protocol for the agent's model-based anticipation process. Results suggest that an ${\\in}$-greedy fear policy ${\\left(\\in=0.1\\right)}$ combined with a long anticipation horizon provides a plausible fear estimation. This is, to our knowledge, the first experimental evidence detailing some of the predictions of the TDRL Theory of Emotion. Our results are of interest to the design of agent learning methods that are transparent to the user.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/aciiw52867.2021.9666259","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Humans use emotional expressions to communicate appraisals. Humans also use emotions in evaluating how they are doing compared to their current goals and desires. The Temporal Difference Reinforcement Learning (TDRL) Theory of Emotion proposes a structure for agents to simulate appropriate emotions during the learning process. In previous work, simulations have shown to reproduce plausible emotion dynamics. In this paper we examine the plausibility and intepretability of TDRL-simulated fear, when expressed by the agent. We presented different TDRL-based fear simulation methods to participants ${\left(n=237\right)}$ in an online study. Each method used a different action selection protocol for the agent's model-based anticipation process. Results suggest that an ${\in}$-greedy fear policy ${\left(\in=0.1\right)}$ combined with a long anticipation horizon provides a plausible fear estimation. This is, to our knowledge, the first experimental evidence detailing some of the predictions of the TDRL Theory of Emotion. Our results are of interest to the design of agent learning methods that are transparent to the user.