{"title":"分层强化学习的自动复合动作发现","authors":"Josiah Laivins, Minwoo Lee","doi":"10.1109/SSCI44817.2019.9003053","DOIUrl":null,"url":null,"abstract":"Even with recent advances in standard reinforcement learning, hierarchical reinforcement learning has been discussed as a promising approach to solve complex problems. From human-designed abstraction, planning or learning with composite actions are well-understood, but without human intervention, producing abstract (or composite) actions automatically is one of the remaining challenges. We separate this action discovery from reinforcement learning problem and investigate on searching impactful composite actions that can make meaningful changes in state space. We discuss the efficiency and flexibility of the suggested model by interpreting and analyzing the discovered composite actions with different deep reinforcement learning algorithms in different environments.","PeriodicalId":6729,"journal":{"name":"2019 IEEE Symposium Series on Computational Intelligence (SSCI)","volume":"56 1","pages":"198-205"},"PeriodicalIF":0.0000,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Automatic Composite Action Discovery for Hierarchical Reinforcement Learning\",\"authors\":\"Josiah Laivins, Minwoo Lee\",\"doi\":\"10.1109/SSCI44817.2019.9003053\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Even with recent advances in standard reinforcement learning, hierarchical reinforcement learning has been discussed as a promising approach to solve complex problems. From human-designed abstraction, planning or learning with composite actions are well-understood, but without human intervention, producing abstract (or composite) actions automatically is one of the remaining challenges. We separate this action discovery from reinforcement learning problem and investigate on searching impactful composite actions that can make meaningful changes in state space. We discuss the efficiency and flexibility of the suggested model by interpreting and analyzing the discovered composite actions with different deep reinforcement learning algorithms in different environments.\",\"PeriodicalId\":6729,\"journal\":{\"name\":\"2019 IEEE Symposium Series on Computational Intelligence (SSCI)\",\"volume\":\"56 1\",\"pages\":\"198-205\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 IEEE Symposium Series on Computational Intelligence (SSCI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SSCI44817.2019.9003053\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE Symposium Series on Computational Intelligence (SSCI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SSCI44817.2019.9003053","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Automatic Composite Action Discovery for Hierarchical Reinforcement Learning
Even with recent advances in standard reinforcement learning, hierarchical reinforcement learning has been discussed as a promising approach to solve complex problems. From human-designed abstraction, planning or learning with composite actions are well-understood, but without human intervention, producing abstract (or composite) actions automatically is one of the remaining challenges. We separate this action discovery from reinforcement learning problem and investigate on searching impactful composite actions that can make meaningful changes in state space. We discuss the efficiency and flexibility of the suggested model by interpreting and analyzing the discovered composite actions with different deep reinforcement learning algorithms in different environments.