{"title":"基于课程学习的区域划分对象传输","authors":"Gyuho Eoh, T. Park","doi":"10.23919/ICCAS52745.2021.9649829","DOIUrl":null,"url":null,"abstract":"This paper presents a deep reinforcement learning (DRL)-based object transportation technique using a region-partitioning curriculum. Previous studies on object transportation using DRL algorithms have suffered a sparse reward problem where a robot cannot gain success experiences frequently due to random actions at the learning stage. To solve the sparse reward problem, we partition pose-initialization regions based on the distance between an object and goal, then a robot gradually extends the partitioned regions as training episodes increase. The robot has more success opportunities using this method, and thus, it can learn effective object transportation methods quickly. We demonstrate simulations to verify the proposed method.","PeriodicalId":411064,"journal":{"name":"2021 21st International Conference on Control, Automation and Systems (ICCAS)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Curriculum Learning-based Object Transportation using Region Partitioning\",\"authors\":\"Gyuho Eoh, T. Park\",\"doi\":\"10.23919/ICCAS52745.2021.9649829\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper presents a deep reinforcement learning (DRL)-based object transportation technique using a region-partitioning curriculum. Previous studies on object transportation using DRL algorithms have suffered a sparse reward problem where a robot cannot gain success experiences frequently due to random actions at the learning stage. To solve the sparse reward problem, we partition pose-initialization regions based on the distance between an object and goal, then a robot gradually extends the partitioned regions as training episodes increase. The robot has more success opportunities using this method, and thus, it can learn effective object transportation methods quickly. We demonstrate simulations to verify the proposed method.\",\"PeriodicalId\":411064,\"journal\":{\"name\":\"2021 21st International Conference on Control, Automation and Systems (ICCAS)\",\"volume\":\"40 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-10-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 21st International Conference on Control, Automation and Systems (ICCAS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.23919/ICCAS52745.2021.9649829\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 21st International Conference on Control, Automation and Systems (ICCAS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.23919/ICCAS52745.2021.9649829","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Curriculum Learning-based Object Transportation using Region Partitioning
This paper presents a deep reinforcement learning (DRL)-based object transportation technique using a region-partitioning curriculum. Previous studies on object transportation using DRL algorithms have suffered a sparse reward problem where a robot cannot gain success experiences frequently due to random actions at the learning stage. To solve the sparse reward problem, we partition pose-initialization regions based on the distance between an object and goal, then a robot gradually extends the partitioned regions as training episodes increase. The robot has more success opportunities using this method, and thus, it can learn effective object transportation methods quickly. We demonstrate simulations to verify the proposed method.