{"title":"多任务训练","authors":"Upasana Pattnaik, Minwoo Lee","doi":"10.1109/SSCI50451.2021.9659943","DOIUrl":null,"url":null,"abstract":"Adapting feedback-driven deep reinforcement learning (DRL) algorithms to real-world problems requires developing robust systems that balance generalization and specialization. DRL algorithms powered by deep neural network function approximation tend to over-fit and perform poorly in new situations. Multi-task learning is a popular approach to reduce over-fitting by increasing input diversity, which in turn improves generalization capabilities. However, optimizing for multiple tasks often leads to distraction and performance oscillation. In this work, transfer learning paradigm Practice is introduced as an auxiliary task to stabilize distributed multi-task learning and enhance generalization. Experimental results demonstrate that the DRL algorithm supplemented with the state dynamics information produced by Practice improves performance.","PeriodicalId":255763,"journal":{"name":"2021 IEEE Symposium Series on Computational Intelligence (SSCI)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multi-task Transfer with Practice\",\"authors\":\"Upasana Pattnaik, Minwoo Lee\",\"doi\":\"10.1109/SSCI50451.2021.9659943\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Adapting feedback-driven deep reinforcement learning (DRL) algorithms to real-world problems requires developing robust systems that balance generalization and specialization. DRL algorithms powered by deep neural network function approximation tend to over-fit and perform poorly in new situations. Multi-task learning is a popular approach to reduce over-fitting by increasing input diversity, which in turn improves generalization capabilities. However, optimizing for multiple tasks often leads to distraction and performance oscillation. In this work, transfer learning paradigm Practice is introduced as an auxiliary task to stabilize distributed multi-task learning and enhance generalization. Experimental results demonstrate that the DRL algorithm supplemented with the state dynamics information produced by Practice improves performance.\",\"PeriodicalId\":255763,\"journal\":{\"name\":\"2021 IEEE Symposium Series on Computational Intelligence (SSCI)\",\"volume\":\"67 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE Symposium Series on Computational Intelligence (SSCI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SSCI50451.2021.9659943\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE Symposium Series on Computational Intelligence (SSCI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SSCI50451.2021.9659943","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Adapting feedback-driven deep reinforcement learning (DRL) algorithms to real-world problems requires developing robust systems that balance generalization and specialization. DRL algorithms powered by deep neural network function approximation tend to over-fit and perform poorly in new situations. Multi-task learning is a popular approach to reduce over-fitting by increasing input diversity, which in turn improves generalization capabilities. However, optimizing for multiple tasks often leads to distraction and performance oscillation. In this work, transfer learning paradigm Practice is introduced as an auxiliary task to stabilize distributed multi-task learning and enhance generalization. Experimental results demonstrate that the DRL algorithm supplemented with the state dynamics information produced by Practice improves performance.