{"title":"多任务学习与增强模块","authors":"Zishuo Zheng, Yadong Wei, Zixu Zhao, Xindi Wu, Zhengcheng Li, Pengju Ren","doi":"10.1109/ICDSP.2018.8631696","DOIUrl":null,"url":null,"abstract":"In multitask learning (MTL) paradigm, modularity is an effective way to achieve component and parameter reuse as well as system extensibility. In this work, we introduce two enhanced modules named res-fire module (RF) and dimension reduction module(DR) to improve the performance of modular MTL network – PathNet. In addition, in order to further improve the transfer ability of the network, we apply learnable scale parameters to merge the outputs of the modules in the same layer and then scatter to the next layer. Experiments on MNIST, CIFAR, SVHN and MiniImageNet demonstrate that, with the similar scale as PathNet, our architecture achieves remarkable improvement in both transfer ability and expression ability. Our design used x5.23 fewer generations to achieve 99% accuracy on a source-to-target MNIST classification task compared with DeepMind’s PathNet. We also increase the accuracy of CIFARSVHN transfer task by x1.9. Also we get 70.75% accuracy on miniImageNet.","PeriodicalId":218806,"journal":{"name":"2018 IEEE 23rd International Conference on Digital Signal Processing (DSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multitask Learning With Enhanced Modules\",\"authors\":\"Zishuo Zheng, Yadong Wei, Zixu Zhao, Xindi Wu, Zhengcheng Li, Pengju Ren\",\"doi\":\"10.1109/ICDSP.2018.8631696\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In multitask learning (MTL) paradigm, modularity is an effective way to achieve component and parameter reuse as well as system extensibility. In this work, we introduce two enhanced modules named res-fire module (RF) and dimension reduction module(DR) to improve the performance of modular MTL network – PathNet. In addition, in order to further improve the transfer ability of the network, we apply learnable scale parameters to merge the outputs of the modules in the same layer and then scatter to the next layer. Experiments on MNIST, CIFAR, SVHN and MiniImageNet demonstrate that, with the similar scale as PathNet, our architecture achieves remarkable improvement in both transfer ability and expression ability. Our design used x5.23 fewer generations to achieve 99% accuracy on a source-to-target MNIST classification task compared with DeepMind’s PathNet. We also increase the accuracy of CIFARSVHN transfer task by x1.9. Also we get 70.75% accuracy on miniImageNet.\",\"PeriodicalId\":218806,\"journal\":{\"name\":\"2018 IEEE 23rd International Conference on Digital Signal Processing (DSP)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2018 IEEE 23rd International Conference on Digital Signal Processing (DSP)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICDSP.2018.8631696\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE 23rd International Conference on Digital Signal Processing (DSP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDSP.2018.8631696","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
In multitask learning (MTL) paradigm, modularity is an effective way to achieve component and parameter reuse as well as system extensibility. In this work, we introduce two enhanced modules named res-fire module (RF) and dimension reduction module(DR) to improve the performance of modular MTL network – PathNet. In addition, in order to further improve the transfer ability of the network, we apply learnable scale parameters to merge the outputs of the modules in the same layer and then scatter to the next layer. Experiments on MNIST, CIFAR, SVHN and MiniImageNet demonstrate that, with the similar scale as PathNet, our architecture achieves remarkable improvement in both transfer ability and expression ability. Our design used x5.23 fewer generations to achieve 99% accuracy on a source-to-target MNIST classification task compared with DeepMind’s PathNet. We also increase the accuracy of CIFARSVHN transfer task by x1.9. Also we get 70.75% accuracy on miniImageNet.