Boosting share routing for multi-task learning.

Chen Xiaokai, Gu Xiaoguang, Fu Libo
{"title":"Boosting share routing for multi-task learning.","authors":"Chen Xiaokai, Gu Xiaoguang, Fu Libo","doi":"10.1145/3442442.3452323","DOIUrl":null,"url":null,"abstract":"Multi-task learning (MTL) aims to make full use of the knowledge contained in multi-task supervision signals to improve the overall performance. How to make the knowledge of multiple tasks shared appropriately is an open problem for MTL. Most existing deep MTL models are based on parameter sharing. However, suitable sharing mechanism is hard to design as the relationship among tasks is complicated. In this paper, we propose a general framework called Multi-Task Neural Architecture Search (MTNAS) to efficiently find a suitable sharing route for a given MTL problem. MTNAS modularizes the sharing part into multiple layers of sub-networks. It allows sparse connection among these sub-networks and soft sharing based on gating is enabled for a certain route. Benefiting from such setting, each candidate architecture in our search space defines a dynamic sparse sharing route which is more flexible compared with full-sharing in previous approaches. We show that existing typical sharing approaches are sub-graphs in our search space. Extensive experiments on three real-world recommendation datasets demonstrate MTANS achieves consistent improvement compared with single-task models and typical multi-task methods while maintaining high computation efficiency. Furthermore, in-depth experiments demonstrates that MTNAS can learn suitable sparse route to mitigate negative transfer.","PeriodicalId":8468,"journal":{"name":"arXiv: Learning","volume":"102 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv: Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3442442.3452323","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 8

Abstract

Multi-task learning (MTL) aims to make full use of the knowledge contained in multi-task supervision signals to improve the overall performance. How to make the knowledge of multiple tasks shared appropriately is an open problem for MTL. Most existing deep MTL models are based on parameter sharing. However, suitable sharing mechanism is hard to design as the relationship among tasks is complicated. In this paper, we propose a general framework called Multi-Task Neural Architecture Search (MTNAS) to efficiently find a suitable sharing route for a given MTL problem. MTNAS modularizes the sharing part into multiple layers of sub-networks. It allows sparse connection among these sub-networks and soft sharing based on gating is enabled for a certain route. Benefiting from such setting, each candidate architecture in our search space defines a dynamic sparse sharing route which is more flexible compared with full-sharing in previous approaches. We show that existing typical sharing approaches are sub-graphs in our search space. Extensive experiments on three real-world recommendation datasets demonstrate MTANS achieves consistent improvement compared with single-task models and typical multi-task methods while maintaining high computation efficiency. Furthermore, in-depth experiments demonstrates that MTNAS can learn suitable sparse route to mitigate negative transfer.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
促进多任务学习的共享路由。
多任务学习(MTL)旨在充分利用多任务监督信号中包含的知识来提高整体绩效。如何使多任务的知识适当地共享是MTL的一个开放性问题。现有的深度MTL模型大多是基于参数共享的。但由于任务间关系复杂,难以设计合适的共享机制。在本文中,我们提出了一个通用的框架,称为多任务神经结构搜索(MTNAS),以有效地为给定的MTL问题找到合适的共享路由。MTNAS将共享部分模块化为多层子网。它允许这些子网之间的稀疏连接,并对某条路由启用基于门控的软共享。得益于这样的设置,我们的搜索空间中的每个候选体系结构都定义了一个动态的稀疏共享路由,与之前的完全共享方法相比,该路由更加灵活。我们表明,现有的典型共享方法是搜索空间中的子图。在三个真实推荐数据集上的大量实验表明,与单任务模型和典型的多任务方法相比,MTANS在保持较高计算效率的同时取得了一致的改进。此外,深入的实验表明,MTNAS可以学习合适的稀疏路径来减轻负迁移。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
High dimensional robust M-estimation : arbitrary corruption and heavy tails Boosting share routing for multi-task learning. Clustering Residential Electricity Consumption Data to Create Archetypes that Capture Household Behaviour in South Africa Synthetic Observational Health Data with GANs: from slow adoption to a boom in medical research and ultimately digital twins? A Review of Privacy-Preserving Federated Learning for the Internet-of-Things
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1