Clustering-Guided Incremental Learning of Tasks

Y. Kim, Eunwoo Kim
{"title":"Clustering-Guided Incremental Learning of Tasks","authors":"Y. Kim, Eunwoo Kim","doi":"10.1109/ICOIN50884.2021.9334003","DOIUrl":null,"url":null,"abstract":"Incremental deep learning aims to learn a sequence of tasks while avoiding forgetting their knowledge. One naïve approach using a deep architecture is to increase the capacity of the architecture as the number of tasks increases. However, this is followed by heavy memory consumption and makes the approach not practical. If we attempt to avoid such an issue with a fixed capacity, we encounter another challenging problem called catastrophic forgetting, which leads to a notable degradation of performance on previously learned tasks. To overcome these problems, we propose a clustering-guided incremental learning approach that can mitigate catastrophic forgetting while not increasing the capacity of an architecture. The proposed approach adopts a parameter-splitting strategy to assign a subset of parameters in an architecture for each task to prevent forgetting. It uses a clustering approach to discover the relationship between tasks by storing a few samples per task. When we learn a new task, we utilize the knowledge of the relevant tasks together with the current task to improve performance. This approach could maximize the efficiency of the approach realized in a single fixed architecture. Experimental results with a number of fine-grained datasets show that our method outperforms existing competitors.","PeriodicalId":6741,"journal":{"name":"2021 International Conference on Information Networking (ICOIN)","volume":"1 1","pages":"417-421"},"PeriodicalIF":0.0000,"publicationDate":"2021-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Information Networking (ICOIN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICOIN50884.2021.9334003","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Incremental deep learning aims to learn a sequence of tasks while avoiding forgetting their knowledge. One naïve approach using a deep architecture is to increase the capacity of the architecture as the number of tasks increases. However, this is followed by heavy memory consumption and makes the approach not practical. If we attempt to avoid such an issue with a fixed capacity, we encounter another challenging problem called catastrophic forgetting, which leads to a notable degradation of performance on previously learned tasks. To overcome these problems, we propose a clustering-guided incremental learning approach that can mitigate catastrophic forgetting while not increasing the capacity of an architecture. The proposed approach adopts a parameter-splitting strategy to assign a subset of parameters in an architecture for each task to prevent forgetting. It uses a clustering approach to discover the relationship between tasks by storing a few samples per task. When we learn a new task, we utilize the knowledge of the relevant tasks together with the current task to improve performance. This approach could maximize the efficiency of the approach realized in a single fixed architecture. Experimental results with a number of fine-grained datasets show that our method outperforms existing competitors.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
聚类引导的任务增量学习
增量深度学习旨在学习一系列任务,同时避免忘记他们的知识。使用深度架构的一种naïve方法是随着任务数量的增加而增加架构的容量。然而,随之而来的是大量的内存消耗,使得该方法不实用。如果我们试图用固定的能力来避免这样的问题,我们就会遇到另一个具有挑战性的问题,即灾难性遗忘,它会导致我们在完成之前学习过的任务时的表现显著下降。为了克服这些问题,我们提出了一种聚类引导的增量学习方法,该方法可以减轻灾难性遗忘,同时不增加架构的容量。该方法采用参数分离策略,为每个任务分配一个体系结构的参数子集,以防止遗忘。它使用聚类方法通过为每个任务存储一些样本来发现任务之间的关系。当我们学习一项新任务时,我们利用相关任务的知识和当前任务来提高表现。这种方法可以最大限度地提高在单一固定体系结构中实现的方法的效率。在大量细粒度数据集上的实验结果表明,我们的方法优于现有的竞争对手。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
A Study on the Cluster-wise Regression Model for Bead Width in the Automatic GMA Welding GDFed: Dynamic Federated Learning for Heterogenous Device Using Graph Neural Network A Solution for Recovering Network Topology with Missing Links using Sparse Modeling Real-time health monitoring system design based on optical camera communication Multimedia Contents Retrieval based on 12-Mood Vector
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1