Low Rank Optimization for Efficient Deep Learning: Making a Balance Between Compact Architecture And Fast Training

IF 1.9 3区 计算机科学 Q3 AUTOMATION & CONTROL SYSTEMS Journal of Systems Engineering and Electronics Pub Date : 2023-12-12 DOI:10.23919/jsee.2023.000159
Xinwei Ou, Zhangxin Chen, Ce Zhu, Yipeng Liu
{"title":"Low Rank Optimization for Efficient Deep Learning: Making a Balance Between Compact Architecture And Fast Training","authors":"Xinwei Ou, Zhangxin Chen, Ce Zhu, Yipeng Liu","doi":"10.23919/jsee.2023.000159","DOIUrl":null,"url":null,"abstract":"Deep neural networks (DNNs) have achieved great success in many data processing applications. However, high computational complexity and storage cost make deep learning difficult to be used on resource-constrained devices, and it is not environmental-friendly with much power cost. In this paper, we focus on low-rank optimization for efficient deep learning techniques. In the space domain, DNNs are compressed by low rank approximation of the network parameters, which directly reduces the storage requirement with a smaller number of network parameters. In the time domain, the network parameters can be trained in a few subspaces, which enables efficient training for fast convergence. The model compression in the spatial domain is summarized into three categories as pre-train, pre-set, and compression-aware methods, respectively. With a series of integrable techniques discussed, such as sparse pruning, quantization, and entropy coding, we can ensemble them in an integration framework with lower computational complexity and storage. In addition to summary of recent technical advances, we have two findings for motivating future works. One is that the effective rank, derived from the Shannon entropy of the normalized singular values, outperforms other conventional sparse measures such as the <tex xmlns:mml=\"http://www.w3.org/1998/Math/MathML\" xmlns:xlink=\"http://www.w3.org/1999/xlink\">$\\ell_{1}$</tex> norm for network compression. The other is a spatial and temporal balance for tensorized neural networks. For accelerating the training of tensorized neural networks, it is crucial to leverage redundancy for both model compression and subspace training.","PeriodicalId":50030,"journal":{"name":"Journal of Systems Engineering and Electronics","volume":"8 1","pages":""},"PeriodicalIF":1.9000,"publicationDate":"2023-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Systems Engineering and Electronics","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.23919/jsee.2023.000159","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Deep neural networks (DNNs) have achieved great success in many data processing applications. However, high computational complexity and storage cost make deep learning difficult to be used on resource-constrained devices, and it is not environmental-friendly with much power cost. In this paper, we focus on low-rank optimization for efficient deep learning techniques. In the space domain, DNNs are compressed by low rank approximation of the network parameters, which directly reduces the storage requirement with a smaller number of network parameters. In the time domain, the network parameters can be trained in a few subspaces, which enables efficient training for fast convergence. The model compression in the spatial domain is summarized into three categories as pre-train, pre-set, and compression-aware methods, respectively. With a series of integrable techniques discussed, such as sparse pruning, quantization, and entropy coding, we can ensemble them in an integration framework with lower computational complexity and storage. In addition to summary of recent technical advances, we have two findings for motivating future works. One is that the effective rank, derived from the Shannon entropy of the normalized singular values, outperforms other conventional sparse measures such as the $\ell_{1}$ norm for network compression. The other is a spatial and temporal balance for tensorized neural networks. For accelerating the training of tensorized neural networks, it is crucial to leverage redundancy for both model compression and subspace training.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
高效深度学习的低等级优化:在紧凑架构和快速训练之间取得平衡
深度神经网络(DNN)在许多数据处理应用中取得了巨大成功。然而,高计算复杂度和存储成本使得深度学习难以在资源受限的设备上使用,而且它不环保,耗电量大。在本文中,我们将重点研究低秩优化的高效深度学习技术。在空间域,DNN 通过网络参数的低秩逼近进行压缩,从而以较少的网络参数数量直接降低存储需求。在时域中,网络参数可以在几个子空间中进行训练,从而实现高效训练,快速收敛。空间域的模型压缩可归纳为三类,分别是预训练法、预设法和压缩感知法。通过所讨论的一系列可积分技术,如稀疏剪枝、量化和熵编码,我们可以将它们集合在一个积分框架中,从而降低计算复杂度和存储量。除了对近期技术进展的总结,我们还有两个发现可以激励未来的工作。一个是由归一化奇异值的香农熵推导出的有效秩优于其他传统的稀疏度量,如用于网络压缩的 $\ell_{1}$ norm。另一个是张量神经网络的时空平衡。为了加速张量神经网络的训练,利用冗余进行模型压缩和子空间训练至关重要。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Journal of Systems Engineering and Electronics
Journal of Systems Engineering and Electronics 工程技术-工程:电子与电气
CiteScore
4.10
自引率
14.30%
发文量
131
审稿时长
7.5 months
期刊介绍: Information not localized
期刊最新文献
System Error Iterative Identification for Underwater Positioning Based on Spectral Clustering Cloud Control for IIoT in a Cloud-Edge Environment Multi-Network-Region Traffic Cooperative Scheduling in Large-Scale LEO Satellite Networks Quantitative Method for Calculating Spatial Release Region for Laser-Guided Bomb Early Warning of Core Network Capacity in Space-Terrestrial Integrated Networks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1