TinyFEL: Communication, Computation, and Memory Efficient Tiny Federated Edge Learning via Model Sparse Update

IF 8.9 1区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS IEEE Internet of Things Journal Pub Date : 2024-11-15 DOI:10.1109/JIOT.2024.3499375
Qimei Chen;Han Cheng;Yipeng Liang;Guangxu Zhu;Ming Li;Hao Jiang
{"title":"TinyFEL: Communication, Computation, and Memory Efficient Tiny Federated Edge Learning via Model Sparse Update","authors":"Qimei Chen;Han Cheng;Yipeng Liang;Guangxu Zhu;Ming Li;Hao Jiang","doi":"10.1109/JIOT.2024.3499375","DOIUrl":null,"url":null,"abstract":"Federated edge learning (FEL) is regarded as a promising distributed machine learning paradigm to reduce transmission latency and resources as well as preserve raw data privacy by collaboratively training local deep learning models across multiple edge devices. However, with the development of artificial intelligence (AI) technologies, the size of neural network models grows exponentially with their parameters to meet variable application requirements, which poses significant challenges to the computation, communication, and memory abilities of edge devices. Existing designs typically focus on either communication or computation efficiency without caring each device’s memory ability. To deal with the above issues, we first introduce a novel model sparse update enabled tiny FEL (TinyFEL) architecture, which terminates the backpropagation early in local model training processes. Therefore, the proposed TinyFEL can reduce local memory occupation and lessen the communication-and-computation burden. Furthermore, we propose a parameter splitting mechanism instead of transmitting the full model, only a part of updated layers of parameters is transmitted for aggregation, which significantly reduced the communication overheads. Thereafter, we develop a communication and computation latency minimization problem to accelerate the training of TinyFEL. To this end, we theoretically analyze the convergence performance of TinyFEL, which unveils the mathematical relationship among sparse update ratio assignment, device selection, and learning performance. Then, a joint sparse update ratio assignment, device selection, and resource allocation strategy is introduced based on the alternating direction method of multipliers (ADMMs) and block coordinate descent (BCD) algorithms. Numerical results indicate that our proposed TinyFEL can reduce training memory occupation by over 40% than the traditional FEL at the cost of negligible accuracy loss.","PeriodicalId":54347,"journal":{"name":"IEEE Internet of Things Journal","volume":"12 7","pages":"8247-8260"},"PeriodicalIF":8.9000,"publicationDate":"2024-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Internet of Things Journal","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10753457/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Federated edge learning (FEL) is regarded as a promising distributed machine learning paradigm to reduce transmission latency and resources as well as preserve raw data privacy by collaboratively training local deep learning models across multiple edge devices. However, with the development of artificial intelligence (AI) technologies, the size of neural network models grows exponentially with their parameters to meet variable application requirements, which poses significant challenges to the computation, communication, and memory abilities of edge devices. Existing designs typically focus on either communication or computation efficiency without caring each device’s memory ability. To deal with the above issues, we first introduce a novel model sparse update enabled tiny FEL (TinyFEL) architecture, which terminates the backpropagation early in local model training processes. Therefore, the proposed TinyFEL can reduce local memory occupation and lessen the communication-and-computation burden. Furthermore, we propose a parameter splitting mechanism instead of transmitting the full model, only a part of updated layers of parameters is transmitted for aggregation, which significantly reduced the communication overheads. Thereafter, we develop a communication and computation latency minimization problem to accelerate the training of TinyFEL. To this end, we theoretically analyze the convergence performance of TinyFEL, which unveils the mathematical relationship among sparse update ratio assignment, device selection, and learning performance. Then, a joint sparse update ratio assignment, device selection, and resource allocation strategy is introduced based on the alternating direction method of multipliers (ADMMs) and block coordinate descent (BCD) algorithms. Numerical results indicate that our proposed TinyFEL can reduce training memory occupation by over 40% than the traditional FEL at the cost of negligible accuracy loss.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
TinyFEL:通过模型稀疏更新实现通信、计算和内存高效的微型联合边缘学习
联邦边缘学习(FEL)被认为是一种很有前途的分布式机器学习范式,通过跨多个边缘设备协作训练本地深度学习模型来减少传输延迟和资源,并保护原始数据隐私。然而,随着人工智能技术的发展,神经网络模型的规模随着参数的变化呈指数级增长,以满足不同的应用需求,这对边缘设备的计算、通信和存储能力提出了重大挑战。现有的设计通常侧重于通信或计算效率,而不关心每个设备的存储能力。为了解决上述问题,我们首先引入了一种新的支持模型稀疏更新的微小FEL (TinyFEL)架构,该架构可以在局部模型训练过程中提前终止反向传播。因此,提出的TinyFEL可以减少局部内存占用,减轻通信和计算负担。此外,我们提出了一种参数分割机制,而不是传输整个模型,只传输更新后的部分参数层进行聚合,从而大大降低了通信开销。在此基础上,我们提出了通信和计算延迟最小化问题,以加速TinyFEL的训练。为此,我们从理论上分析了TinyFEL的收敛性能,揭示了稀疏更新比分配、器件选择和学习性能之间的数学关系。然后,提出了一种基于乘法器交替方向法(admm)和块坐标下降(BCD)算法的联合稀疏更新比分配、设备选择和资源分配策略。数值结果表明,我们提出的TinyFEL比传统的FEL减少了40%以上的训练内存占用,而精度损失可以忽略不计。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE Internet of Things Journal
IEEE Internet of Things Journal Computer Science-Information Systems
CiteScore
17.60
自引率
13.20%
发文量
1982
期刊介绍: The EEE Internet of Things (IoT) Journal publishes articles and review articles covering various aspects of IoT, including IoT system architecture, IoT enabling technologies, IoT communication and networking protocols such as network coding, and IoT services and applications. Topics encompass IoT's impacts on sensor technologies, big data management, and future internet design for applications like smart cities and smart homes. Fields of interest include IoT architecture such as things-centric, data-centric, service-oriented IoT architecture; IoT enabling technologies and systematic integration such as sensor technologies, big sensor data management, and future Internet design for IoT; IoT services, applications, and test-beds such as IoT service middleware, IoT application programming interface (API), IoT application design, and IoT trials/experiments; IoT standardization activities and technology development in different standard development organizations (SDO) such as IEEE, IETF, ITU, 3GPP, ETSI, etc.
期刊最新文献
Differential STFLC with Resource Grid Mapping for Robust IoT Connectivity in Time-Varying Frequency-Selective Channels ViPSN 2.0: A Reconfigurable Battery-free IoT Platform for Vibration Energy Harvesting Optimizing Resource Utilization and Performance in LEO Satellite Edge Computing: A Joint Service Deployment and Task Offloading Approach A Mobile Aerial Semi-Quantum Communication Protocol with Time-Batched Polarization Encoding for Securing Low-Altitude Networks AI-Enhanced 2D-Material Terahertz Sensor for 6G IoT-Based Smart Agricultural Monitoring
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1