{"title":"TinyFEL: Communication, Computation, and Memory Efficient Tiny Federated Edge Learning via Model Sparse Update","authors":"Qimei Chen;Han Cheng;Yipeng Liang;Guangxu Zhu;Ming Li;Hao Jiang","doi":"10.1109/JIOT.2024.3499375","DOIUrl":null,"url":null,"abstract":"Federated edge learning (FEL) is regarded as a promising distributed machine learning paradigm to reduce transmission latency and resources as well as preserve raw data privacy by collaboratively training local deep learning models across multiple edge devices. However, with the development of artificial intelligence (AI) technologies, the size of neural network models grows exponentially with their parameters to meet variable application requirements, which poses significant challenges to the computation, communication, and memory abilities of edge devices. Existing designs typically focus on either communication or computation efficiency without caring each device’s memory ability. To deal with the above issues, we first introduce a novel model sparse update enabled tiny FEL (TinyFEL) architecture, which terminates the backpropagation early in local model training processes. Therefore, the proposed TinyFEL can reduce local memory occupation and lessen the communication-and-computation burden. Furthermore, we propose a parameter splitting mechanism instead of transmitting the full model, only a part of updated layers of parameters is transmitted for aggregation, which significantly reduced the communication overheads. Thereafter, we develop a communication and computation latency minimization problem to accelerate the training of TinyFEL. To this end, we theoretically analyze the convergence performance of TinyFEL, which unveils the mathematical relationship among sparse update ratio assignment, device selection, and learning performance. Then, a joint sparse update ratio assignment, device selection, and resource allocation strategy is introduced based on the alternating direction method of multipliers (ADMMs) and block coordinate descent (BCD) algorithms. Numerical results indicate that our proposed TinyFEL can reduce training memory occupation by over 40% than the traditional FEL at the cost of negligible accuracy loss.","PeriodicalId":54347,"journal":{"name":"IEEE Internet of Things Journal","volume":"12 7","pages":"8247-8260"},"PeriodicalIF":8.9000,"publicationDate":"2024-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Internet of Things Journal","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10753457/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Federated edge learning (FEL) is regarded as a promising distributed machine learning paradigm to reduce transmission latency and resources as well as preserve raw data privacy by collaboratively training local deep learning models across multiple edge devices. However, with the development of artificial intelligence (AI) technologies, the size of neural network models grows exponentially with their parameters to meet variable application requirements, which poses significant challenges to the computation, communication, and memory abilities of edge devices. Existing designs typically focus on either communication or computation efficiency without caring each device’s memory ability. To deal with the above issues, we first introduce a novel model sparse update enabled tiny FEL (TinyFEL) architecture, which terminates the backpropagation early in local model training processes. Therefore, the proposed TinyFEL can reduce local memory occupation and lessen the communication-and-computation burden. Furthermore, we propose a parameter splitting mechanism instead of transmitting the full model, only a part of updated layers of parameters is transmitted for aggregation, which significantly reduced the communication overheads. Thereafter, we develop a communication and computation latency minimization problem to accelerate the training of TinyFEL. To this end, we theoretically analyze the convergence performance of TinyFEL, which unveils the mathematical relationship among sparse update ratio assignment, device selection, and learning performance. Then, a joint sparse update ratio assignment, device selection, and resource allocation strategy is introduced based on the alternating direction method of multipliers (ADMMs) and block coordinate descent (BCD) algorithms. Numerical results indicate that our proposed TinyFEL can reduce training memory occupation by over 40% than the traditional FEL at the cost of negligible accuracy loss.
期刊介绍:
The EEE Internet of Things (IoT) Journal publishes articles and review articles covering various aspects of IoT, including IoT system architecture, IoT enabling technologies, IoT communication and networking protocols such as network coding, and IoT services and applications. Topics encompass IoT's impacts on sensor technologies, big data management, and future internet design for applications like smart cities and smart homes. Fields of interest include IoT architecture such as things-centric, data-centric, service-oriented IoT architecture; IoT enabling technologies and systematic integration such as sensor technologies, big sensor data management, and future Internet design for IoT; IoT services, applications, and test-beds such as IoT service middleware, IoT application programming interface (API), IoT application design, and IoT trials/experiments; IoT standardization activities and technology development in different standard development organizations (SDO) such as IEEE, IETF, ITU, 3GPP, ETSI, etc.