Jaewon Yun;Yongjeong Oh;Yo-Seb Jeon;H. Vincent Poor
{"title":"Communication-Efficient Federated Learning Over Capacity-Limited Wireless Networks","authors":"Jaewon Yun;Yongjeong Oh;Yo-Seb Jeon;H. Vincent Poor","doi":"10.1109/TCCN.2024.3419039","DOIUrl":null,"url":null,"abstract":"In this paper, we propose a communication-efficient federated learning (FL) framework to enhance the convergence rate of FL under limited uplink capacity. The core idea of our framework is to transmit the values and positions of the top-S entries of a local model update, determined in terms of magnitude. When transmitting the top-S values, we first apply a linear transformation that enforces the transformed values to behave like Gaussian random variables. We then employ a scalar quantizer optimized for Gaussian distributions, leading to minimizing compression errors. When reconstructing the top-S values, we develop a linear minimum mean squared error method based on the Bussgang decomposition. Additionally, we introduce an error feedback strategy to compensate for both compression and reconstruction errors. We analyze the convergence rate of our framework under general considerations, including a non-convex loss function. Based on our analytical results, we optimize the key parameters of our framework to maximize the convergence rate for a given uplink capacity. Simulation results demonstrate that our framework achieves more than a 2.2%, 1.1%, and 1.4% increase in classification accuracy for the MNIST, CIFAR-10, and CIFAR-100 datasets, respectively, compared to state-of-the-art FL frameworks under limited uplink capacity.","PeriodicalId":13069,"journal":{"name":"IEEE Transactions on Cognitive Communications and Networking","volume":"11 1","pages":"621-637"},"PeriodicalIF":7.0000,"publicationDate":"2024-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Cognitive Communications and Networking","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10571556/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"TELECOMMUNICATIONS","Score":null,"Total":0}
引用次数: 0
Abstract
In this paper, we propose a communication-efficient federated learning (FL) framework to enhance the convergence rate of FL under limited uplink capacity. The core idea of our framework is to transmit the values and positions of the top-S entries of a local model update, determined in terms of magnitude. When transmitting the top-S values, we first apply a linear transformation that enforces the transformed values to behave like Gaussian random variables. We then employ a scalar quantizer optimized for Gaussian distributions, leading to minimizing compression errors. When reconstructing the top-S values, we develop a linear minimum mean squared error method based on the Bussgang decomposition. Additionally, we introduce an error feedback strategy to compensate for both compression and reconstruction errors. We analyze the convergence rate of our framework under general considerations, including a non-convex loss function. Based on our analytical results, we optimize the key parameters of our framework to maximize the convergence rate for a given uplink capacity. Simulation results demonstrate that our framework achieves more than a 2.2%, 1.1%, and 1.4% increase in classification accuracy for the MNIST, CIFAR-10, and CIFAR-100 datasets, respectively, compared to state-of-the-art FL frameworks under limited uplink capacity.
期刊介绍:
The IEEE Transactions on Cognitive Communications and Networking (TCCN) aims to publish high-quality manuscripts that push the boundaries of cognitive communications and networking research. Cognitive, in this context, refers to the application of perception, learning, reasoning, memory, and adaptive approaches in communication system design. The transactions welcome submissions that explore various aspects of cognitive communications and networks, focusing on innovative and holistic approaches to complex system design. Key topics covered include architecture, protocols, cross-layer design, and cognition cycle design for cognitive networks. Additionally, research on machine learning, artificial intelligence, end-to-end and distributed intelligence, software-defined networking, cognitive radios, spectrum sharing, and security and privacy issues in cognitive networks are of interest. The publication also encourages papers addressing novel services and applications enabled by these cognitive concepts.