{"title":"FedFT: Improving Communication Performance for Federated Learning with Frequency Space Transformation","authors":"Chamath Palihawadana, Nirmalie Wiratunga, Anjana Wijekoon, Harsha Kalutarage","doi":"arxiv-2409.05242","DOIUrl":null,"url":null,"abstract":"Communication efficiency is a widely recognised research problem in Federated\nLearning (FL), with recent work focused on developing techniques for efficient\ncompression, distribution and aggregation of model parameters between clients\nand the server. Particularly within distributed systems, it is important to\nbalance the need for computational cost and communication efficiency. However,\nexisting methods are often constrained to specific applications and are less\ngeneralisable. In this paper, we introduce FedFT (federated frequency-space\ntransformation), a simple yet effective methodology for communicating model\nparameters in a FL setting. FedFT uses Discrete Cosine Transform (DCT) to\nrepresent model parameters in frequency space, enabling efficient compression\nand reducing communication overhead. FedFT is compatible with various existing\nFL methodologies and neural architectures, and its linear property eliminates\nthe need for multiple transformations during federated aggregation. This\nmethodology is vital for distributed solutions, tackling essential challenges\nlike data privacy, interoperability, and energy efficiency inherent to these\nenvironments. We demonstrate the generalisability of the FedFT methodology on\nfour datasets using comparative studies with three state-of-the-art FL\nbaselines (FedAvg, FedProx, FedSim). Our results demonstrate that using FedFT\nto represent the differences in model parameters between communication rounds\nin frequency space results in a more compact representation compared to\nrepresenting the entire model in frequency space. This leads to a reduction in\ncommunication overhead, while keeping accuracy levels comparable and in some\ncases even improving it. Our results suggest that this reduction can range from\n5% to 30% per client, depending on dataset.","PeriodicalId":501422,"journal":{"name":"arXiv - CS - Distributed, Parallel, and Cluster Computing","volume":"106 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Distributed, Parallel, and Cluster Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.05242","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Communication efficiency is a widely recognised research problem in Federated
Learning (FL), with recent work focused on developing techniques for efficient
compression, distribution and aggregation of model parameters between clients
and the server. Particularly within distributed systems, it is important to
balance the need for computational cost and communication efficiency. However,
existing methods are often constrained to specific applications and are less
generalisable. In this paper, we introduce FedFT (federated frequency-space
transformation), a simple yet effective methodology for communicating model
parameters in a FL setting. FedFT uses Discrete Cosine Transform (DCT) to
represent model parameters in frequency space, enabling efficient compression
and reducing communication overhead. FedFT is compatible with various existing
FL methodologies and neural architectures, and its linear property eliminates
the need for multiple transformations during federated aggregation. This
methodology is vital for distributed solutions, tackling essential challenges
like data privacy, interoperability, and energy efficiency inherent to these
environments. We demonstrate the generalisability of the FedFT methodology on
four datasets using comparative studies with three state-of-the-art FL
baselines (FedAvg, FedProx, FedSim). Our results demonstrate that using FedFT
to represent the differences in model parameters between communication rounds
in frequency space results in a more compact representation compared to
representing the entire model in frequency space. This leads to a reduction in
communication overhead, while keeping accuracy levels comparable and in some
cases even improving it. Our results suggest that this reduction can range from
5% to 30% per client, depending on dataset.