FedCust: Offloading hyperparameter customization for federated learning

IF 1 4区 计算机科学 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Performance Evaluation Pub Date : 2024-11-16 DOI:10.1016/j.peva.2024.102450
Syed Zawad , Xiaolong Ma , Jun Yi , Cheng Li , Minjia Zhang , Lei Yang , Feng Yan , Yuxiong He
{"title":"FedCust: Offloading hyperparameter customization for federated learning","authors":"Syed Zawad ,&nbsp;Xiaolong Ma ,&nbsp;Jun Yi ,&nbsp;Cheng Li ,&nbsp;Minjia Zhang ,&nbsp;Lei Yang ,&nbsp;Feng Yan ,&nbsp;Yuxiong He","doi":"10.1016/j.peva.2024.102450","DOIUrl":null,"url":null,"abstract":"<div><div>Federated Learning (FL) is a new machine learning paradigm that enables training models collaboratively across clients without sharing private data. In FL, data is non-uniformly distributed among clients (i.e., data heterogeneity) and cannot be redistributed nor monitored like in conventional machine learning due to privacy constraints. Such data heterogeneity and privacy requirements bring new challenges for learning hyperparameter optimization as the training dynamics change across clients even within the same training round and they are difficult to be measured due to privacy. The state-of-the-art in hyperparameter customization can greatly improve FL model accuracy but also incur significant computing overheads and power consumption on client devices, and slowdown the training process. To address the prohibitively expensive cost challenge, we explore the possibility of offloading hyperparameter customization to servers. We propose <em>FedCust</em>, a framework that offloads expensive hyperparameter customization cost from the client devices to the central server without violating privacy constraints. Our key discovery is that it is not necessary to do hyperparameter customization for every client, and clients with similar data heterogeneity can use the same hyperparameters to achieve good training performance. We propose heterogeneity measurement metrics for clustering clients into groups such that clients within the same group share hyperparameters. <em>FedCust</em> uses the proxy data from initial model design to emulate different heterogeneity groups and perform hyperparameter customization on the server side without accessing client data nor information. To make the hyperparameter customization scalable, <em>FedCust</em> further employs a Bayesian-strengthened tuner to significantly accelerates the hyperparameter customization speed. Extensive evaluation demonstrates that <em>FedCust</em> achieves up to 7/2/4/4/6% better accuracy than the widely adopted one-size-fits-all approach on popular FL benchmarks FEMNIST, Shakespeare, Cifar100, Cifar10, and Fashion-MNIST respectively, while being scalable and reducing computation, memory, and energy consumption on the client devices, without compromising privacy constraints.</div></div>","PeriodicalId":19964,"journal":{"name":"Performance Evaluation","volume":"167 ","pages":"Article 102450"},"PeriodicalIF":1.0000,"publicationDate":"2024-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Performance Evaluation","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0166531624000555","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
引用次数: 0

Abstract

Federated Learning (FL) is a new machine learning paradigm that enables training models collaboratively across clients without sharing private data. In FL, data is non-uniformly distributed among clients (i.e., data heterogeneity) and cannot be redistributed nor monitored like in conventional machine learning due to privacy constraints. Such data heterogeneity and privacy requirements bring new challenges for learning hyperparameter optimization as the training dynamics change across clients even within the same training round and they are difficult to be measured due to privacy. The state-of-the-art in hyperparameter customization can greatly improve FL model accuracy but also incur significant computing overheads and power consumption on client devices, and slowdown the training process. To address the prohibitively expensive cost challenge, we explore the possibility of offloading hyperparameter customization to servers. We propose FedCust, a framework that offloads expensive hyperparameter customization cost from the client devices to the central server without violating privacy constraints. Our key discovery is that it is not necessary to do hyperparameter customization for every client, and clients with similar data heterogeneity can use the same hyperparameters to achieve good training performance. We propose heterogeneity measurement metrics for clustering clients into groups such that clients within the same group share hyperparameters. FedCust uses the proxy data from initial model design to emulate different heterogeneity groups and perform hyperparameter customization on the server side without accessing client data nor information. To make the hyperparameter customization scalable, FedCust further employs a Bayesian-strengthened tuner to significantly accelerates the hyperparameter customization speed. Extensive evaluation demonstrates that FedCust achieves up to 7/2/4/4/6% better accuracy than the widely adopted one-size-fits-all approach on popular FL benchmarks FEMNIST, Shakespeare, Cifar100, Cifar10, and Fashion-MNIST respectively, while being scalable and reducing computation, memory, and energy consumption on the client devices, without compromising privacy constraints.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
FedCust:为联合学习卸载超参数定制功能
联合学习(FL)是一种新的机器学习范式,它能在不共享隐私数据的情况下跨客户端协作训练模型。在联机学习中,数据在客户端之间是非均匀分布的(即数据异构性),由于隐私限制,不能像传统机器学习那样进行再分配或监控。这种数据异质性和隐私要求给学习超参数优化带来了新的挑战,因为即使在同一轮训练中,不同客户的训练动态也会发生变化,而且由于隐私原因,很难对其进行测量。最先进的超参数定制技术可以大大提高 FL 模型的准确性,但同时也会在客户端设备上产生巨大的计算开销和功耗,并减慢训练过程。为了解决成本过高的难题,我们探索了将超参数定制卸载到服务器上的可能性。我们提出了 FedCust,这是一个在不违反隐私约束的情况下将昂贵的超参数定制成本从客户端设备卸载到中央服务器的框架。我们的主要发现是,没有必要为每个客户端进行超参数定制,具有相似数据异质性的客户端可以使用相同的超参数来实现良好的训练性能。我们提出了异质性测量指标,用于将客户机聚类成组,使同组内的客户机共享超参数。FedCust 使用初始模型设计中的代理数据来模拟不同的异质性组,并在服务器端执行超参数定制,而无需访问客户端数据或信息。为了使超参数定制具有可扩展性,FedCust 进一步采用了贝叶斯强化调谐器,显著加快了超参数定制速度。广泛的评估表明,在流行的 FL 基准 FEMNIST、Shakespeare、Cifar100、Cifar10 和 Fashion-MNIST 上,FedCust 比广泛采用的 "一刀切 "方法分别提高了高达 7/2/4/4/6%的准确率,同时还具有可扩展性,降低了客户端设备的计算量、内存和能耗,而且不影响隐私约束。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Performance Evaluation
Performance Evaluation 工程技术-计算机:理论方法
CiteScore
3.10
自引率
0.00%
发文量
20
审稿时长
24 days
期刊介绍: Performance Evaluation functions as a leading journal in the area of modeling, measurement, and evaluation of performance aspects of computing and communication systems. As such, it aims to present a balanced and complete view of the entire Performance Evaluation profession. Hence, the journal is interested in papers that focus on one or more of the following dimensions: -Define new performance evaluation tools, including measurement and monitoring tools as well as modeling and analytic techniques -Provide new insights into the performance of computing and communication systems -Introduce new application areas where performance evaluation tools can play an important role and creative new uses for performance evaluation tools. More specifically, common application areas of interest include the performance of: -Resource allocation and control methods and algorithms (e.g. routing and flow control in networks, bandwidth allocation, processor scheduling, memory management) -System architecture, design and implementation -Cognitive radio -VANETs -Social networks and media -Energy efficient ICT -Energy harvesting -Data centers -Data centric networks -System reliability -System tuning and capacity planning -Wireless and sensor networks -Autonomic and self-organizing systems -Embedded systems -Network science
期刊最新文献
Analysis of a queue-length-dependent vacation queue with bulk service, N-policy, set-up time and cost optimization FedCust: Offloading hyperparameter customization for federated learning Trust your local scaler: A continuous, decentralized approach to autoscaling Enabling scalable and adaptive machine learning training via serverless computing on public cloud Symbolic state-space exploration meets statistical model checking
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1