{"title":"Exploring System-Heterogeneous Federated Learning with Dynamic Model Selection","authors":"Dixi Yao","doi":"arxiv-2409.08858","DOIUrl":null,"url":null,"abstract":"Federated learning is a distributed learning paradigm in which multiple\nmobile clients train a global model while keeping data local. These mobile\nclients can have various available memory and network bandwidth. However, to\nachieve the best global model performance, how we can utilize available memory\nand network bandwidth to the maximum remains an open challenge. In this paper,\nwe propose to assign each client a subset of the global model, having different\nlayers and channels on each layer. To realize that, we design a constrained\nmodel search process with early stop to improve efficiency of finding the\nmodels from such a very large space; and a data-free knowledge distillation\nmechanism to improve the global model performance when aggregating models of\nsuch different structures. For fair and reproducible comparison between\ndifferent solutions, we develop a new system, which can directly allocate\ndifferent memory and bandwidth to each client according to memory and bandwidth\nlogs collected on mobile devices. The evaluation shows that our solution can\nhave accuracy increase ranging from 2.43\\% to 15.81\\% and provide 5\\% to 40\\%\nmore memory and bandwidth utilization with negligible extra running time,\ncomparing to existing state-of-the-art system-heterogeneous federated learning\nmethods under different available memory and bandwidth, non-i.i.d.~datasets,\nimage and text tasks.","PeriodicalId":501422,"journal":{"name":"arXiv - CS - Distributed, Parallel, and Cluster Computing","volume":"20 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Distributed, Parallel, and Cluster Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.08858","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Federated learning is a distributed learning paradigm in which multiple
mobile clients train a global model while keeping data local. These mobile
clients can have various available memory and network bandwidth. However, to
achieve the best global model performance, how we can utilize available memory
and network bandwidth to the maximum remains an open challenge. In this paper,
we propose to assign each client a subset of the global model, having different
layers and channels on each layer. To realize that, we design a constrained
model search process with early stop to improve efficiency of finding the
models from such a very large space; and a data-free knowledge distillation
mechanism to improve the global model performance when aggregating models of
such different structures. For fair and reproducible comparison between
different solutions, we develop a new system, which can directly allocate
different memory and bandwidth to each client according to memory and bandwidth
logs collected on mobile devices. The evaluation shows that our solution can
have accuracy increase ranging from 2.43\% to 15.81\% and provide 5\% to 40\%
more memory and bandwidth utilization with negligible extra running time,
comparing to existing state-of-the-art system-heterogeneous federated learning
methods under different available memory and bandwidth, non-i.i.d.~datasets,
image and text tasks.