{"title":"POLARIS:利用客户端选择加速异步联合学习","authors":"Yufei Kang;Baochun Li","doi":"10.1109/TCC.2024.3370688","DOIUrl":null,"url":null,"abstract":"Federated learning has garnered significant research attention as a privacy-preserving learning paradigm. Asynchronous federated learning has been proposed to improve scalability by accommodating slower clients, commonly referred to as stragglers. However, asynchronous federated learning suffers from slow convergence with respect to wall-clock time, due to the existence of data heterogeneity and staleness. Existing strategies struggled to tackle both difficulties for a wide range of deep learning models. To address the problem, we propose \n<italic>Polaris</i>\n, a theoretically sound design and a new take to client selection for asynchronous federated learning. With \n<italic>Polaris</i>\n, we first theoretically investigated the design space of client sampling strategies from a geometric optimization perspective, taking both data heterogeneity and staleness into account. Our design is not only theoretically proven, but also thoroughly tested in our reproducible experimental open-source testbed. Our experimental results demonstrates overwhelming evidence that \n<italic>Polaris</i>\n outperformed existing state-of-the-art client selection strategies by a substantial margin over a wide variety of tasks and datasets, as we train image classification models using \n<monospace>CIFAR-10</monospace>\n, \n<monospace>CIFAR-100</monospace>\n, \n<monospace>CINIC-10</monospace>\n, \n<monospace>Federated EMNIST</monospace>\n, and a language modeling model using the \n<monospace>Tiny Shakespeare</monospace>\n dataset. Further, our extensive array of ablation studies have also shown that \n<italic>Polaris</i>\n is both scalable and robust as the size of datasets scale up and data heterogeneity vary.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"12 2","pages":"446-458"},"PeriodicalIF":5.3000,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Polaris: Accelerating Asynchronous Federated Learning With Client Selection\",\"authors\":\"Yufei Kang;Baochun Li\",\"doi\":\"10.1109/TCC.2024.3370688\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Federated learning has garnered significant research attention as a privacy-preserving learning paradigm. Asynchronous federated learning has been proposed to improve scalability by accommodating slower clients, commonly referred to as stragglers. However, asynchronous federated learning suffers from slow convergence with respect to wall-clock time, due to the existence of data heterogeneity and staleness. Existing strategies struggled to tackle both difficulties for a wide range of deep learning models. To address the problem, we propose \\n<italic>Polaris</i>\\n, a theoretically sound design and a new take to client selection for asynchronous federated learning. With \\n<italic>Polaris</i>\\n, we first theoretically investigated the design space of client sampling strategies from a geometric optimization perspective, taking both data heterogeneity and staleness into account. Our design is not only theoretically proven, but also thoroughly tested in our reproducible experimental open-source testbed. Our experimental results demonstrates overwhelming evidence that \\n<italic>Polaris</i>\\n outperformed existing state-of-the-art client selection strategies by a substantial margin over a wide variety of tasks and datasets, as we train image classification models using \\n<monospace>CIFAR-10</monospace>\\n, \\n<monospace>CIFAR-100</monospace>\\n, \\n<monospace>CINIC-10</monospace>\\n, \\n<monospace>Federated EMNIST</monospace>\\n, and a language modeling model using the \\n<monospace>Tiny Shakespeare</monospace>\\n dataset. Further, our extensive array of ablation studies have also shown that \\n<italic>Polaris</i>\\n is both scalable and robust as the size of datasets scale up and data heterogeneity vary.\",\"PeriodicalId\":13202,\"journal\":{\"name\":\"IEEE Transactions on Cloud Computing\",\"volume\":\"12 2\",\"pages\":\"446-458\"},\"PeriodicalIF\":5.3000,\"publicationDate\":\"2024-02-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Cloud Computing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10452821/\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Cloud Computing","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10452821/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
Polaris: Accelerating Asynchronous Federated Learning With Client Selection
Federated learning has garnered significant research attention as a privacy-preserving learning paradigm. Asynchronous federated learning has been proposed to improve scalability by accommodating slower clients, commonly referred to as stragglers. However, asynchronous federated learning suffers from slow convergence with respect to wall-clock time, due to the existence of data heterogeneity and staleness. Existing strategies struggled to tackle both difficulties for a wide range of deep learning models. To address the problem, we propose
Polaris
, a theoretically sound design and a new take to client selection for asynchronous federated learning. With
Polaris
, we first theoretically investigated the design space of client sampling strategies from a geometric optimization perspective, taking both data heterogeneity and staleness into account. Our design is not only theoretically proven, but also thoroughly tested in our reproducible experimental open-source testbed. Our experimental results demonstrates overwhelming evidence that
Polaris
outperformed existing state-of-the-art client selection strategies by a substantial margin over a wide variety of tasks and datasets, as we train image classification models using
CIFAR-10
,
CIFAR-100
,
CINIC-10
,
Federated EMNIST
, and a language modeling model using the
Tiny Shakespeare
dataset. Further, our extensive array of ablation studies have also shown that
Polaris
is both scalable and robust as the size of datasets scale up and data heterogeneity vary.
期刊介绍:
The IEEE Transactions on Cloud Computing (TCC) is dedicated to the multidisciplinary field of cloud computing. It is committed to the publication of articles that present innovative research ideas, application results, and case studies in cloud computing, focusing on key technical issues related to theory, algorithms, systems, applications, and performance.