{"title":"Overlay-based Decentralized Federated Learning in Bandwidth-limited Networks","authors":"Yudi Huang, Tingyang Sun, Ting He","doi":"arxiv-2408.04705","DOIUrl":null,"url":null,"abstract":"The emerging machine learning paradigm of decentralized federated learning\n(DFL) has the promise of greatly boosting the deployment of artificial\nintelligence (AI) by directly learning across distributed agents without\ncentralized coordination. Despite significant efforts on improving the\ncommunication efficiency of DFL, most existing solutions were based on the\nsimplistic assumption that neighboring agents are physically adjacent in the\nunderlying communication network, which fails to correctly capture the\ncommunication cost when learning over a general bandwidth-limited network, as\nencountered in many edge networks. In this work, we address this gap by\nleveraging recent advances in network tomography to jointly design the\ncommunication demands and the communication schedule for overlay-based DFL in\nbandwidth-limited networks without requiring explicit cooperation from the\nunderlying network. By carefully analyzing the structure of our problem, we\ndecompose it into a series of optimization problems that can each be solved\nefficiently, to collectively minimize the total training time. Extensive\ndata-driven simulations show that our solution can significantly accelerate DFL\nin comparison with state-of-the-art designs.","PeriodicalId":501280,"journal":{"name":"arXiv - CS - Networking and Internet Architecture","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Networking and Internet Architecture","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.04705","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The emerging machine learning paradigm of decentralized federated learning
(DFL) has the promise of greatly boosting the deployment of artificial
intelligence (AI) by directly learning across distributed agents without
centralized coordination. Despite significant efforts on improving the
communication efficiency of DFL, most existing solutions were based on the
simplistic assumption that neighboring agents are physically adjacent in the
underlying communication network, which fails to correctly capture the
communication cost when learning over a general bandwidth-limited network, as
encountered in many edge networks. In this work, we address this gap by
leveraging recent advances in network tomography to jointly design the
communication demands and the communication schedule for overlay-based DFL in
bandwidth-limited networks without requiring explicit cooperation from the
underlying network. By carefully analyzing the structure of our problem, we
decompose it into a series of optimization problems that can each be solved
efficiently, to collectively minimize the total training time. Extensive
data-driven simulations show that our solution can significantly accelerate DFL
in comparison with state-of-the-art designs.