{"title":"通过自适应聚合权重应对联合学习中的数据异质性变化","authors":"","doi":"10.1016/j.knosys.2024.112484","DOIUrl":null,"url":null,"abstract":"<div><p>In federated learning (FL), ensuring the efficiency of global models generated from the weighted aggregation of local models with data heterogeneity remains challenging. Moreover, the contradiction between imprecise aggregation weights and changing data distributions leads to aggregation errors that increase in an accelerated manner throughout the process. Therefore, we present federated learning using adaptive aggregate weights (FedAAW) to change the optimization direction in steps, including local training and global aggregation, and reduce inefficiencies in the global model due to the accelerated growth of aggregation errors resulting from changes in heterogeneity. In each round, the global- and local-model information is dynamically combined to generate an initial model at the beginning of the local training. The key module in FedAAW is adaptive aggregate weights (AAW), which are used to update the aggregation weight by sharing an optimization objective with global training and using the gradient information from other clients to accurately compute the updated aggregation weight direction. AAW guarantee consistency between weight update and global optimization, theoretically demonstrating convergence. The results of our comprehensive experiments on public datasets demonstrate that the test accuracy metrics of FedAAW are higher than those of six state-of-the-art algorithms and that FedAAW is capable of up to 50% improvement. FedAAW also results in an improvement of 14% on CIFAR100, a complex dataset, when compared with the best-performing baseline. FedAAW is faster than other algorithms in attaining the specified accuracy in experiments; in particular, it is approximately three times faster than federated learning with adaptive local aggregation. In addition, the results obtained in experimental environments with different client sizes and heterogeneous data confirm that FedAAW is robust and scalable.</p></div>","PeriodicalId":49939,"journal":{"name":"Knowledge-Based Systems","volume":null,"pages":null},"PeriodicalIF":7.2000,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Tackling data-heterogeneity variations in federated learning via adaptive aggregate weights\",\"authors\":\"\",\"doi\":\"10.1016/j.knosys.2024.112484\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>In federated learning (FL), ensuring the efficiency of global models generated from the weighted aggregation of local models with data heterogeneity remains challenging. Moreover, the contradiction between imprecise aggregation weights and changing data distributions leads to aggregation errors that increase in an accelerated manner throughout the process. Therefore, we present federated learning using adaptive aggregate weights (FedAAW) to change the optimization direction in steps, including local training and global aggregation, and reduce inefficiencies in the global model due to the accelerated growth of aggregation errors resulting from changes in heterogeneity. In each round, the global- and local-model information is dynamically combined to generate an initial model at the beginning of the local training. The key module in FedAAW is adaptive aggregate weights (AAW), which are used to update the aggregation weight by sharing an optimization objective with global training and using the gradient information from other clients to accurately compute the updated aggregation weight direction. AAW guarantee consistency between weight update and global optimization, theoretically demonstrating convergence. The results of our comprehensive experiments on public datasets demonstrate that the test accuracy metrics of FedAAW are higher than those of six state-of-the-art algorithms and that FedAAW is capable of up to 50% improvement. FedAAW also results in an improvement of 14% on CIFAR100, a complex dataset, when compared with the best-performing baseline. FedAAW is faster than other algorithms in attaining the specified accuracy in experiments; in particular, it is approximately three times faster than federated learning with adaptive local aggregation. In addition, the results obtained in experimental environments with different client sizes and heterogeneous data confirm that FedAAW is robust and scalable.</p></div>\",\"PeriodicalId\":49939,\"journal\":{\"name\":\"Knowledge-Based Systems\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":7.2000,\"publicationDate\":\"2024-09-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Knowledge-Based Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0950705124011183\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Knowledge-Based Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0950705124011183","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Tackling data-heterogeneity variations in federated learning via adaptive aggregate weights
In federated learning (FL), ensuring the efficiency of global models generated from the weighted aggregation of local models with data heterogeneity remains challenging. Moreover, the contradiction between imprecise aggregation weights and changing data distributions leads to aggregation errors that increase in an accelerated manner throughout the process. Therefore, we present federated learning using adaptive aggregate weights (FedAAW) to change the optimization direction in steps, including local training and global aggregation, and reduce inefficiencies in the global model due to the accelerated growth of aggregation errors resulting from changes in heterogeneity. In each round, the global- and local-model information is dynamically combined to generate an initial model at the beginning of the local training. The key module in FedAAW is adaptive aggregate weights (AAW), which are used to update the aggregation weight by sharing an optimization objective with global training and using the gradient information from other clients to accurately compute the updated aggregation weight direction. AAW guarantee consistency between weight update and global optimization, theoretically demonstrating convergence. The results of our comprehensive experiments on public datasets demonstrate that the test accuracy metrics of FedAAW are higher than those of six state-of-the-art algorithms and that FedAAW is capable of up to 50% improvement. FedAAW also results in an improvement of 14% on CIFAR100, a complex dataset, when compared with the best-performing baseline. FedAAW is faster than other algorithms in attaining the specified accuracy in experiments; in particular, it is approximately three times faster than federated learning with adaptive local aggregation. In addition, the results obtained in experimental environments with different client sizes and heterogeneous data confirm that FedAAW is robust and scalable.
期刊介绍:
Knowledge-Based Systems, an international and interdisciplinary journal in artificial intelligence, publishes original, innovative, and creative research results in the field. It focuses on knowledge-based and other artificial intelligence techniques-based systems. The journal aims to support human prediction and decision-making through data science and computation techniques, provide a balanced coverage of theory and practical study, and encourage the development and implementation of knowledge-based intelligence models, methods, systems, and software tools. Applications in business, government, education, engineering, and healthcare are emphasized.