{"title":"非虚拟环境下联邦学习的显著加权聚合方法","authors":"Wei-Jong Yang, P. Chung","doi":"10.1109/IS3C57901.2023.00095","DOIUrl":null,"url":null,"abstract":"Federated learning provides a decentralized learning without data exchange. Among them, the Federated Average (FedAVG) framework is the most likely to be implemented in real world application due to its low communication overhead. However, this architecture can easily affect the efficiency of global model convergence when there are differences data distribution in individual user. Therefore, in this paper, we propose an aggregation strategy called significant Weighted feature aggregation method, in which the features with large variation are appropriately weighted at the server side to improve the model convergence speed even in not identically and independently distributed (non-iid) environments. As shown in our experiments, our approach had over 10% of improvements compared to the FedAVG.","PeriodicalId":142483,"journal":{"name":"2023 Sixth International Symposium on Computer, Consumer and Control (IS3C)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Significant Weighted Aggregation Method for Federated Learning in Non-iid Environment\",\"authors\":\"Wei-Jong Yang, P. Chung\",\"doi\":\"10.1109/IS3C57901.2023.00095\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Federated learning provides a decentralized learning without data exchange. Among them, the Federated Average (FedAVG) framework is the most likely to be implemented in real world application due to its low communication overhead. However, this architecture can easily affect the efficiency of global model convergence when there are differences data distribution in individual user. Therefore, in this paper, we propose an aggregation strategy called significant Weighted feature aggregation method, in which the features with large variation are appropriately weighted at the server side to improve the model convergence speed even in not identically and independently distributed (non-iid) environments. As shown in our experiments, our approach had over 10% of improvements compared to the FedAVG.\",\"PeriodicalId\":142483,\"journal\":{\"name\":\"2023 Sixth International Symposium on Computer, Consumer and Control (IS3C)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-06-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 Sixth International Symposium on Computer, Consumer and Control (IS3C)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IS3C57901.2023.00095\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 Sixth International Symposium on Computer, Consumer and Control (IS3C)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IS3C57901.2023.00095","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Significant Weighted Aggregation Method for Federated Learning in Non-iid Environment
Federated learning provides a decentralized learning without data exchange. Among them, the Federated Average (FedAVG) framework is the most likely to be implemented in real world application due to its low communication overhead. However, this architecture can easily affect the efficiency of global model convergence when there are differences data distribution in individual user. Therefore, in this paper, we propose an aggregation strategy called significant Weighted feature aggregation method, in which the features with large variation are appropriately weighted at the server side to improve the model convergence speed even in not identically and independently distributed (non-iid) environments. As shown in our experiments, our approach had over 10% of improvements compared to the FedAVG.