Pub Date : 2024-03-22DOI: 10.1109/TSUSC.2024.3395350
Na Hu;Wei Liang;Dafang Zhang;Kun Xie;Kuanching Li;Albert Y. Zomaya
Traffic prediction is crucial for intelligent transportation systems, assisting in making travel decisions, minimizing traffic congestion, and improving traffic operation efficiency. Although effective, existing centralized traffic prediction methods have privacy leakage risks. Federated learning-based traffic prediction methods keep raw data local and train the global model in a distributed way, thus preserving data privacy. Nevertheless, the spatial correlations between local clients will be broken as data exchange between local clients is not allowed in federated learning, leading to missing spatial information and inferior prediction accuracy. To this end, we propose a federated graph neural network with spatial information completion (FedGCN) for privacy-preserving traffic prediction by adopting a federated learning scheme to protect confidentiality and presenting a mending graph convolutional neural network to mend the missing spatial information during capturing spatial dependency to improve prediction accuracy. To complete the missing spatial information efficiently and capture the client-specific spatial pattern, we design a personalized training scheme for the mending graph neural network, reducing communication overhead. The experiments on four public traffic datasets demonstrate that the proposed model outperforms the best baseline with a ratio of 3.82%, 1.82%, 2.13%, and 1.49% in terms of absolute mean error while preserving privacy.
{"title":"FedGCN: A Federated Graph Convolutional Network for Privacy-Preserving Traffic Prediction","authors":"Na Hu;Wei Liang;Dafang Zhang;Kun Xie;Kuanching Li;Albert Y. Zomaya","doi":"10.1109/TSUSC.2024.3395350","DOIUrl":"https://doi.org/10.1109/TSUSC.2024.3395350","url":null,"abstract":"Traffic prediction is crucial for intelligent transportation systems, assisting in making travel decisions, minimizing traffic congestion, and improving traffic operation efficiency. Although effective, existing centralized traffic prediction methods have privacy leakage risks. Federated learning-based traffic prediction methods keep raw data local and train the global model in a distributed way, thus preserving data privacy. Nevertheless, the spatial correlations between local clients will be broken as data exchange between local clients is not allowed in federated learning, leading to missing spatial information and inferior prediction accuracy. To this end, we propose a federated graph neural network with spatial information completion (FedGCN) for privacy-preserving traffic prediction by adopting a federated learning scheme to protect confidentiality and presenting a mending graph convolutional neural network to mend the missing spatial information during capturing spatial dependency to improve prediction accuracy. To complete the missing spatial information efficiently and capture the client-specific spatial pattern, we design a personalized training scheme for the mending graph neural network, reducing communication overhead. The experiments on four public traffic datasets demonstrate that the proposed model outperforms the best baseline with a ratio of 3.82%, 1.82%, 2.13%, and 1.49% in terms of absolute mean error while preserving privacy.","PeriodicalId":13268,"journal":{"name":"IEEE Transactions on Sustainable Computing","volume":"9 6","pages":"925-935"},"PeriodicalIF":3.0,"publicationDate":"2024-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142810511","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Federated Learning (FL), as a distributed machine learning technique, has promise for training models with distributed data in Artificial Intelligence of Things (AIoT). However, FL is vulnerable to Byzantine attacks from diverse participants. While numerous Byzantine-robust FL solutions have been proposed, most of them involve deploying defenses at either the aggregation server or the participant level, significantly impacting the original FL process. Moreover, it will bring extra computational burden to the server or the participant, which is especially unsuitable for the resource-constrained AIoT domain. To resolve the aforementioned concerns, we propose FL-Auditor, a Byzantine-robust FL approach based on public auditing. Its core idea is to use a Third-Party Auditor (TPA) to audit samples from the FL training process, analyzing the trustworthiness of different participants, thereby helping FL obtain a more robust global model. In addition, we also design a lazy update mechanism to reduce the negative impact of sampling audit on the performance of the global model. Extensive experiments have demonstrated the effectiveness of our FL-Auditor in terms of accuracy, robustness against attacks, and flexibility. In particular, compared to the existing method, our FL-Auditor significantly reduces the computation time on the aggregation server by 8×-17×.
{"title":"Using Third-Party Auditor to Help Federated Learning: An Efficient Byzantine-Robust Federated Learning","authors":"Zhuangzhuang Zhang;Libing Wu;Debiao He;Jianxin Li;Na Lu;Xuejiang Wei","doi":"10.1109/TSUSC.2024.3379440","DOIUrl":"https://doi.org/10.1109/TSUSC.2024.3379440","url":null,"abstract":"Federated Learning (FL), as a distributed machine learning technique, has promise for training models with distributed data in Artificial Intelligence of Things (AIoT). However, FL is vulnerable to Byzantine attacks from diverse participants. While numerous Byzantine-robust FL solutions have been proposed, most of them involve deploying defenses at either the aggregation server or the participant level, significantly impacting the original FL process. Moreover, it will bring extra computational burden to the server or the participant, which is especially unsuitable for the resource-constrained AIoT domain. To resolve the aforementioned concerns, we propose FL-Auditor, a Byzantine-robust FL approach based on public auditing. Its core idea is to use a Third-Party Auditor (TPA) to audit samples from the FL training process, analyzing the trustworthiness of different participants, thereby helping FL obtain a more robust global model. In addition, we also design a lazy update mechanism to reduce the negative impact of sampling audit on the performance of the global model. Extensive experiments have demonstrated the effectiveness of our FL-Auditor in terms of accuracy, robustness against attacks, and flexibility. In particular, compared to the existing method, our FL-Auditor significantly reduces the computation time on the aggregation server by 8×-17×.","PeriodicalId":13268,"journal":{"name":"IEEE Transactions on Sustainable Computing","volume":"9 6","pages":"848-861"},"PeriodicalIF":3.0,"publicationDate":"2024-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142810544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Personalized federated learning (PFL) combines client needs and data characteristics to train personalized models for local clients. However, the most of previous PFL schemes encountered challenges such as low model prediction accuracy and privacy leakage when applied to practical datasets. Besides, the existing privacy protection methods fail to achieve satisfactory results in terms of model prediction accuracy and security simultaneously. In this paper, we propose a Privacy-preserving Personalized Federated Learning under Secure Multi-party Computation (SMC-PPFL), which can preserve privacy while obtaining a local personalized model with high prediction accuracy. In SMC-PPFL, noise perturbation is utilized to protect similarity computation, and secure multi-party computation is employed for model sub-aggregations. This combination ensures that clients’ privacy is preserved, and the computed values remain unbiased without compromising security. Then, we propose a weighted sub-aggregation strategy based on the similarity of clients and introduce a regularization term in the local training to improve prediction accuracy. Finally, we evaluate the performance of SMC-PPFL on three common datasets. The experimental results show that SMC-PPFL achieves $2%!sim! 15%$ higher prediction accuracy compared to the previous PFL schemes. Besides, the security analysis also verifies that SMC-PPFL can resist model inversion attacks and membership inference attacks.
{"title":"Secure and Accurate Personalized Federated Learning With Similarity-Based Model Aggregation","authors":"Zhouyong Tan;Junqing Le;Fan Yang;Min Huang;Tao Xiang;Xiaofeng Liao","doi":"10.1109/TSUSC.2024.3403427","DOIUrl":"https://doi.org/10.1109/TSUSC.2024.3403427","url":null,"abstract":"Personalized federated learning (PFL) combines client needs and data characteristics to train personalized models for local clients. However, the most of previous PFL schemes encountered challenges such as low model prediction accuracy and privacy leakage when applied to practical datasets. Besides, the existing privacy protection methods fail to achieve satisfactory results in terms of model prediction accuracy and security simultaneously. In this paper, we propose a <u>P</u>rivacy-preserving <u>P</u>ersonalized <u>F</u>ederated <u>L</u>earning under <u>S</u>ecure <u>M</u>ulti-party <u>C</u>omputation (SMC-PPFL), which can preserve privacy while obtaining a local personalized model with high prediction accuracy. In SMC-PPFL, noise perturbation is utilized to protect similarity computation, and secure multi-party computation is employed for model sub-aggregations. This combination ensures that clients’ privacy is preserved, and the computed values remain unbiased without compromising security. Then, we propose a weighted sub-aggregation strategy based on the similarity of clients and introduce a regularization term in the local training to improve prediction accuracy. Finally, we evaluate the performance of SMC-PPFL on three common datasets. The experimental results show that SMC-PPFL achieves <inline-formula><tex-math>$2%!sim! 15%$</tex-math></inline-formula> higher prediction accuracy compared to the previous PFL schemes. Besides, the security analysis also verifies that SMC-PPFL can resist model inversion attacks and membership inference attacks.","PeriodicalId":13268,"journal":{"name":"IEEE Transactions on Sustainable Computing","volume":"10 1","pages":"132-145"},"PeriodicalIF":3.0,"publicationDate":"2024-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143184036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-20DOI: 10.1109/TSUSC.2024.3379550
Duc Van Le;Jing Zhou;Rongrong Wang;Rui Tan;Fei Duan
Data centers (DCs) are power-intensive facilities which use a significant amount of energy for cooling the servers. Increasing the temperature and relative humidity (RH) setpoints is a rule-of-thumb approach to reducing the DC energy usage. However, the high temperature and RH may undermine the server's reliability. Before we can choose the proper temperature and RH settings, it is essential to understand how the temperature and RH setpoints affect the DC power usage and server's reliability. To this end, we constructed and experimented with an air-cooled DC testbed in Singapore, which consists of a direct expansion cooling system and 521 servers running real-world application workloads. This paper presents the key measurement results and observations from our 11-month experiments. Our results suggest that by operating at a supply air temperature setpoints of 29 $^{circ }$