{"title":"Enabling Foundation Models: A Distributed Collaboration Framework Based on Graph Federated Learning","authors":"Jiewei Chen;Shaoyong Guo;Qi Qi;Jiakai Hao;Song Guo;Xuesong Qiu","doi":"10.1109/TSC.2024.3436695","DOIUrl":null,"url":null,"abstract":"Foundation models (FMs), known as pre-trained models, have garnered significant interest in Industrial Internet due to their remarkable performance and robust generalization capabilities in downstream tasks. However, with the increasing requirements of computing infrastructure and data privacy protection for large foundation models, existing learning frameworks face challenges such as data privacy leakage, poor scalability, and deployment difficulties. To address these issues, this paper proposes a novel collaborative Transformer Block (TB)-wise training framework based on Federated Learning (FL), which consists of three stages: pre-training, graph regularization, and personalized training. To tackle the challenge of statistical heterogeneity in distributed data, we design a Graph Convolutional Network (GCN)-based update operator that captures local training representations. Besides, we conduct an analysis based on feature similarity to enhance the interpretability of our algorithm. We choose popular vision Transformer models for the experiments, extensive results demonstrate that our framework can jointly train multiple clients to build a foundation model while improving the single client's personalized performance. The proposed method outperforms state-of-the-art frameworks under various data distributions and system heterogeneity settings, highlighting its robust performance.","PeriodicalId":13255,"journal":{"name":"IEEE Transactions on Services Computing","volume":"17 6","pages":"3569-3582"},"PeriodicalIF":5.8000,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Services Computing","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10620405/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Foundation models (FMs), known as pre-trained models, have garnered significant interest in Industrial Internet due to their remarkable performance and robust generalization capabilities in downstream tasks. However, with the increasing requirements of computing infrastructure and data privacy protection for large foundation models, existing learning frameworks face challenges such as data privacy leakage, poor scalability, and deployment difficulties. To address these issues, this paper proposes a novel collaborative Transformer Block (TB)-wise training framework based on Federated Learning (FL), which consists of three stages: pre-training, graph regularization, and personalized training. To tackle the challenge of statistical heterogeneity in distributed data, we design a Graph Convolutional Network (GCN)-based update operator that captures local training representations. Besides, we conduct an analysis based on feature similarity to enhance the interpretability of our algorithm. We choose popular vision Transformer models for the experiments, extensive results demonstrate that our framework can jointly train multiple clients to build a foundation model while improving the single client's personalized performance. The proposed method outperforms state-of-the-art frameworks under various data distributions and system heterogeneity settings, highlighting its robust performance.
期刊介绍:
IEEE Transactions on Services Computing encompasses the computing and software aspects of the science and technology of services innovation research and development. It places emphasis on algorithmic, mathematical, statistical, and computational methods central to services computing. Topics covered include Service Oriented Architecture, Web Services, Business Process Integration, Solution Performance Management, and Services Operations and Management. The transactions address mathematical foundations, security, privacy, agreement, contract, discovery, negotiation, collaboration, and quality of service for web services. It also covers areas like composite web service creation, business and scientific applications, standards, utility models, business process modeling, integration, collaboration, and more in the realm of Services Computing.