Communication Efficient Federated Learning for Multi-Organ Segmentation via Knowledge Distillation With Image Synthesis

Soopil Kim;Heejung Park;Philip Chikontwe;Myeongkyun Kang;Kyong Hwan Jin;Ehsan Adeli;Kilian M. Pohl;Sang Hyun Park
{"title":"Communication Efficient Federated Learning for Multi-Organ Segmentation via Knowledge Distillation With Image Synthesis","authors":"Soopil Kim;Heejung Park;Philip Chikontwe;Myeongkyun Kang;Kyong Hwan Jin;Ehsan Adeli;Kilian M. Pohl;Sang Hyun Park","doi":"10.1109/TMI.2025.3525581","DOIUrl":null,"url":null,"abstract":"Federated learning (FL) methods for multi-organ segmentation in CT scans are gaining popularity, but generally require numerous rounds of parameter exchange between a central server and clients. This repetitive sharing of parameters between server and clients may not be practical due to the varying network infrastructures of clients and the large transmission of data. Further increasing repetitive sharing results from data heterogeneity among clients, i.e., clients may differ with respect to the type of data they share. For example, they might provide label maps of different organs (i.e. partial labels) as segmentations of all organs shown in the CT are not part of their clinical protocol. To this end, we propose an efficient communication approach for FL with partial labels. Specifically, parameters of local models are transmitted once to a central server and the global model is trained via knowledge distillation (KD) of the local models. While one can make use of unlabeled public data as inputs for KD, the model accuracy is often limited due to distribution shifts between local and public datasets. Herein, we propose to generate synthetic images from clients’ models as additional inputs to mitigate data shifts between public and local data. In addition, our proposed method offers flexibility for additional finetuning through several rounds of communication using existing FL algorithms, leading to enhanced performance. Extensive evaluation on public datasets in few communication FL scenario reveals that our approach substantially improves over state-of-the-art methods.","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"44 5","pages":"2079-2092"},"PeriodicalIF":0.0000,"publicationDate":"2025-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on medical imaging","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10829700/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Federated learning (FL) methods for multi-organ segmentation in CT scans are gaining popularity, but generally require numerous rounds of parameter exchange between a central server and clients. This repetitive sharing of parameters between server and clients may not be practical due to the varying network infrastructures of clients and the large transmission of data. Further increasing repetitive sharing results from data heterogeneity among clients, i.e., clients may differ with respect to the type of data they share. For example, they might provide label maps of different organs (i.e. partial labels) as segmentations of all organs shown in the CT are not part of their clinical protocol. To this end, we propose an efficient communication approach for FL with partial labels. Specifically, parameters of local models are transmitted once to a central server and the global model is trained via knowledge distillation (KD) of the local models. While one can make use of unlabeled public data as inputs for KD, the model accuracy is often limited due to distribution shifts between local and public datasets. Herein, we propose to generate synthetic images from clients’ models as additional inputs to mitigate data shifts between public and local data. In addition, our proposed method offers flexibility for additional finetuning through several rounds of communication using existing FL algorithms, leading to enhanced performance. Extensive evaluation on public datasets in few communication FL scenario reveals that our approach substantially improves over state-of-the-art methods.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于知识蒸馏和图像合成的通信高效联邦学习多器官分割
用于CT扫描中多器官分割的联邦学习(FL)方法越来越受欢迎,但通常需要在中央服务器和客户端之间进行多轮参数交换。由于客户机的网络基础结构不同以及数据的大量传输,这种在服务器和客户机之间重复共享参数的方法可能不实用。客户机之间的数据异构会导致重复共享的进一步增加,也就是说,客户机共享的数据类型可能不同。例如,他们可能会提供不同器官的标签图(即部分标签),因为CT显示的所有器官的分割不是他们临床方案的一部分。为此,我们提出了一种具有部分标签的FL的有效通信方法。具体来说,局部模型的参数被一次传送到中央服务器,然后通过局部模型的知识蒸馏(KD)来训练全局模型。虽然可以使用未标记的公共数据作为KD的输入,但由于本地和公共数据集之间的分布变化,模型精度通常受到限制。在此,我们建议从客户的模型中生成合成图像,作为额外的输入,以减轻公共数据和本地数据之间的数据转换。此外,我们提出的方法通过使用现有FL算法的几轮通信提供了额外微调的灵活性,从而提高了性能。在少数通信FL场景中对公共数据集的广泛评估表明,我们的方法大大改进了最先进的方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
A 3D Cross-modal Keypoint Descriptor for MR-US Matching and Registration. FairFedMed: Benchmarking Group Fairness in Federated Medical Imaging With FairLoRA. MIRROR: Multi-Modal Pathological Self-Supervised Representation Learning via Modality Alignment and Retention Extractive Radiology Reporting With Memory-Based Cross-Modal Representations Wave-Aware Weakly Supervised Histopathological Tissue Segmentation With Cross-Scale Logits Distillation
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1