On the Fundamental Limit of Distributed Learning With Interchangable Constrained Statistics

Xinyi Tong;Jian Xu;Shao-Lun Huang
{"title":"On the Fundamental Limit of Distributed Learning With Interchangable Constrained Statistics","authors":"Xinyi Tong;Jian Xu;Shao-Lun Huang","doi":"10.1109/JSAIT.2024.3409426","DOIUrl":null,"url":null,"abstract":"In the popular federated learning scenarios, distributed nodes often represent and exchange information through functions or statistics of data, with communicative processes constrained by the dimensionality of transmitted information. This paper investigates the fundamental limits of distributed parameter estimation and model training problems under such constraints. Specifically, we assume that each node can observe a sequence of i.i.d. sampled data and communicate statistics of the observed data with dimensionality constraints. We first show the Cramer-Rao lower bound (CRLB) and the corresponding achievable estimators for the distributed parameter estimation problems, and the geometric insights and the computable algorithms of designing efficient estimators are also presented. Moreover, we consider model parameters training for distributed nodes with limited communicable statistics. We demonstrate that in order to optimize the excess risk, the feature functions of the statistics shall be designed along the largest eigenvectors of a matrix induced by the model training loss function. In summary, our results potentially provide theoretical guidelines of designing efficient algorithms for enhancing the performance of distributed learning systems.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"396-406"},"PeriodicalIF":0.0000,"publicationDate":"2024-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE journal on selected areas in information theory","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10547488/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In the popular federated learning scenarios, distributed nodes often represent and exchange information through functions or statistics of data, with communicative processes constrained by the dimensionality of transmitted information. This paper investigates the fundamental limits of distributed parameter estimation and model training problems under such constraints. Specifically, we assume that each node can observe a sequence of i.i.d. sampled data and communicate statistics of the observed data with dimensionality constraints. We first show the Cramer-Rao lower bound (CRLB) and the corresponding achievable estimators for the distributed parameter estimation problems, and the geometric insights and the computable algorithms of designing efficient estimators are also presented. Moreover, we consider model parameters training for distributed nodes with limited communicable statistics. We demonstrate that in order to optimize the excess risk, the feature functions of the statistics shall be designed along the largest eigenvectors of a matrix induced by the model training loss function. In summary, our results potentially provide theoretical guidelines of designing efficient algorithms for enhancing the performance of distributed learning systems.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
论可互换约束统计分布式学习的基本极限
在流行的联合学习场景中,分布式节点通常通过函数或数据统计来表示和交换信息,通信过程受到传输信息维度的限制。本文研究了这种约束条件下分布式参数估计和模型训练问题的基本限制。具体来说,我们假设每个节点都能观察到 i.i.d. 采样数据序列,并在维度限制下交流观察到的数据统计量。我们首先展示了分布式参数估计问题的 Cramer-Rao 下界(CRLB)和相应的可实现估计器,还介绍了设计高效估计器的几何见解和可计算算法。此外,我们还考虑了具有有限可通信统计量的分布式节点的模型参数训练问题。我们证明,为了优化超额风险,应沿着模型训练损失函数诱导的矩阵的最大特征向量设计统计特征函数。总之,我们的研究结果为设计高效算法以提高分布式学习系统的性能提供了理论指导。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
8.20
自引率
0.00%
发文量
0
期刊最新文献
Source Coding for Markov Sources With Partial Memoryless Side Information at the Decoder Deviation From Maximal Entanglement for Mid-Spectrum Eigenstates of Local Hamiltonians Statistical Inference With Limited Memory: A Survey Tightening Continuity Bounds for Entropies and Bounds on Quantum Capacities Dynamic Group Testing to Control and Monitor Disease Progression in a Population
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1