首页 > 最新文献

IEEE journal on selected areas in information theory最新文献

英文 中文
Throughput and Latency Analysis for Line Networks With Outage Links 带中断链路的线路网络吞吐量和延迟分析
Pub Date : 2024-06-25 DOI: 10.1109/JSAIT.2024.3419054
Yanyan Dong;Shenghao Yang;Jie Wang;Fan Cheng
Wireless communication links suffer from outage events caused by fading and interference. To facilitate a tractable analysis of network communication throughput and latency, we propose an outage link model to represent a communication link in the slow fading phenomenon. For a line-topology network with outage links, we study three types of intermediate network node schemes: random linear network coding, store-and-forward, and hop-by-hop retransmission. We provide the analytical formulas for the maximum throughputs and the end-to-end latency for each scheme. To gain a more explicit understanding, we perform a scalability analysis of the throughput and latency as the network length increases. We observe that the same order of throughput/latency holds across a wide range of outage functions for each scheme. We illustrate how our exact formulae and scalability results can be applied to compare different schemes.
无线通信链路存在由衰落和干扰引起的中断事件。为了便于分析网络通信的吞吐量和延迟,我们提出了一个中断链路模型来表示慢衰落现象中的通信链路。对于具有中断链路的线路拓扑网络,我们研究了三种中间网络节点方案:随机线性网络编码、存储转发和逐跳重传。我们提供了每种方案的最大吞吐量和端到端延迟的解析公式。为了获得更清晰的理解,我们对网络长度增加时的吞吐量和延迟进行了可扩展性分析。我们发现,每种方案的吞吐量/延迟在各种中断函数中都保持相同的顺序。我们将说明如何应用我们的精确公式和可扩展性结果来比较不同的方案。
{"title":"Throughput and Latency Analysis for Line Networks With Outage Links","authors":"Yanyan Dong;Shenghao Yang;Jie Wang;Fan Cheng","doi":"10.1109/JSAIT.2024.3419054","DOIUrl":"https://doi.org/10.1109/JSAIT.2024.3419054","url":null,"abstract":"Wireless communication links suffer from outage events caused by fading and interference. To facilitate a tractable analysis of network communication throughput and latency, we propose an outage link model to represent a communication link in the slow fading phenomenon. For a line-topology network with outage links, we study three types of intermediate network node schemes: random linear network coding, store-and-forward, and hop-by-hop retransmission. We provide the analytical formulas for the maximum throughputs and the end-to-end latency for each scheme. To gain a more explicit understanding, we perform a scalability analysis of the throughput and latency as the network length increases. We observe that the same order of throughput/latency holds across a wide range of outage functions for each scheme. We illustrate how our exact formulae and scalability results can be applied to compare different schemes.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"464-477"},"PeriodicalIF":0.0,"publicationDate":"2024-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10571545","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141966000","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Addressing GAN Training Instabilities via Tunable Classification Losses 通过可调分类损失解决 GAN 训练不稳定性问题
Pub Date : 2024-06-19 DOI: 10.1109/JSAIT.2024.3415670
Monica Welfert;Gowtham R. Kurri;Kyle Otstot;Lalitha Sankar
Generative adversarial networks (GANs), modeled as a zero-sum game between a generator (G) and a discriminator (D), allow generating synthetic data with formal guarantees. Noting that D is a classifier, we begin by reformulating the GAN value function using class probability estimation (CPE) losses. We prove a two-way correspondence between CPE loss GANs and f-GANs which minimize f-divergences. We also show that all symmetric f-divergences are equivalent in convergence. In the finite sample and model capacity setting, we define and obtain bounds on estimation and generalization errors. We specialize these results to $alpha $ -GANs, defined using $alpha $ -loss, a tunable CPE loss family parametrized by $alpha in (0,infty $ ]. We next introduce a class of dual-objective GANs to address training instabilities of GANs by modeling each player’s objective using $alpha $ -loss to obtain $(alpha _{D},alpha _{G})$ -GANs. We show that the resulting non-zero sum game simplifies to minimizing an f-divergence under appropriate conditions on $(alpha _{D},alpha _{G})$ . Generalizing this dual-objective formulation using CPE losses, we define and obtain upper bounds on an appropriately defined estimation error. Finally, we highlight the value of tuning $(alpha _{D},alpha _{G})$ in alleviating training instabilities for the synthetic 2D Gaussian mixture ring as well as the large publicly available Celeb-A and LSUN Classroom image datasets.
生成式对抗网络(GAN)被模拟为生成器(G)和判别器(D)之间的零和博弈,可以生成具有形式保证的合成数据。注意到 D 是一个分类器,我们首先使用类概率估计(CPE)损失重新表述了 GAN 的价值函数。我们证明了 CPE 损失 GAN 与 f-GAN 之间的双向对应关系,后者最大限度地减小了 f 分歧。我们还证明了所有对称的 f-divergences 在收敛性上是等价的。在有限样本和模型容量设置中,我们定义并获得了估计误差和泛化误差的界限。我们将这些结果专门应用于$alpha $ -GANs,使用$alpha $ -loss定义,$alpha in (0,infty $ ]是一个参数为$alpha in (0,infty $ ]的可调CPE损失族。接下来,我们引入了一类双目标 GAN,通过使用 $alpha $ -loss 对每个参与者的目标进行建模,得到 $(alpha _{D},alpha _{G})$ -GAN,从而解决 GAN 的训练不稳定性问题。我们证明,由此产生的非零和博弈在 $(alpha _{D},alpha _{G})$ 的适当条件下简化为最小化 f-发散。通过使用 CPE 损失对这一双目标表述进行推广,我们定义并获得了适当定义的估计误差上限。最后,我们强调了调整 $(alpha _{D},alpha _{G})$ 在缓解合成二维高斯混合环以及大型公开 Celeb-A 和 LSUN 课堂图像数据集的训练不稳定性方面的价值。
{"title":"Addressing GAN Training Instabilities via Tunable Classification Losses","authors":"Monica Welfert;Gowtham R. Kurri;Kyle Otstot;Lalitha Sankar","doi":"10.1109/JSAIT.2024.3415670","DOIUrl":"https://doi.org/10.1109/JSAIT.2024.3415670","url":null,"abstract":"Generative adversarial networks (GANs), modeled as a zero-sum game between a generator (G) and a discriminator (D), allow generating synthetic data with formal guarantees. Noting that D is a classifier, we begin by reformulating the GAN value function using class probability estimation (CPE) losses. We prove a two-way correspondence between CPE loss GANs and f-GANs which minimize f-divergences. We also show that all symmetric f-divergences are equivalent in convergence. In the finite sample and model capacity setting, we define and obtain bounds on estimation and generalization errors. We specialize these results to \u0000<inline-formula> <tex-math>$alpha $ </tex-math></inline-formula>\u0000-GANs, defined using \u0000<inline-formula> <tex-math>$alpha $ </tex-math></inline-formula>\u0000-loss, a tunable CPE loss family parametrized by \u0000<inline-formula> <tex-math>$alpha in (0,infty $ </tex-math></inline-formula>\u0000]. We next introduce a class of dual-objective GANs to address training instabilities of GANs by modeling each player’s objective using \u0000<inline-formula> <tex-math>$alpha $ </tex-math></inline-formula>\u0000-loss to obtain \u0000<inline-formula> <tex-math>$(alpha _{D},alpha _{G})$ </tex-math></inline-formula>\u0000-GANs. We show that the resulting non-zero sum game simplifies to minimizing an f-divergence under appropriate conditions on \u0000<inline-formula> <tex-math>$(alpha _{D},alpha _{G})$ </tex-math></inline-formula>\u0000. Generalizing this dual-objective formulation using CPE losses, we define and obtain upper bounds on an appropriately defined estimation error. Finally, we highlight the value of tuning \u0000<inline-formula> <tex-math>$(alpha _{D},alpha _{G})$ </tex-math></inline-formula>\u0000 in alleviating training instabilities for the synthetic 2D Gaussian mixture ring as well as the large publicly available Celeb-A and LSUN Classroom image datasets.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"534-553"},"PeriodicalIF":0.0,"publicationDate":"2024-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141965916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Long-Term Fairness in Sequential Multi-Agent Selection With Positive Reinforcement 带正向强化的连续多代理选择中的长期公平性
Pub Date : 2024-06-18 DOI: 10.1109/JSAIT.2024.3416078
Bhagyashree Puranik;Ozgur Guldogan;Upamanyu Madhow;Ramtin Pedarsani
While much of the rapidly growing literature on fair decision-making focuses on metrics for one-shot decisions, recent work has raised the intriguing possibility of designing sequential decision-making to positively impact long-term social fairness. In selection processes such as college admissions or hiring, biasing slightly towards applicants from under-represented groups is hypothesized to provide positive feedback that increases the pool of under-represented applicants in future selection rounds, thus enhancing fairness in the long term. In this paper, we examine this hypothesis and its consequences in a setting in which multiple agents are selecting from a common pool of applicants. We propose the Multi-agent Fair-Greedy policy, that balances greedy score maximization and fairness. Under this policy, we prove that the resource pool and the admissions converge to a long-term fairness target set by the agents when the score distributions across the groups in the population are identical. We provide empirical evidence of existence of equilibria under non-identical score distributions through synthetic and adapted real-world datasets. We then sound a cautionary note for more complex applicant pool evolution models, under which uncoordinated behavior by the agents can cause negative reinforcement, leading to a reduction in the fraction of under-represented applicants. Our results indicate that, while positive reinforcement is a promising mechanism for long-term fairness, policies must be designed carefully to be robust to variations in the evolution model, with a number of open issues that remain to be explored by algorithm designers, social scientists, and policymakers.
尽管快速增长的有关公平决策的文献大多侧重于一次性决策的衡量标准,但最近的研究提出了一种令人感兴趣的可能性,即通过设计连续决策来对长期社会公平性产生积极影响。在大学录取或招聘等选拔过程中,如果对来自代表性不足群体的申请人略有偏向,就会产生积极的反馈,从而在未来的选拔中增加代表性不足的申请人的数量,从而提高长期的公平性。在本文中,我们将在多个代理从一个共同的申请人库中进行遴选的情况下,对这一假设及其结果进行研究。我们提出了多代理公平-贪婪政策,在贪婪分数最大化和公平性之间取得了平衡。在这一政策下,我们证明了当群体中各组的分数分布相同时,资源池和录取率会趋同于代理设定的长期公平目标。我们通过合成和改编的现实世界数据集,提供了非相同分数分布下存在均衡的经验证据。然后,我们对更复杂的申请者群体演化模型提出了警告,在这种情况下,代理人的不协调行为可能会导致负强化,从而导致代表性不足的申请者比例下降。我们的研究结果表明,虽然正强化是一种有希望实现长期公平的机制,但政策的设计必须谨慎,以适应演化模型的变化,同时还有许多开放性问题有待算法设计者、社会科学家和政策制定者去探索。
{"title":"Long-Term Fairness in Sequential Multi-Agent Selection With Positive Reinforcement","authors":"Bhagyashree Puranik;Ozgur Guldogan;Upamanyu Madhow;Ramtin Pedarsani","doi":"10.1109/JSAIT.2024.3416078","DOIUrl":"https://doi.org/10.1109/JSAIT.2024.3416078","url":null,"abstract":"While much of the rapidly growing literature on fair decision-making focuses on metrics for one-shot decisions, recent work has raised the intriguing possibility of designing sequential decision-making to positively impact long-term social fairness. In selection processes such as college admissions or hiring, biasing slightly towards applicants from under-represented groups is hypothesized to provide positive feedback that increases the pool of under-represented applicants in future selection rounds, thus enhancing fairness in the long term. In this paper, we examine this hypothesis and its consequences in a setting in which multiple agents are selecting from a common pool of applicants. We propose the Multi-agent Fair-Greedy policy, that balances greedy score maximization and fairness. Under this policy, we prove that the resource pool and the admissions converge to a long-term fairness target set by the agents when the score distributions across the groups in the population are identical. We provide empirical evidence of existence of equilibria under non-identical score distributions through synthetic and adapted real-world datasets. We then sound a cautionary note for more complex applicant pool evolution models, under which uncoordinated behavior by the agents can cause negative reinforcement, leading to a reduction in the fraction of under-represented applicants. Our results indicate that, while positive reinforcement is a promising mechanism for long-term fairness, policies must be designed carefully to be robust to variations in the evolution model, with a number of open issues that remain to be explored by algorithm designers, social scientists, and policymakers.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"424-441"},"PeriodicalIF":0.0,"publicationDate":"2024-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141624148","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Controlled Privacy Leakage Propagation Throughout Overlapping Grouped Learning 通过重叠分组学习控制隐私泄露传播
Pub Date : 2024-06-18 DOI: 10.1109/JSAIT.2024.3416089
Shahrzad Kiani;Franziska Boenisch;Stark C. Draper
Federated Learning (FL) is the standard protocol for collaborative learning. In FL, multiple workers jointly train a shared model. They exchange model updates calculated on their data, while keeping the raw data itself local. Since workers naturally form groups based on common interests and privacy policies, we are motivated to extend standard FL to reflect a setting with multiple, potentially overlapping groups. In this setup where workers can belong and contribute to more than one group at a time, complexities arise in understanding privacy leakage and in adhering to privacy policies. To address the challenges, we propose differential private overlapping grouped learning (DP-OGL), a novel method to implement privacy guarantees within overlapping groups. Under the honest-but-curious threat model, we derive novel privacy guarantees between arbitrary pairs of workers. These privacy guarantees describe and quantify two key effects of privacy leakage in DP-OGL: propagation delay, i.e., the fact that information from one group will leak to other groups only with temporal offset through the common workers and information degradation, i.e., the fact that noise addition over model updates limits information leakage between workers. Our experiments show that applying DP-OGL enhances utility while maintaining strong privacy compared to standard FL setups.
联合学习(FL)是协作学习的标准协议。在 FL 中,多个工作人员共同训练一个共享模型。他们交换根据各自数据计算的模型更新,同时保持原始数据本身的本地化。由于工作人员会根据共同的兴趣和隐私政策自然地组成小组,因此我们有动力对标准 FL 进行扩展,以反映具有多个潜在重叠小组的环境。在这种情况下,工人可以同时属于一个以上的小组并为其做出贡献,因此在理解隐私泄露和遵守隐私政策方面出现了复杂的问题。为了应对这些挑战,我们提出了差分隐私重叠分组学习(DP-OGL),这是一种在重叠组内实现隐私保证的新方法。在 "诚实但好奇 "的威胁模型下,我们得出了任意工人对之间的新型隐私保证。这些隐私保证描述并量化了 DP-OGL 中隐私泄漏的两个关键影响:传播延迟,即一个组的信息只有通过共同工作者的时间偏移才会泄漏到其他组;信息退化,即模型更新时的噪声增加限制了工作者之间的信息泄漏。我们的实验表明,与标准的 FL 设置相比,应用 DP-OGL 可以提高效用,同时保持较高的隐私性。
{"title":"Controlled Privacy Leakage Propagation Throughout Overlapping Grouped Learning","authors":"Shahrzad Kiani;Franziska Boenisch;Stark C. Draper","doi":"10.1109/JSAIT.2024.3416089","DOIUrl":"https://doi.org/10.1109/JSAIT.2024.3416089","url":null,"abstract":"Federated Learning (FL) is the standard protocol for collaborative learning. In FL, multiple workers jointly train a shared model. They exchange model updates calculated on their data, while keeping the raw data itself local. Since workers naturally form groups based on common interests and privacy policies, we are motivated to extend standard FL to reflect a setting with multiple, potentially overlapping groups. In this setup where workers can belong and contribute to more than one group at a time, complexities arise in understanding privacy leakage and in adhering to privacy policies. To address the challenges, we propose differential private overlapping grouped learning (DP-OGL), a novel method to implement privacy guarantees within overlapping groups. Under the honest-but-curious threat model, we derive novel privacy guarantees between arbitrary pairs of workers. These privacy guarantees describe and quantify two key effects of privacy leakage in DP-OGL: propagation delay, i.e., the fact that information from one group will leak to other groups only with temporal offset through the common workers and information degradation, i.e., the fact that noise addition over model updates limits information leakage between workers. Our experiments show that applying DP-OGL enhances utility while maintaining strong privacy compared to standard FL setups.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"442-463"},"PeriodicalIF":0.0,"publicationDate":"2024-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141624126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Information Velocity of Cascaded Gaussian Channels With Feedback 带反馈的级联高斯信道的信息速度
Pub Date : 2024-06-18 DOI: 10.1109/JSAIT.2024.3416310
Elad Domanovitz;Anatoly Khina;Tal Philosof;Yuval Kochman
We consider a line network of nodes, connected by additive white noise channels, equipped with local feedback. We study the velocity at which information spreads over this network. For transmission of a data packet, we give an explicit positive lower bound on the velocity, for any packet size. Furthermore, we consider streaming, that is, transmission of data packets generated at a given average arrival rate. We show that a positive velocity exists as long as the arrival rate is below the individual Gaussian channel capacity, and provide an explicit lower bound. Our analysis involves applying pulse-amplitude modulation to the data (successively in the streaming case), and using linear mean-squared error estimation at the network nodes. For general white noise, we derive exponential error-probability bounds. For single-packet transmission over channels with (sub-)Gaussian noise, we show a doubly-exponential behavior, which reduces to the celebrated Schalkwijk–Kailath scheme when considering a single node. Viewing the constellation as an “analog source”, we also provide bounds on the exponential decay of the mean-squared error of source transmission over the network.
我们考虑了一个由节点组成的线性网络,该网络通过加性白噪声信道连接,并配有局部反馈。我们研究信息在该网络中的传播速度。对于数据包的传输,无论数据包大小如何,我们都给出了速度的明确正下限。此外,我们还考虑了流式传输,即传输以给定平均到达率生成的数据包。我们证明,只要到达率低于单个高斯信道容量,就存在正速度,并给出了明确的下限。我们的分析包括对数据进行脉冲幅度调制(在流式情况下连续进行),并在网络节点使用线性均方误差估计。对于一般白噪声,我们推导出指数误差概率边界。对于在具有(亚)高斯噪声的信道上进行的单包传输,我们展示了一种双指数行为,当考虑单个节点时,这种行为简化为著名的 Schalkwijk-Kailath 方案。将星座视为 "模拟源",我们还提供了网络上源传输均方误差的指数衰减约束。
{"title":"Information Velocity of Cascaded Gaussian Channels With Feedback","authors":"Elad Domanovitz;Anatoly Khina;Tal Philosof;Yuval Kochman","doi":"10.1109/JSAIT.2024.3416310","DOIUrl":"https://doi.org/10.1109/JSAIT.2024.3416310","url":null,"abstract":"We consider a line network of nodes, connected by additive white noise channels, equipped with local feedback. We study the velocity at which information spreads over this network. For transmission of a data packet, we give an explicit positive lower bound on the velocity, for any packet size. Furthermore, we consider streaming, that is, transmission of data packets generated at a given average arrival rate. We show that a positive velocity exists as long as the arrival rate is below the individual Gaussian channel capacity, and provide an explicit lower bound. Our analysis involves applying pulse-amplitude modulation to the data (successively in the streaming case), and using linear mean-squared error estimation at the network nodes. For general white noise, we derive exponential error-probability bounds. For single-packet transmission over channels with (sub-)Gaussian noise, we show a doubly-exponential behavior, which reduces to the celebrated Schalkwijk–Kailath scheme when considering a single node. Viewing the constellation as an “analog source”, we also provide bounds on the exponential decay of the mean-squared error of source transmission over the network.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"554-569"},"PeriodicalIF":0.0,"publicationDate":"2024-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141980053","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neural Distributed Source Coding 神经分布式源编码
Pub Date : 2024-06-14 DOI: 10.1109/JSAIT.2024.3412976
Jay Whang;Alliot Nagle;Anish Acharya;Hyeji Kim;Alexandros G. Dimakis
We consider the Distributed Source Coding (DSC) problem concerning the task of encoding an input in the absence of correlated side information that is only available to the decoder. Remarkably, Slepian and Wolf showed in 1973 that an encoder without access to the side information can asymptotically achieve the same compression rate as when the side information is available to it. This seminal result was later extended to lossy compression of distributed sources by Wyner, Ziv, Berger, and Tung. While there is vast prior work on this topic, practical DSC has been limited to synthetic datasets and specific correlation structures. Here we present a framework for lossy DSC that is agnostic to the correlation structure and can scale to high dimensions. Rather than relying on hand-crafted source modeling, our method utilizes a conditional Vector-Quantized Variational auto-encoder (VQ-VAE) to learn the distributed encoder and decoder. We evaluate our method on multiple datasets and show that our method can handle complex correlations and achieves state-of-the-art PSNR.
我们考虑的分布式源编码(DSC)问题涉及在没有相关边信息的情况下对输入进行编码的任务,而这些边信息只有解码器才能获得。令人瞩目的是,Slepian 和 Wolf 于 1973 年证明,在无法获得边信息的情况下,编码器可以渐进地达到与获得边信息时相同的压缩率。这一开创性成果后来被 Wyner、Ziv、Berger 和 Tung 扩展到分布式信号源的有损压缩。虽然之前有大量关于这一主题的研究,但实用的 DSC 一直局限于合成数据集和特定的相关结构。在这里,我们提出了一种有损 DSC 框架,它与相关结构无关,并能扩展到高维度。我们的方法不依赖手工制作的源建模,而是利用条件矢量量化变异自动编码器(VQ-VAE)来学习分布式编码器和解码器。我们在多个数据集上对我们的方法进行了评估,结果表明我们的方法可以处理复杂的相关性,并达到最先进的 PSNR。
{"title":"Neural Distributed Source Coding","authors":"Jay Whang;Alliot Nagle;Anish Acharya;Hyeji Kim;Alexandros G. Dimakis","doi":"10.1109/JSAIT.2024.3412976","DOIUrl":"https://doi.org/10.1109/JSAIT.2024.3412976","url":null,"abstract":"We consider the Distributed Source Coding (DSC) problem concerning the task of encoding an input in the absence of correlated side information that is only available to the decoder. Remarkably, Slepian and Wolf showed in 1973 that an encoder without access to the side information can asymptotically achieve the same compression rate as when the side information is available to it. This seminal result was later extended to lossy compression of distributed sources by Wyner, Ziv, Berger, and Tung. While there is vast prior work on this topic, practical DSC has been limited to synthetic datasets and specific correlation structures. Here we present a framework for lossy DSC that is agnostic to the correlation structure and can scale to high dimensions. Rather than relying on hand-crafted source modeling, our method utilizes a conditional Vector-Quantized Variational auto-encoder (VQ-VAE) to learn the distributed encoder and decoder. We evaluate our method on multiple datasets and show that our method can handle complex correlations and achieves state-of-the-art PSNR.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"493-508"},"PeriodicalIF":0.0,"publicationDate":"2024-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141965287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Secure Source Coding Resilient Against Compromised Users via an Access Structure 安全源代码编码通过访问结构抵御受攻击用户的攻击
Pub Date : 2024-06-10 DOI: 10.1109/JSAIT.2024.3410235
Hassan ZivariFard;Rémi A. Chou
Consider a source and multiple users who observe the independent and identically distributed (i.i.d.) copies of correlated Gaussian random variables. The source wishes to compress its observations and store the result in a public database such that (i) authorized sets of users are able to reconstruct the source with a certain distortion level, and (ii) information leakage to non-authorized sets of colluding users is minimized. In other words, the recovery of the source is restricted to a predefined access structure. The main result of this paper is a closed-form characterization of the fundamental trade-off between the source coding rate and the information leakage rate. As an example, threshold access structures are studied, i.e., the case where any set of at least t users is able to reconstruct the source with some predefined distortion level and the information leakage at any set of users with a size smaller than t is minimized.
考虑一个信号源和多个用户,他们观察相关高斯随机变量的独立且同分布(i.i.d.)副本。数据源希望压缩其观测数据并将结果存储在公共数据库中,以便(i)授权用户集能够以一定的失真度重建数据源,以及(ii)向非授权串通用户集的信息泄漏最小化。换句话说,信息源的恢复仅限于预定义的访问结构。本文的主要成果是对信源编码率和信息泄漏率之间基本权衡的闭式表征。举例来说,本文研究了阈值访问结构,即任何至少由 t 个用户组成的用户集都能以某种预定的失真水平重构信源,且任何小于 t 的用户集的信息泄漏最小。
{"title":"Secure Source Coding Resilient Against Compromised Users via an Access Structure","authors":"Hassan ZivariFard;Rémi A. Chou","doi":"10.1109/JSAIT.2024.3410235","DOIUrl":"https://doi.org/10.1109/JSAIT.2024.3410235","url":null,"abstract":"Consider a source and multiple users who observe the independent and identically distributed (i.i.d.) copies of correlated Gaussian random variables. The source wishes to compress its observations and store the result in a public database such that (i) authorized sets of users are able to reconstruct the source with a certain distortion level, and (ii) information leakage to non-authorized sets of colluding users is minimized. In other words, the recovery of the source is restricted to a predefined access structure. The main result of this paper is a closed-form characterization of the fundamental trade-off between the source coding rate and the information leakage rate. As an example, threshold access structures are studied, i.e., the case where any set of at least \u0000<italic>t</i>\u0000 users is able to reconstruct the source with some predefined distortion level and the information leakage at any set of users with a size smaller than \u0000<italic>t</i>\u0000 is minimized.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"478-492"},"PeriodicalIF":0.0,"publicationDate":"2024-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141965288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Information-Theoretic Tools to Understand Distributed Source Coding in Neuroscience 理解神经科学中分布式源编码的信息论工具
Pub Date : 2024-06-10 DOI: 10.1109/JSAIT.2024.3409683
Ariel K. Feldman;Praveen Venkatesh;Douglas J. Weber;Pulkit Grover
This paper brings together topics of two of Berger’s main contributions to information theory: distributed source coding, and living information theory. Our goal is to understand which information theory techniques can be helpful in understanding a distributed source coding strategy used by the natural world. Towards this goal, we study the example of the encoding of location of an animal by grid cells in its brain. We use information measures of partial information decomposition (PID) to assess the unique, redundant, and synergistic information carried by multiple grid cells, first for simulated grid cells utilizing known encodings, and subsequently for data from real grid cells. In all cases, we make simplifying assumptions so we can assess the consistency of specific PID definitions with intuition. Our results suggest that the measure of PID proposed by Bertschinger et al. (Entropy, 2014) provides intuitive insights on distributed source coding by grid cells, and can be used for subsequent studies for understanding grid-cell encoding as well as broadly in neuroscience.
本文汇集了伯杰对信息论的两大贡献:分布式源编码和生命信息论。我们的目标是了解哪些信息论技术有助于理解自然界使用的分布式源编码策略。为了实现这一目标,我们以动物大脑中的网格细胞对其位置进行编码为例进行研究。我们使用部分信息分解(PID)的信息量来评估多个网格细胞所携带的独特、冗余和协同信息,首先是利用已知编码的模拟网格细胞,然后是真实网格细胞的数据。在所有情况下,我们都做了简化假设,以便评估特定 PID 定义与直觉的一致性。我们的结果表明,Bertschinger 等人提出的 PID 测量方法(熵,2014 年)为网格细胞的分布式源编码提供了直观的见解,可用于后续研究,以了解网格细胞编码以及广泛的神经科学。
{"title":"Information-Theoretic Tools to Understand Distributed Source Coding in Neuroscience","authors":"Ariel K. Feldman;Praveen Venkatesh;Douglas J. Weber;Pulkit Grover","doi":"10.1109/JSAIT.2024.3409683","DOIUrl":"https://doi.org/10.1109/JSAIT.2024.3409683","url":null,"abstract":"This paper brings together topics of two of Berger’s main contributions to information theory: distributed source coding, and living information theory. Our goal is to understand which information theory techniques can be helpful in understanding a distributed source coding strategy used by the natural world. Towards this goal, we study the example of the encoding of location of an animal by grid cells in its brain. We use information measures of partial information decomposition (PID) to assess the unique, redundant, and synergistic information carried by multiple grid cells, first for simulated grid cells utilizing known encodings, and subsequently for data from real grid cells. In all cases, we make simplifying assumptions so we can assess the consistency of specific PID definitions with intuition. Our results suggest that the measure of PID proposed by Bertschinger et al. (Entropy, 2014) provides intuitive insights on distributed source coding by grid cells, and can be used for subsequent studies for understanding grid-cell encoding as well as broadly in neuroscience.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"509-519"},"PeriodicalIF":0.0,"publicationDate":"2024-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141965286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Fundamental Limit of Distributed Learning With Interchangable Constrained Statistics 论可互换约束统计分布式学习的基本极限
Pub Date : 2024-06-04 DOI: 10.1109/JSAIT.2024.3409426
Xinyi Tong;Jian Xu;Shao-Lun Huang
In the popular federated learning scenarios, distributed nodes often represent and exchange information through functions or statistics of data, with communicative processes constrained by the dimensionality of transmitted information. This paper investigates the fundamental limits of distributed parameter estimation and model training problems under such constraints. Specifically, we assume that each node can observe a sequence of i.i.d. sampled data and communicate statistics of the observed data with dimensionality constraints. We first show the Cramer-Rao lower bound (CRLB) and the corresponding achievable estimators for the distributed parameter estimation problems, and the geometric insights and the computable algorithms of designing efficient estimators are also presented. Moreover, we consider model parameters training for distributed nodes with limited communicable statistics. We demonstrate that in order to optimize the excess risk, the feature functions of the statistics shall be designed along the largest eigenvectors of a matrix induced by the model training loss function. In summary, our results potentially provide theoretical guidelines of designing efficient algorithms for enhancing the performance of distributed learning systems.
在流行的联合学习场景中,分布式节点通常通过函数或数据统计来表示和交换信息,通信过程受到传输信息维度的限制。本文研究了这种约束条件下分布式参数估计和模型训练问题的基本限制。具体来说,我们假设每个节点都能观察到 i.i.d. 采样数据序列,并在维度限制下交流观察到的数据统计量。我们首先展示了分布式参数估计问题的 Cramer-Rao 下界(CRLB)和相应的可实现估计器,还介绍了设计高效估计器的几何见解和可计算算法。此外,我们还考虑了具有有限可通信统计量的分布式节点的模型参数训练问题。我们证明,为了优化超额风险,应沿着模型训练损失函数诱导的矩阵的最大特征向量设计统计特征函数。总之,我们的研究结果为设计高效算法以提高分布式学习系统的性能提供了理论指导。
{"title":"On the Fundamental Limit of Distributed Learning With Interchangable Constrained Statistics","authors":"Xinyi Tong;Jian Xu;Shao-Lun Huang","doi":"10.1109/JSAIT.2024.3409426","DOIUrl":"https://doi.org/10.1109/JSAIT.2024.3409426","url":null,"abstract":"In the popular federated learning scenarios, distributed nodes often represent and exchange information through functions or statistics of data, with communicative processes constrained by the dimensionality of transmitted information. This paper investigates the fundamental limits of distributed parameter estimation and model training problems under such constraints. Specifically, we assume that each node can observe a sequence of i.i.d. sampled data and communicate statistics of the observed data with dimensionality constraints. We first show the Cramer-Rao lower bound (CRLB) and the corresponding achievable estimators for the distributed parameter estimation problems, and the geometric insights and the computable algorithms of designing efficient estimators are also presented. Moreover, we consider model parameters training for distributed nodes with limited communicable statistics. We demonstrate that in order to optimize the excess risk, the feature functions of the statistics shall be designed along the largest eigenvectors of a matrix induced by the model training loss function. In summary, our results potentially provide theoretical guidelines of designing efficient algorithms for enhancing the performance of distributed learning systems.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"396-406"},"PeriodicalIF":0.0,"publicationDate":"2024-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141624095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LightVeriFL: A Lightweight and Verifiable Secure Aggregation for Federated Learning LightVeriFL:用于联合学习的轻量级可验证安全聚合系统
Pub Date : 2024-04-29 DOI: 10.1109/JSAIT.2024.3391849
Baturalp Buyukates;Jinhyun So;Hessam Mahdavifar;Salman Avestimehr
Secure aggregation protects the local models of the users in federated learning, by not allowing the server to obtain any information beyond the aggregate model at each iteration. Naively implementing secure aggregation fails to protect the integrity of the aggregate model in the possible presence of a malicious server forging the aggregation result, which motivates verifiable aggregation in federated learning. Existing verifiable aggregation schemes either have a linear complexity in model size or require time-consuming reconstruction at the server, that is quadratic in the number of users, in case of likely user dropouts. To overcome these limitations, we propose LightVeriFL, a lightweight and communication-efficient secure verifiable aggregation protocol, that provides the same guarantees for verifiability against a malicious server, data privacy, and dropout-resilience as the state-of-the-art protocols without incurring substantial communication and computation overheads. The proposed LightVeriFL protocol utilizes homomorphic hash and commitment functions of constant length, that are independent of the model size, to enable verification at the users. In case of dropouts, LightVeriFL uses a one-shot aggregate hash recovery of the dropped-out users, instead of a one-by-one recovery, making the verification process significantly faster than the existing approaches. Comprehensive experiments show the advantage of LightVeriFL in practical settings.
安全聚合可以保护联合学习中用户的本地模型,在每次迭代时不允许服务器获取聚合模型之外的任何信息。在可能存在恶意服务器伪造聚合结果的情况下,天真地实施安全聚合无法保护聚合模型的完整性,这就是联合学习中可验证聚合的动机。现有的可验证聚合方案要么复杂度与模型大小呈线性关系,要么需要在服务器上进行耗时的重构,而重构的复杂度与用户数量呈二次方关系,以防可能出现的用户放弃情况。为了克服这些局限性,我们提出了轻量级、通信效率高的安全可验证聚合协议--LightVeriFL,它能提供与最先进协议相同的针对恶意服务器的可验证性、数据私密性和抗丢弃性保证,而不会产生大量通信和计算开销。所提出的 LightVeriFL 协议利用长度恒定的同态哈希和承诺函数(与模型大小无关)来实现用户验证。在出现掉线的情况下,LightVeriFL 采用对掉线用户进行一次聚合哈希恢复,而不是逐个恢复,从而使验证过程明显快于现有方法。综合实验显示了 LightVeriFL 在实际应用中的优势。
{"title":"LightVeriFL: A Lightweight and Verifiable Secure Aggregation for Federated Learning","authors":"Baturalp Buyukates;Jinhyun So;Hessam Mahdavifar;Salman Avestimehr","doi":"10.1109/JSAIT.2024.3391849","DOIUrl":"https://doi.org/10.1109/JSAIT.2024.3391849","url":null,"abstract":"Secure aggregation protects the local models of the users in federated learning, by not allowing the server to obtain any information beyond the aggregate model at each iteration. Naively implementing secure aggregation fails to protect the integrity of the aggregate model in the possible presence of a malicious server forging the aggregation result, which motivates verifiable aggregation in federated learning. Existing verifiable aggregation schemes either have a linear complexity in model size or require time-consuming reconstruction at the server, that is quadratic in the number of users, in case of likely user dropouts. To overcome these limitations, we propose \u0000<monospace>LightVeriFL</monospace>\u0000, a lightweight and communication-efficient secure verifiable aggregation protocol, that provides the same guarantees for verifiability against a malicious server, data privacy, and dropout-resilience as the state-of-the-art protocols without incurring substantial communication and computation overheads. The proposed \u0000<monospace>LightVeriFL</monospace>\u0000 protocol utilizes homomorphic hash and commitment functions of constant length, that are independent of the model size, to enable verification at the users. In case of dropouts, \u0000<monospace>LightVeriFL</monospace>\u0000 uses a one-shot aggregate hash recovery of the dropped-out users, instead of a one-by-one recovery, making the verification process significantly faster than the existing approaches. Comprehensive experiments show the advantage of \u0000<monospace>LightVeriFL</monospace>\u0000 in practical settings.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"285-301"},"PeriodicalIF":0.0,"publicationDate":"2024-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141096202","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE journal on selected areas in information theory
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1