首页 > 最新文献

2022 IEEE International Symposium on Information Theory (ISIT)最新文献

英文 中文
Capacity-Achieving Constrained Codes with GC-Content and Runlength Limits for DNA Storage 具有gc含量和运行长度限制的DNA存储容量受限代码
Pub Date : 2022-06-26 DOI: 10.1109/ISIT50566.2022.9834494
Yajuan Liu, Xuan He, Xiaohu Tang
GC-content and homopolymer run are two constraints of interest in DNA storage systems. Extensive experiments showed that if GC-content is too high (low), or homopolymer run exceeds six in a DNA sequence, there will give rise to dramatical increase of insertion, deletion and substitution errors. Committing to study the DNA sequences with both constraints, a recent work (Nguyen et al. 2020) proposed a class of (ϵ, ℓ)-constrained codes that can only asymptotically approach the capacity, but may have reasonable loss for finite code lengths.In this paper, we design the first (ϵ, ℓ)-constrained codes based on the enumeration coding technique which can always achieve capacity regardless of code lengths. In addition, motivated by the influence of local GC-content, we consider a nontrivial case that the prefixes of a DNA sequence also hold GC-content constraint for the first time, called (δ,ℓ)-prefix constrained codes.
gc含量和均聚物运行是DNA存储系统的两个限制因素。大量实验表明,如果一个DNA序列中gc含量过高(过低)或均聚物数超过6个,插入、删除和替换错误将显著增加。致力于研究具有这两种约束的DNA序列,最近的一项工作(Nguyen et al. 2020)提出了一类(λ, λ)约束的编码,它只能渐近地接近容量,但对于有限的编码长度可能有合理的损失。在本文中,我们设计了第一个基于枚举编码技术的(λ, λ)约束码,无论码长如何,都能获得容量。此外,受局部gc含量的影响,我们考虑了一种非平凡的情况,即DNA序列的前缀也首次具有gc含量约束,称为(δ, r)-前缀约束码。
{"title":"Capacity-Achieving Constrained Codes with GC-Content and Runlength Limits for DNA Storage","authors":"Yajuan Liu, Xuan He, Xiaohu Tang","doi":"10.1109/ISIT50566.2022.9834494","DOIUrl":"https://doi.org/10.1109/ISIT50566.2022.9834494","url":null,"abstract":"GC-content and homopolymer run are two constraints of interest in DNA storage systems. Extensive experiments showed that if GC-content is too high (low), or homopolymer run exceeds six in a DNA sequence, there will give rise to dramatical increase of insertion, deletion and substitution errors. Committing to study the DNA sequences with both constraints, a recent work (Nguyen et al. 2020) proposed a class of (ϵ, ℓ)-constrained codes that can only asymptotically approach the capacity, but may have reasonable loss for finite code lengths.In this paper, we design the first (ϵ, ℓ)-constrained codes based on the enumeration coding technique which can always achieve capacity regardless of code lengths. In addition, motivated by the influence of local GC-content, we consider a nontrivial case that the prefixes of a DNA sequence also hold GC-content constraint for the first time, called (δ,ℓ)-prefix constrained codes.","PeriodicalId":348168,"journal":{"name":"2022 IEEE International Symposium on Information Theory (ISIT)","volume":"141 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115758673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
PCR, Tropical Arithmetic, and Group Testing 聚合酶链反应,热带算术,和组检验
Pub Date : 2022-06-26 DOI: 10.1109/ISIT50566.2022.9834718
Hsin-Po Wang, Ryan Gabrys, A. Vardy
Polymerase chain reaction (PCR) testing is the gold standard for diagnosing COVID-19. Unfortunately, the outputs of these tests are imprecise and therefore quantitative group testing methods, which rely on precise measurements, are not applicable. Motivated by the ever-increasing demand to identify individuals infected with SARS-CoV-19, we propose a new model that leverages tropical arithmetic to characterize the PCR testing process. In many cases, some of which are highlighted in this work, tropical group testing is provably more powerful than traditional binary group testing in that it requires fewer tests than classical approaches, while additionally providing a mechanism to identify the viral load of each infected individual.
聚合酶链反应(PCR)检测是诊断COVID-19的金标准。不幸的是,这些测试的输出是不精确的,因此依赖于精确测量的定量组测试方法不适用。由于识别SARS-CoV-19感染者的需求不断增加,我们提出了一种利用热带算法表征PCR检测过程的新模型。在许多情况下,其中一些在本工作中得到了强调,热带群体检测被证明比传统的二元群体检测更强大,因为它比经典方法需要更少的测试,同时还提供了一种识别每个感染者病毒载量的机制。
{"title":"PCR, Tropical Arithmetic, and Group Testing","authors":"Hsin-Po Wang, Ryan Gabrys, A. Vardy","doi":"10.1109/ISIT50566.2022.9834718","DOIUrl":"https://doi.org/10.1109/ISIT50566.2022.9834718","url":null,"abstract":"Polymerase chain reaction (PCR) testing is the gold standard for diagnosing COVID-19. Unfortunately, the outputs of these tests are imprecise and therefore quantitative group testing methods, which rely on precise measurements, are not applicable. Motivated by the ever-increasing demand to identify individuals infected with SARS-CoV-19, we propose a new model that leverages tropical arithmetic to characterize the PCR testing process. In many cases, some of which are highlighted in this work, tropical group testing is provably more powerful than traditional binary group testing in that it requires fewer tests than classical approaches, while additionally providing a mechanism to identify the viral load of each infected individual.","PeriodicalId":348168,"journal":{"name":"2022 IEEE International Symposium on Information Theory (ISIT)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124306213","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Group Testing with Geometric Ranges 几何范围的群检验
Pub Date : 2022-06-26 DOI: 10.1109/ISIT50566.2022.9834574
Benjamin Aram Berendsohn, L. Kozma
Group testing is a well-studied approach for identifying t defective items in a set X of m items, by testing appropriately chosen subsets of X. In classical group testing any subset of X can be tested, and for $t in {mathcal{O}}(1)$ the optimal number of (non-adaptive) tests is known to be Θ(logm).In this work we consider a novel geometric setting for group testing, where the items are points in Euclidean space and the tests are axis-parallel boxes (hyperrectangles), corresponding to the scenario where tests are defined by parameter-ranges (say, according to physical measurements). We present upper and lower bounds on the required number of tests in this setting, observing that in contrast to the unrestricted, combinatorial case, the bounds are polynomial in m. For instance, we show that with two parameters, identifying a defective pair of items requires Ω(m3/5) tests, and there exist configurations for which ${mathcal{O}}left({{m^{2/3}}}right)$ tests are sufficient, whereas to identify a single defective item Θ(m1/2) tests are always necessary and sometimes sufficient. Perhaps most interestingly, our work brings to the study of group testing a set of techniques from extremal combinatorics.
群测试是一种经过充分研究的方法,通过测试X的适当选择的子集来识别m个项目集合X中的t个缺陷项目。在经典的群测试中,X的任何子集都可以被测试,对于$t in {mathcal{O}}(1)$,(非自适应)测试的最佳数量已知为Θ(logm)。在这项工作中,我们考虑了一种新的组测试几何设置,其中项目是欧几里得空间中的点,测试是轴平行盒(超矩形),对应于测试由参数范围定义的场景(例如,根据物理测量)。在这种情况下,我们给出了所需测试次数的上界和下界,观察到与不受限制的组合情况相反,边界是m中的多项式。例如,我们表明,对于两个参数,识别缺陷对需要Ω(m3/5)测试,并且存在${mathcal{O}}左({{m^{2/3}}}}右)$测试足够的配置,然而,为了识别单个缺陷项目Θ(m1/2),测试总是必要的,有时是足够的。也许最有趣的是,我们的工作将极值组合学中的一组技术引入了对群体测试的研究。
{"title":"Group Testing with Geometric Ranges","authors":"Benjamin Aram Berendsohn, L. Kozma","doi":"10.1109/ISIT50566.2022.9834574","DOIUrl":"https://doi.org/10.1109/ISIT50566.2022.9834574","url":null,"abstract":"Group testing is a well-studied approach for identifying t defective items in a set X of m items, by testing appropriately chosen subsets of X. In classical group testing any subset of X can be tested, and for $t in {mathcal{O}}(1)$ the optimal number of (non-adaptive) tests is known to be Θ(logm).In this work we consider a novel geometric setting for group testing, where the items are points in Euclidean space and the tests are axis-parallel boxes (hyperrectangles), corresponding to the scenario where tests are defined by parameter-ranges (say, according to physical measurements). We present upper and lower bounds on the required number of tests in this setting, observing that in contrast to the unrestricted, combinatorial case, the bounds are polynomial in m. For instance, we show that with two parameters, identifying a defective pair of items requires Ω(m3/5) tests, and there exist configurations for which ${mathcal{O}}left({{m^{2/3}}}right)$ tests are sufficient, whereas to identify a single defective item Θ(m1/2) tests are always necessary and sometimes sufficient. Perhaps most interestingly, our work brings to the study of group testing a set of techniques from extremal combinatorics.","PeriodicalId":348168,"journal":{"name":"2022 IEEE International Symposium on Information Theory (ISIT)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124538043","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Capacity of the Shotgun Sequencing Channel 霰弹枪序列通道的容量
Pub Date : 2022-06-26 DOI: 10.1109/ISIT50566.2022.9834409
Aditya Narayan Ravi, Alireza Vahid, Ilan Shomorony
Most DNA sequencing technologies are based on the shotgun paradigm: many short reads are obtained from random unknown locations in the DNA sequence. A fundamental question, studied in [1], is what read length and coverage depth (i.e., the total number of reads) are needed to guarantee reliable sequence reconstruction. Motivated by DNA-based storage, we study the coded version of this problem; i.e., the scenario in which the DNA molecule being sequenced is a codeword from a predefined codebook. Our main result is an exact characterization of the capacity of the resulting shotgun sequencing channel as a function of the read length and coverage depth. In particular, our results imply that while in the uncoded case, O(n) reads of length greater than 2logn are needed for reliable reconstruction of a length-n binary sequence, in the coded case, only O(n/log n) reads of length greater than log n are needed for the capacity to be arbitrarily close to 1.
大多数DNA测序技术都是基于霰弹枪模式:许多短读是从DNA序列中随机未知的位置获得的。[1]研究的一个基本问题是,需要多大的读取长度和覆盖深度(即总读取次数)才能保证可靠的序列重建。基于dna存储的动机,我们研究了这个问题的编码版本;即,被测序的DNA分子是来自预定义码本的码字的场景。我们的主要结果是准确表征了由此产生的鸟枪测序通道的容量作为读取长度和覆盖深度的函数。特别是,我们的结果表明,在未编码情况下,为了可靠地重建长度为n的二进制序列,需要O(n)次长度大于2logn的读取,而在编码情况下,只需要O(n/log n)次长度大于logn的读取,容量就可以任意接近1。
{"title":"Capacity of the Shotgun Sequencing Channel","authors":"Aditya Narayan Ravi, Alireza Vahid, Ilan Shomorony","doi":"10.1109/ISIT50566.2022.9834409","DOIUrl":"https://doi.org/10.1109/ISIT50566.2022.9834409","url":null,"abstract":"Most DNA sequencing technologies are based on the shotgun paradigm: many short reads are obtained from random unknown locations in the DNA sequence. A fundamental question, studied in [1], is what read length and coverage depth (i.e., the total number of reads) are needed to guarantee reliable sequence reconstruction. Motivated by DNA-based storage, we study the coded version of this problem; i.e., the scenario in which the DNA molecule being sequenced is a codeword from a predefined codebook. Our main result is an exact characterization of the capacity of the resulting shotgun sequencing channel as a function of the read length and coverage depth. In particular, our results imply that while in the uncoded case, O(n) reads of length greater than 2logn are needed for reliable reconstruction of a length-n binary sequence, in the coded case, only O(n/log n) reads of length greater than log n are needed for the capacity to be arbitrarily close to 1.","PeriodicalId":348168,"journal":{"name":"2022 IEEE International Symposium on Information Theory (ISIT)","volume":"217 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114370202","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning neural codes for perceptual uncertainty 学习感知不确定性的神经编码
Pub Date : 2022-06-26 DOI: 10.1109/ISIT50566.2022.9834606
M. Salmasi, M. Sahani
Perception is an inferential process, in which the state of the immediate environment must be estimated from sensory input. Inference in the face of noise and ambiguity requires reasoning with uncertainty, and much animal behaviour appears close to Bayes optimal. This observation has inspired hypotheses for how the activity of neurons in the brain might represent the distributional beliefs necessary to implement explicit Bayesian computation. While previous work has focused on the sufficiency of these hypothesised codes for computation, relatively little consideration has been given to optimality in the representation itself. Here, we adopt an encoder-decoder approach to study representational optimisation within one hypothesised belief encoding framework: the distributed distributional code (DDC). We consider a setting in which typical belief distribution functions take the form of a sparse combination of an underlying set of basis functions, and the corresponding DDC signals are corrupted by neural variability. We estimate the conditional entropy over beliefs induced by these DDC signals using an appropriate decoder. Like other hypothesised frameworks, a DDC representation of a belief depends on a set of fixed encoding functions that are usually set arbitrarily. Our approach allows us to seek the encoding functions that minimise the decoder conditional entropy and thus optimise representational accuracy in an information theoretic sense. We apply the approach to show how optimal encoding properties may adapt to represent beliefs in new environments, relating the results to experimentally reported neural responses.
感知是一个推理过程,在这个过程中,必须通过感官输入来估计周围环境的状态。面对噪音和模糊性的推理需要不确定性推理,而许多动物行为似乎接近贝叶斯最优。这一观察启发了一些假设,即大脑中神经元的活动如何代表实现显式贝叶斯计算所必需的分布信念。虽然以前的工作主要集中在这些假设代码的充分性上,但相对较少考虑到表示本身的最优性。在这里,我们采用编码器-解码器方法来研究一个假设的信念编码框架中的表征优化:分布式分布代码(DDC)。我们考虑了一种典型的信念分布函数采用底层基函数集的稀疏组合形式的设置,并且相应的DDC信号被神经变异性破坏。我们使用合适的解码器估计由这些DDC信号引起的信念的条件熵。与其他假设框架一样,信念的DDC表示依赖于一组固定的编码函数,这些函数通常是任意设置的。我们的方法允许我们寻找编码函数,使解码器条件熵最小化,从而在信息理论意义上优化表征精度。我们应用该方法来展示最佳编码特性如何适应新环境中的信念,并将结果与实验报告的神经反应联系起来。
{"title":"Learning neural codes for perceptual uncertainty","authors":"M. Salmasi, M. Sahani","doi":"10.1109/ISIT50566.2022.9834606","DOIUrl":"https://doi.org/10.1109/ISIT50566.2022.9834606","url":null,"abstract":"Perception is an inferential process, in which the state of the immediate environment must be estimated from sensory input. Inference in the face of noise and ambiguity requires reasoning with uncertainty, and much animal behaviour appears close to Bayes optimal. This observation has inspired hypotheses for how the activity of neurons in the brain might represent the distributional beliefs necessary to implement explicit Bayesian computation. While previous work has focused on the sufficiency of these hypothesised codes for computation, relatively little consideration has been given to optimality in the representation itself. Here, we adopt an encoder-decoder approach to study representational optimisation within one hypothesised belief encoding framework: the distributed distributional code (DDC). We consider a setting in which typical belief distribution functions take the form of a sparse combination of an underlying set of basis functions, and the corresponding DDC signals are corrupted by neural variability. We estimate the conditional entropy over beliefs induced by these DDC signals using an appropriate decoder. Like other hypothesised frameworks, a DDC representation of a belief depends on a set of fixed encoding functions that are usually set arbitrarily. Our approach allows us to seek the encoding functions that minimise the decoder conditional entropy and thus optimise representational accuracy in an information theoretic sense. We apply the approach to show how optimal encoding properties may adapt to represent beliefs in new environments, relating the results to experimentally reported neural responses.","PeriodicalId":348168,"journal":{"name":"2022 IEEE International Symposium on Information Theory (ISIT)","volume":"301 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114583776","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
On Information-Debt-Optimal Streaming Codes With Small Memory 小内存下信息债务最优流码研究
Pub Date : 2022-06-26 DOI: 10.1109/ISIT50566.2022.9834842
Vinayak Ramkumar, M. Krishnan, Myna Vajha, P. V. Kumar
In the context of an (n,k,m) convolutional code where k is the number of message symbols, n the number of code symbols and m the memory, Martinian [1] introduced the concept of information debt whose value at time t is the number of additional coded symbols needed to decode all prior message symbols. The same paper shows the existence of (n,k,m) convolutional codes that can recover all prior message symbols whenever the symbol-erasure pattern is such that the maximum time interval τ between successive returns to zero of the information debt function is at most m. The parameter τ also represents the worst-case delay in decoding a message symbol. In the present paper, we study (n,k,m) convolutional codes that possess the analogous property for the case τ > m whenever it is possible to do so. We will refer to such codes as information-debt-optimal streaming (iDOS) codes. We prove the existence of periodically time-varying iDOS codes for all possible {n,k,m,τ} parameters. We also show that m-MDS codes and Maximum Distance Profile convolutional codes are iDOS codes for certain parameter ranges. As a by-product of our existence result, the minimum memory needed for a particular class of streaming codes studied earlier in the literature, is determined.
对于一个(n,k,m)卷积码,其中k为消息符号数,n为编码符号数,m为存储器,Martinian[1]引入了信息债务的概念,其在时刻t的值是解码所有先验消息符号所需的额外编码符号数。同一篇论文表明,存在(n,k,m)卷积码,只要符号擦除模式使得信息债务函数连续返回零之间的最大时间间隔τ不超过m,就可以恢复所有先前的消息符号。参数τ也表示解码消息符号时的最坏情况延迟。在本文中,我们研究了(n,k,m)卷积码在τ > m的情况下,只要有可能,就具有类似的性质。我们将把这样的代码称为信息债务最优流(iDOS)代码。对于所有可能的{n,k,m,τ}参数,我们证明了周期时变iDOS码的存在性。我们还证明了m-MDS码和最大距离轮廓卷积码在一定参数范围内是iDOS码。作为我们存在结果的副产品,确定了先前文献中研究的一类特定流码所需的最小内存。
{"title":"On Information-Debt-Optimal Streaming Codes With Small Memory","authors":"Vinayak Ramkumar, M. Krishnan, Myna Vajha, P. V. Kumar","doi":"10.1109/ISIT50566.2022.9834842","DOIUrl":"https://doi.org/10.1109/ISIT50566.2022.9834842","url":null,"abstract":"In the context of an (n,k,m) convolutional code where k is the number of message symbols, n the number of code symbols and m the memory, Martinian [1] introduced the concept of information debt whose value at time t is the number of additional coded symbols needed to decode all prior message symbols. The same paper shows the existence of (n,k,m) convolutional codes that can recover all prior message symbols whenever the symbol-erasure pattern is such that the maximum time interval τ between successive returns to zero of the information debt function is at most m. The parameter τ also represents the worst-case delay in decoding a message symbol. In the present paper, we study (n,k,m) convolutional codes that possess the analogous property for the case τ > m whenever it is possible to do so. We will refer to such codes as information-debt-optimal streaming (iDOS) codes. We prove the existence of periodically time-varying iDOS codes for all possible {n,k,m,τ} parameters. We also show that m-MDS codes and Maximum Distance Profile convolutional codes are iDOS codes for certain parameter ranges. As a by-product of our existence result, the minimum memory needed for a particular class of streaming codes studied earlier in the literature, is determined.","PeriodicalId":348168,"journal":{"name":"2022 IEEE International Symposium on Information Theory (ISIT)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114653299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Interpreting Deep-Learned Error-Correcting Codes 解释深度学习纠错码
Pub Date : 2022-06-26 DOI: 10.1109/ISIT50566.2022.9834599
N. Devroye, N. Mohammadi, A. Mulgund, H. Naik, R. Shekhar, Gyoergy Turan, Y. Wei, M. Žefran
Deep learning has been used recently to learn error-correcting encoders and decoders which may improve upon previously known codes in certain regimes. The encoders and decoders are learned "black-boxes", and interpreting their behavior is of interest both for further applications and for incorporating this work into coding theory. Understanding these codes provides a compelling case study for Explainable Artificial Intelligence (XAI): since coding theory is a well-developed and quantitative field, the interpretability problems that arise differ from those traditionally considered. We develop post-hoc interpretability techniques to analyze the deep-learned, autoencoder-based encoders of TurboAE-binary codes, using influence heatmaps, mixed integer linear programming (MILP), Fourier analysis, and property testing. We compare the learned, interpretable encoders combined with BCJR decoders to the original black-box code.
深度学习最近被用于学习纠错编码器和解码器,这可能会在某些制度下改进先前已知的代码。编码器和解码器是经过学习的“黑盒”,解释它们的行为对进一步的应用和将这项工作纳入编码理论都很有意义。理解这些代码为可解释的人工智能(XAI)提供了一个引人注目的案例研究:由于编码理论是一个发展良好的定量领域,因此产生的可解释性问题与传统考虑的问题不同。我们利用影响热图、混合整数线性规划(MILP)、傅立叶分析和性能测试,开发了基于自编码器的turbo ae二进制码深度学习编码器的即时可解释性技术。我们将学习的、可解释的编码器与BCJR解码器结合起来与原始的黑箱代码进行比较。
{"title":"Interpreting Deep-Learned Error-Correcting Codes","authors":"N. Devroye, N. Mohammadi, A. Mulgund, H. Naik, R. Shekhar, Gyoergy Turan, Y. Wei, M. Žefran","doi":"10.1109/ISIT50566.2022.9834599","DOIUrl":"https://doi.org/10.1109/ISIT50566.2022.9834599","url":null,"abstract":"Deep learning has been used recently to learn error-correcting encoders and decoders which may improve upon previously known codes in certain regimes. The encoders and decoders are learned \"black-boxes\", and interpreting their behavior is of interest both for further applications and for incorporating this work into coding theory. Understanding these codes provides a compelling case study for Explainable Artificial Intelligence (XAI): since coding theory is a well-developed and quantitative field, the interpretability problems that arise differ from those traditionally considered. We develop post-hoc interpretability techniques to analyze the deep-learned, autoencoder-based encoders of TurboAE-binary codes, using influence heatmaps, mixed integer linear programming (MILP), Fourier analysis, and property testing. We compare the learned, interpretable encoders combined with BCJR decoders to the original black-box code.","PeriodicalId":348168,"journal":{"name":"2022 IEEE International Symposium on Information Theory (ISIT)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117081601","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Fundamental Limits of Personalized Federated Linear Regression with Data Heterogeneity 具有数据异质性的个性化联邦线性回归的基本限制
Pub Date : 2022-06-26 DOI: 10.1109/ISIT50566.2022.9834894
Chun-Ying Hou, I-Hsiang Wang
Federated learning is a nascent framework for collaborative machine learning over networks of devices with local data and local model updates. Data heterogeneity across the devices is one of the challenges confronting this emerging field. Personalization is a natural approach to simultaneously utilize information from the other users’ data and take data heterogeneity into account. In this work, we study the linear regression problem where the data across users are generated from different regression vectors. We present an information-theoretic lower bound of the minimax expected excess risk of personalized linear models. We show an upper bound that matches the lower bound within constant factors. The results characterize the effect of data heterogeneity on learning performance and the trade-off between sample size, problem difficulty, and distribution discrepancy, suggesting that the discrepancy-to-difficulty ratio is the key factor governing the effectiveness of heterogeneous data.
联邦学习是一个新兴的框架,用于在具有本地数据和本地模型更新的设备网络上进行协作机器学习。跨设备的数据异构是这个新兴领域面临的挑战之一。个性化是一种自然的方法,可以同时利用来自其他用户数据的信息并考虑到数据的异质性。在这项工作中,我们研究了线性回归问题,其中跨用户的数据由不同的回归向量生成。给出了个性化线性模型的最小、最大期望超额风险的信息论下界。我们给出了在常数因子范围内与下界匹配的上界。研究结果描述了数据异质性对学习绩效的影响,以及样本量、问题难度和分布差异之间的权衡关系,表明差异与难度比是控制异构数据有效性的关键因素。
{"title":"Fundamental Limits of Personalized Federated Linear Regression with Data Heterogeneity","authors":"Chun-Ying Hou, I-Hsiang Wang","doi":"10.1109/ISIT50566.2022.9834894","DOIUrl":"https://doi.org/10.1109/ISIT50566.2022.9834894","url":null,"abstract":"Federated learning is a nascent framework for collaborative machine learning over networks of devices with local data and local model updates. Data heterogeneity across the devices is one of the challenges confronting this emerging field. Personalization is a natural approach to simultaneously utilize information from the other users’ data and take data heterogeneity into account. In this work, we study the linear regression problem where the data across users are generated from different regression vectors. We present an information-theoretic lower bound of the minimax expected excess risk of personalized linear models. We show an upper bound that matches the lower bound within constant factors. The results characterize the effect of data heterogeneity on learning performance and the trade-off between sample size, problem difficulty, and distribution discrepancy, suggesting that the discrepancy-to-difficulty ratio is the key factor governing the effectiveness of heterogeneous data.","PeriodicalId":348168,"journal":{"name":"2022 IEEE International Symposium on Information Theory (ISIT)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124055626","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Ranking Recovery from Noisy Observations up to a Distortion 从噪声观测到失真程度的排序恢复
Pub Date : 2022-06-26 DOI: 10.1109/ISIT50566.2022.9834780
Minoh Jeong, Martina Cardone, Alex Dytso
This paper considers the problem of recovering the ranking of a data vector from noisy observations, up to a distortion. Specifically, the noisy observations consist of the original data vector corrupted by isotropic additive Gaussian noise, and the distortion is measured in terms of a distance function between the estimated ranking and the true ranking of the original data vector. First, it is shown that an optimal (in terms of error probability) decision rule for the estimation task simply outputs the ranking of the noisy observation. Then, the error probability incurred by such a decision rule is characterized in the low-noise regime, and shown to grow sublinearly with the noise standard deviation. This result highlights that the proposed approximate version of the ranking recovery problem is significantly less noise-dominated than the exact recovery considered in [Jeong, ISIT 2021].
本文考虑了从噪声观测中恢复数据向量的排序问题,直至失真。具体来说,噪声观测由被各向同性加性高斯噪声破坏的原始数据向量组成,并且畸变是根据原始数据向量的估计排名与真实排名之间的距离函数来测量的。首先,证明了估计任务的最优决策规则(就错误概率而言)只是输出噪声观测值的排序。在低噪声条件下,该决策规则的误差概率随噪声标准差呈次线性增长。该结果突出表明,与[Jeong, ISIT 2021]中考虑的精确采收率相比,所提出的近似版本的排序采收率问题的噪声占主导地位要小得多。
{"title":"On the Ranking Recovery from Noisy Observations up to a Distortion","authors":"Minoh Jeong, Martina Cardone, Alex Dytso","doi":"10.1109/ISIT50566.2022.9834780","DOIUrl":"https://doi.org/10.1109/ISIT50566.2022.9834780","url":null,"abstract":"This paper considers the problem of recovering the ranking of a data vector from noisy observations, up to a distortion. Specifically, the noisy observations consist of the original data vector corrupted by isotropic additive Gaussian noise, and the distortion is measured in terms of a distance function between the estimated ranking and the true ranking of the original data vector. First, it is shown that an optimal (in terms of error probability) decision rule for the estimation task simply outputs the ranking of the noisy observation. Then, the error probability incurred by such a decision rule is characterized in the low-noise regime, and shown to grow sublinearly with the noise standard deviation. This result highlights that the proposed approximate version of the ranking recovery problem is significantly less noise-dominated than the exact recovery considered in [Jeong, ISIT 2021].","PeriodicalId":348168,"journal":{"name":"2022 IEEE International Symposium on Information Theory (ISIT)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125997084","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Algebraic Chase Decoding of Elliptic Codes Through Computing the Gröbner Basis 通过计算Gröbner基实现椭圆码的代数追码
Pub Date : 2022-06-26 DOI: 10.1109/ISIT50566.2022.9834889
Wan, Li Chen, Fangguo Zhang
This paper proposes two interpolation-based algebraic Chase decoding for elliptic codes. It is introduced from the perspective of computing the Gröbner basis of the interpolation module, for which two Chase interpolation approaches are utilized. They are Kötter’s interpolation and the basis reduction (BR) interpolation. By identifying η unreliable symbols, 2η decoding test-vectors are formulated, and the corresponding interpolation modules can be defined. The re-encoding further helps transform the test-vectors, facilitating the two interpolation techniques. In particular, Kötter’s interpolation is performed for the common elements of the test-vectors, producing an intermediate outcome that is shared by the decoding of all test-vectors. The desired Gröbner bases w.r.t. all test-vectors can be obtained in a binary tree growing fashion, leading to a low complexity but its decoding latency cannot be contained. In contrast, the BR interpolation first performs the common computation in basis construction which is shared by all interpolation modules, and then conducts the module basis construction and reduction for all test-vectors in parallel. It results in a significantly lower decoding latency. Finally, simulation results are also presented to demonstrate the effectiveness of the proposed Chase decoding.
提出了两种基于插值的椭圆码代数Chase译码方法。从计算插补模块Gröbner基的角度进行了介绍,其中采用了两种Chase插补方法。它们是Kötter插值和基约简(BR)插值。通过识别η不可靠符号,推导出2η译码测试向量,并定义相应的插值模块。重新编码进一步帮助转换测试向量,促进两种插值技术。特别是,Kötter的插值是针对测试向量的公共元素执行的,从而产生一个中间结果,该结果由所有测试向量的解码共享。所需的Gröbner碱基w.r.t.所有测试向量都可以用二叉树生长的方式获得,导致较低的复杂性,但其解码延迟无法控制。相比之下,BR插值首先进行所有插值模块共享的基构造公共计算,然后并行地对所有测试向量进行模块基构造和约简。它可以显著降低解码延迟。最后给出了仿真结果,验证了所提出的Chase解码方法的有效性。
{"title":"Algebraic Chase Decoding of Elliptic Codes Through Computing the Gröbner Basis","authors":"Wan, Li Chen, Fangguo Zhang","doi":"10.1109/ISIT50566.2022.9834889","DOIUrl":"https://doi.org/10.1109/ISIT50566.2022.9834889","url":null,"abstract":"This paper proposes two interpolation-based algebraic Chase decoding for elliptic codes. It is introduced from the perspective of computing the Gröbner basis of the interpolation module, for which two Chase interpolation approaches are utilized. They are Kötter’s interpolation and the basis reduction (BR) interpolation. By identifying η unreliable symbols, 2η decoding test-vectors are formulated, and the corresponding interpolation modules can be defined. The re-encoding further helps transform the test-vectors, facilitating the two interpolation techniques. In particular, Kötter’s interpolation is performed for the common elements of the test-vectors, producing an intermediate outcome that is shared by the decoding of all test-vectors. The desired Gröbner bases w.r.t. all test-vectors can be obtained in a binary tree growing fashion, leading to a low complexity but its decoding latency cannot be contained. In contrast, the BR interpolation first performs the common computation in basis construction which is shared by all interpolation modules, and then conducts the module basis construction and reduction for all test-vectors in parallel. It results in a significantly lower decoding latency. Finally, simulation results are also presented to demonstrate the effectiveness of the proposed Chase decoding.","PeriodicalId":348168,"journal":{"name":"2022 IEEE International Symposium on Information Theory (ISIT)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129807707","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2022 IEEE International Symposium on Information Theory (ISIT)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1