Pub Date : 2024-11-11DOI: 10.1109/JSAIT.2024.3496197
Yasutada Oohama
We consider the one helper source coding problem posed and investigated by Ahlswede, Körner, and Wyner for a class of information sources with memory. For this class of information sources we give explicit inner and outer bounds of the admissible rate region. We also give a certain nontrivial class of information sources where the inner and outer bounds match.
{"title":"Source Coding for Markov Sources With Partial Memoryless Side Information at the Decoder","authors":"Yasutada Oohama","doi":"10.1109/JSAIT.2024.3496197","DOIUrl":"https://doi.org/10.1109/JSAIT.2024.3496197","url":null,"abstract":"We consider the one helper source coding problem posed and investigated by Ahlswede, Körner, and Wyner for a class of information sources with memory. For this class of information sources we give explicit inner and outer bounds of the admissible rate region. We also give a certain nontrivial class of information sources where the inner and outer bounds match.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"675-693"},"PeriodicalIF":0.0,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10750312","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142880454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-31DOI: 10.1109/JSAIT.2024.3487856
Yichen Huang
In a spin chain governed by a local Hamiltonian, we consider a microcanonical ensemble in the middle of the energy spectrum and a contiguous subsystem whose length is a constant fraction of the system size. We prove that if the bandwidth of the ensemble is greater than a certain constant, then the average entanglement entropy (between the subsystem and the rest of the system) of eigenstates in the ensemble deviates from the maximum entropy by at least a positive constant. This result highlights the difference between the entanglement entropy of mid-spectrum eigenstates of (chaotic) local Hamiltonians and that of random states. We also prove that the former deviates from the thermodynamic entropy at the same energy by at least a positive constant.
{"title":"Deviation From Maximal Entanglement for Mid-Spectrum Eigenstates of Local Hamiltonians","authors":"Yichen Huang","doi":"10.1109/JSAIT.2024.3487856","DOIUrl":"https://doi.org/10.1109/JSAIT.2024.3487856","url":null,"abstract":"In a spin chain governed by a local Hamiltonian, we consider a microcanonical ensemble in the middle of the energy spectrum and a contiguous subsystem whose length is a constant fraction of the system size. We prove that if the bandwidth of the ensemble is greater than a certain constant, then the average entanglement entropy (between the subsystem and the rest of the system) of eigenstates in the ensemble deviates from the maximum entropy by at least a positive constant. This result highlights the difference between the entanglement entropy of mid-spectrum eigenstates of (chaotic) local Hamiltonians and that of random states. We also prove that the former deviates from the thermodynamic entropy at the same energy by at least a positive constant.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"694-701"},"PeriodicalIF":0.0,"publicationDate":"2024-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142825942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-14DOI: 10.1109/JSAIT.2024.3481296
Tomer Berg;Or Ordentlich;Ofer Shayevitz
The problem of statistical inference in its various forms has been the subject of decades-long extensive research. Most of the effort has been focused on characterizing the behavior as a function of the number of available samples, with far less attention given to the effect of memory limitations on performance. Recently, this latter topic has drawn much interest in the engineering and computer science literature. In this survey paper, we attempt to review the state-of-the-art of statistical inference under memory constraints in several canonical problems, including hypothesis testing, parameter estimation, and distribution property testing/estimation. We discuss the main results in this developing field, and by identifying recurrent themes, we extract some fundamental building blocks for algorithmic construction, as well as useful techniques for lower bound derivations.
{"title":"Statistical Inference With Limited Memory: A Survey","authors":"Tomer Berg;Or Ordentlich;Ofer Shayevitz","doi":"10.1109/JSAIT.2024.3481296","DOIUrl":"https://doi.org/10.1109/JSAIT.2024.3481296","url":null,"abstract":"The problem of statistical inference in its various forms has been the subject of decades-long extensive research. Most of the effort has been focused on characterizing the behavior as a function of the number of available samples, with far less attention given to the effect of memory limitations on performance. Recently, this latter topic has drawn much interest in the engineering and computer science literature. In this survey paper, we attempt to review the state-of-the-art of statistical inference under memory constraints in several canonical problems, including hypothesis testing, parameter estimation, and distribution property testing/estimation. We discuss the main results in this developing field, and by identifying recurrent themes, we extract some fundamental building blocks for algorithmic construction, as well as useful techniques for lower bound derivations.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"5 ","pages":"623-644"},"PeriodicalIF":0.0,"publicationDate":"2024-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142595119","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-30DOI: 10.1109/JSAIT.2024.3469929
Michael G. Jabbour;Nilanjana Datta
Uniform continuity bounds on entropies are generally expressed in terms of a single distance measure between probability distributions or quantum states, typically, the total variation- or trace distance. However, if an additional distance measure is known, the continuity bounds can be significantly strengthened. Here, we prove a tight uniform continuity bound for the Shannon entropy in terms of both the local- and total variation distances, sharpening an inequality in (Sason, 2013). We also obtain a uniform continuity bound for the von Neumann entropy in terms of both the operator norm- and trace distances. We then apply our results to compute upper bounds on channel capacities. We first refine the concept of approximate degradable channels by introducing $(varepsilon ,nu)$