{"title":"缺失的 g-质量调查分布的缺失部分","authors":"Prafulla Chandra;Andrew Thangaraj","doi":"10.1109/TIT.2024.3440661","DOIUrl":null,"url":null,"abstract":"Estimating the underlying distribution from iid samples is a classical and important problem in statistics. When the alphabet size is large compared to number of samples, a portion of the distribution is highly likely to be unobserved or sparsely observed. The missing mass, defined as the sum of probabilities \n<inline-formula> <tex-math>$\\Pr (x)$ </tex-math></inline-formula>\n over the missing letters x, and the Good-Turing estimator for missing mass have been important tools in large-alphabet distribution estimation. In this article, given a positive function g from \n<inline-formula> <tex-math>$[{0,1}]$ </tex-math></inline-formula>\n to the reals, the missing g-mass, defined as the sum of \n<inline-formula> <tex-math>$g(\\Pr (x))$ </tex-math></inline-formula>\n over the missing letters x, is introduced and studied. The missing g-mass can be used to investigate the structure of the missing part of the distribution. Specific applications for special cases such as order-\n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n missing mass (\n<inline-formula> <tex-math>$g(p)=p^{\\alpha }$ </tex-math></inline-formula>\n) and the missing Shannon entropy (\n<inline-formula> <tex-math>$g(p)=-p\\log p$ </tex-math></inline-formula>\n) include estimating distance from uniformity of the missing distribution and its partial estimation. Minimax estimation is studied for order-\n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n missing mass for integer values of \n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n and exact minimax convergence rates are obtained. Concentration is studied for a class of functions g and specific results are derived for order-\n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n missing mass and missing Shannon entropy. Sub-Gaussian tail bounds with near-optimal worst-case variance factors are derived. Two new notions of concentration, named strongly sub-Gamma and filtered sub-Gaussian concentration, are introduced and shown to result in right tail bounds that are better than those obtained from sub-Gaussian concentration.","PeriodicalId":13494,"journal":{"name":"IEEE Transactions on Information Theory","volume":"70 10","pages":"7049-7065"},"PeriodicalIF":2.2000,"publicationDate":"2024-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Missing g-Mass: Investigating the Missing Parts of Distributions\",\"authors\":\"Prafulla Chandra;Andrew Thangaraj\",\"doi\":\"10.1109/TIT.2024.3440661\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Estimating the underlying distribution from iid samples is a classical and important problem in statistics. When the alphabet size is large compared to number of samples, a portion of the distribution is highly likely to be unobserved or sparsely observed. The missing mass, defined as the sum of probabilities \\n<inline-formula> <tex-math>$\\\\Pr (x)$ </tex-math></inline-formula>\\n over the missing letters x, and the Good-Turing estimator for missing mass have been important tools in large-alphabet distribution estimation. In this article, given a positive function g from \\n<inline-formula> <tex-math>$[{0,1}]$ </tex-math></inline-formula>\\n to the reals, the missing g-mass, defined as the sum of \\n<inline-formula> <tex-math>$g(\\\\Pr (x))$ </tex-math></inline-formula>\\n over the missing letters x, is introduced and studied. The missing g-mass can be used to investigate the structure of the missing part of the distribution. Specific applications for special cases such as order-\\n<inline-formula> <tex-math>$\\\\alpha $ </tex-math></inline-formula>\\n missing mass (\\n<inline-formula> <tex-math>$g(p)=p^{\\\\alpha }$ </tex-math></inline-formula>\\n) and the missing Shannon entropy (\\n<inline-formula> <tex-math>$g(p)=-p\\\\log p$ </tex-math></inline-formula>\\n) include estimating distance from uniformity of the missing distribution and its partial estimation. Minimax estimation is studied for order-\\n<inline-formula> <tex-math>$\\\\alpha $ </tex-math></inline-formula>\\n missing mass for integer values of \\n<inline-formula> <tex-math>$\\\\alpha $ </tex-math></inline-formula>\\n and exact minimax convergence rates are obtained. Concentration is studied for a class of functions g and specific results are derived for order-\\n<inline-formula> <tex-math>$\\\\alpha $ </tex-math></inline-formula>\\n missing mass and missing Shannon entropy. Sub-Gaussian tail bounds with near-optimal worst-case variance factors are derived. Two new notions of concentration, named strongly sub-Gamma and filtered sub-Gaussian concentration, are introduced and shown to result in right tail bounds that are better than those obtained from sub-Gaussian concentration.\",\"PeriodicalId\":13494,\"journal\":{\"name\":\"IEEE Transactions on Information Theory\",\"volume\":\"70 10\",\"pages\":\"7049-7065\"},\"PeriodicalIF\":2.2000,\"publicationDate\":\"2024-08-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Information Theory\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10630859/\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Information Theory","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10630859/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
摘要
从 iid 样本中估计基本分布是统计学中一个经典而重要的问题。当字母表的大小与样本数量相比较大时,分布的一部分极有可能未被观测到或观测稀少。缺失质量定义为缺失字母 x 的概率总和 $\Pr (x)$,缺失质量的 Good-Turing 估计器一直是大字母分布估计的重要工具。本文介绍并研究了从$[{0,1}]$到实数的正函数g的缺失g-质量,其定义为缺失字母x上的$g(\Pr (x))$之和。缺失 g 质量可用于研究分布中缺失部分的结构。特殊情况下的具体应用,如阶- $\alpha $缺失质量($g(p)=p^{\alpha }$)和缺失香农熵($g(p)=-p\log p$),包括估计缺失分布的均匀性距离及其部分估计。针对 $\alpha $ 的整数值,研究了阶 $\alpha $ 缺失质量的最小估计,并获得了精确的最小收敛率。研究了一类函数 g 的集中性,并得出了阶(order- $\alpha $)缺失质量和缺失香农熵的具体结果。推导出了接近最优最坏情况方差系数的亚高斯尾边界。引入了两个新的集中概念,分别称为强亚伽马集中和滤波亚高斯集中,并证明这两个概念能得到比亚高斯集中更好的右尾边界。
Missing g-Mass: Investigating the Missing Parts of Distributions
Estimating the underlying distribution from iid samples is a classical and important problem in statistics. When the alphabet size is large compared to number of samples, a portion of the distribution is highly likely to be unobserved or sparsely observed. The missing mass, defined as the sum of probabilities
$\Pr (x)$
over the missing letters x, and the Good-Turing estimator for missing mass have been important tools in large-alphabet distribution estimation. In this article, given a positive function g from
$[{0,1}]$
to the reals, the missing g-mass, defined as the sum of
$g(\Pr (x))$
over the missing letters x, is introduced and studied. The missing g-mass can be used to investigate the structure of the missing part of the distribution. Specific applications for special cases such as order-
$\alpha $
missing mass (
$g(p)=p^{\alpha }$
) and the missing Shannon entropy (
$g(p)=-p\log p$
) include estimating distance from uniformity of the missing distribution and its partial estimation. Minimax estimation is studied for order-
$\alpha $
missing mass for integer values of
$\alpha $
and exact minimax convergence rates are obtained. Concentration is studied for a class of functions g and specific results are derived for order-
$\alpha $
missing mass and missing Shannon entropy. Sub-Gaussian tail bounds with near-optimal worst-case variance factors are derived. Two new notions of concentration, named strongly sub-Gamma and filtered sub-Gaussian concentration, are introduced and shown to result in right tail bounds that are better than those obtained from sub-Gaussian concentration.
期刊介绍:
The IEEE Transactions on Information Theory is a journal that publishes theoretical and experimental papers concerned with the transmission, processing, and utilization of information. The boundaries of acceptable subject matter are intentionally not sharply delimited. Rather, it is hoped that as the focus of research activity changes, a flexible policy will permit this Transactions to follow suit. Current appropriate topics are best reflected by recent Tables of Contents; they are summarized in the titles of editorial areas that appear on the inside front cover.