{"title":"Missing g-Mass: Investigating the Missing Parts of Distributions","authors":"Prafulla Chandra;Andrew Thangaraj","doi":"10.1109/TIT.2024.3440661","DOIUrl":null,"url":null,"abstract":"Estimating the underlying distribution from iid samples is a classical and important problem in statistics. When the alphabet size is large compared to number of samples, a portion of the distribution is highly likely to be unobserved or sparsely observed. The missing mass, defined as the sum of probabilities \n<inline-formula> <tex-math>$\\Pr (x)$ </tex-math></inline-formula>\n over the missing letters x, and the Good-Turing estimator for missing mass have been important tools in large-alphabet distribution estimation. In this article, given a positive function g from \n<inline-formula> <tex-math>$[{0,1}]$ </tex-math></inline-formula>\n to the reals, the missing g-mass, defined as the sum of \n<inline-formula> <tex-math>$g(\\Pr (x))$ </tex-math></inline-formula>\n over the missing letters x, is introduced and studied. The missing g-mass can be used to investigate the structure of the missing part of the distribution. Specific applications for special cases such as order-\n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n missing mass (\n<inline-formula> <tex-math>$g(p)=p^{\\alpha }$ </tex-math></inline-formula>\n) and the missing Shannon entropy (\n<inline-formula> <tex-math>$g(p)=-p\\log p$ </tex-math></inline-formula>\n) include estimating distance from uniformity of the missing distribution and its partial estimation. Minimax estimation is studied for order-\n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n missing mass for integer values of \n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n and exact minimax convergence rates are obtained. Concentration is studied for a class of functions g and specific results are derived for order-\n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n missing mass and missing Shannon entropy. Sub-Gaussian tail bounds with near-optimal worst-case variance factors are derived. Two new notions of concentration, named strongly sub-Gamma and filtered sub-Gaussian concentration, are introduced and shown to result in right tail bounds that are better than those obtained from sub-Gaussian concentration.","PeriodicalId":13494,"journal":{"name":"IEEE Transactions on Information Theory","volume":"70 10","pages":"7049-7065"},"PeriodicalIF":2.2000,"publicationDate":"2024-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Information Theory","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10630859/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Estimating the underlying distribution from iid samples is a classical and important problem in statistics. When the alphabet size is large compared to number of samples, a portion of the distribution is highly likely to be unobserved or sparsely observed. The missing mass, defined as the sum of probabilities
$\Pr (x)$
over the missing letters x, and the Good-Turing estimator for missing mass have been important tools in large-alphabet distribution estimation. In this article, given a positive function g from
$[{0,1}]$
to the reals, the missing g-mass, defined as the sum of
$g(\Pr (x))$
over the missing letters x, is introduced and studied. The missing g-mass can be used to investigate the structure of the missing part of the distribution. Specific applications for special cases such as order-
$\alpha $
missing mass (
$g(p)=p^{\alpha }$
) and the missing Shannon entropy (
$g(p)=-p\log p$
) include estimating distance from uniformity of the missing distribution and its partial estimation. Minimax estimation is studied for order-
$\alpha $
missing mass for integer values of
$\alpha $
and exact minimax convergence rates are obtained. Concentration is studied for a class of functions g and specific results are derived for order-
$\alpha $
missing mass and missing Shannon entropy. Sub-Gaussian tail bounds with near-optimal worst-case variance factors are derived. Two new notions of concentration, named strongly sub-Gamma and filtered sub-Gaussian concentration, are introduced and shown to result in right tail bounds that are better than those obtained from sub-Gaussian concentration.
期刊介绍:
The IEEE Transactions on Information Theory is a journal that publishes theoretical and experimental papers concerned with the transmission, processing, and utilization of information. The boundaries of acceptable subject matter are intentionally not sharply delimited. Rather, it is hoped that as the focus of research activity changes, a flexible policy will permit this Transactions to follow suit. Current appropriate topics are best reflected by recent Tables of Contents; they are summarized in the titles of editorial areas that appear on the inside front cover.