{"title":"Information Theory and Statistics: A Tutorial","authors":"I. Csiszár, P. Shields","doi":"10.1561/0100000004","DOIUrl":null,"url":null,"abstract":"This tutorial is concerned with applications of information theory concepts in statistics, in the finite alphabet setting. The information measure known as information divergence or Kullback-Leibler distance or relative entropy plays a key role, often with a geometric flavor as an analogue of squared Euclidean distance, as in the concepts of I-projection, I-radius and I-centroid. The topics covered include large deviations, hypothesis testing, maximum likelihood estimation in exponential families, analysis of contingency tables, and iterative algorithms with an \"information geometry\" background. Also, an introduction is provided to the theory of universal coding, and to statistical inference via the minimum description length principle motivated by that theory.","PeriodicalId":45236,"journal":{"name":"Foundations and Trends in Communications and Information Theory","volume":"23 1","pages":""},"PeriodicalIF":2.0000,"publicationDate":"2004-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"758","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Foundations and Trends in Communications and Information Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1561/0100000004","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 758
Abstract
This tutorial is concerned with applications of information theory concepts in statistics, in the finite alphabet setting. The information measure known as information divergence or Kullback-Leibler distance or relative entropy plays a key role, often with a geometric flavor as an analogue of squared Euclidean distance, as in the concepts of I-projection, I-radius and I-centroid. The topics covered include large deviations, hypothesis testing, maximum likelihood estimation in exponential families, analysis of contingency tables, and iterative algorithms with an "information geometry" background. Also, an introduction is provided to the theory of universal coding, and to statistical inference via the minimum description length principle motivated by that theory.
期刊介绍:
Foundations and Trends® in Communications and Information Theory publishes survey and tutorial articles in the following topics: - Coded modulation - Coding theory and practice - Communication complexity - Communication system design - Cryptology and data security - Data compression - Data networks - Demodulation and Equalization - Denoising - Detection and estimation - Information theory and statistics - Information theory and computer science - Joint source/channel coding - Modulation and signal design - Multiuser detection - Multiuser information theory