{"title":"A New Objective Measure Of Signal Complexity Using Bayesian Inference","authors":"A. Quinn","doi":"10.1109/SSAP.1994.572444","DOIUrl":null,"url":null,"abstract":"An objective Ockham prior which penalizes complexity in parametric signal hypotheses is derived from Bayesian fundamentals. Novel quantitative definitions of complexity are deduced under the procedure. This improves on current variants of the coding theoretic Minimum Message Length (MML) criterion where complexity definitions are imposed as heuristics. It is shown that the Ockham prior arises naturally in marginal Bayesian inference, but is excluded if joint inference is adopted. 1. I N T R O D U C T I O N : BAYESIAN SYSTEMATIC HYPOTHESES The Signal Identification problem arises whenever a systematic hypothesis is adopted to explain an observed data set. Let d = (d l , . . . , d ~ ) ~ be a finite set of one-dimensional observations. The prior hypothesis, 2, asserts that 2. O C K H A M ’ S R A Z O R Systematic hypotheses (1) must be assessed in the context of Ockham’s Razor (i.e. the Desideratum of Simplicity) [3] which, for the purposes of time series analysis, states that randomness must not be fitted with determinism. Consider a set of hypotheses, 21~12, . . ., parameterized with p.p. sets, 0 1 , 0 2 , . . . respectively. The likelihood function (LF) for the kth hypothesis is 1(& I d, 2,) = p(d I &, I k ) . The classical approach is to identify a model for d by maximizing l ( . ) . The resulting point inference, arg . sup, . supel I(& I d, lk), identifies both a mode1 (solving the Model Selection problem) and a set of parameter values (solving the Parameter Estimation problem). In open-ended inference, we may posit models with increasing numbers of degrees of heedom (e.g. increasing model order). In this manner, llell is reduced (l), 11 . 11 being the norm implied by the p.d.f. of e [l]. Ultimately, the likelihood-based inference machine fails because of its insensitivity to Ockham’s Razor. d = s + e 2.1. Subjec t ive vs. Objec t ive Complexi ty s # 0 is the unknown deterministic component (i.e. the ‘signal’) and e is the vector of unknown, non-systematic residuals. If the hypothesis is parametric, then s = s(8). The validity of the decomposition (1) must be assessed for a particular d, since all subsequent steps in the inference taskmodel selection, parameter estimationdepend upon it. 1.1. Probabi l i ty as Belief Calculus The fundamental concept underlying the Bayesian Paradigm is that the beliefs associated with inductive inference are uniquely quantified as probabilities and are consistently manipulated using the Probability Calculus [l]. The equivalence of the Belief Calculus and the Probability Calculus has been deduced from fundamentals [’I. Any unknowneither fixed or random-in a hypothesis has a domain, R, of possible values. The distribution of beliefs across R is expressed by a p.d.f. This constitutes tlie definition of a probabilistic parameter (p.p.) [ I ] which is tlie appropriate Bayesian extension of the random variable (r.v.) concept of orthodox inference. The two definitions merge when the uiikiiown is i~iherently random. The problem of admitting Ockham to an inference tends to be treated in two distinct stages [3-71: (i) an acceptable (quantifiable) definition of complexity is proposed; (ii) an inference procedure which includes a monotonically decreasing function of this complexity is adopted. Jeffreys [3] sought to specify the rate of ezchange between simplicity and closeness of fit to observations. His ‘Simplicity Postulate” stated that ‘simpler laws have the greater prior probabilities”. This subjective approach overlooks the prospect of engendering the principle objectively.","PeriodicalId":151571,"journal":{"name":"IEEE Seventh SP Workshop on Statistical Signal and Array Processing","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"1994-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Seventh SP Workshop on Statistical Signal and Array Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SSAP.1994.572444","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
An objective Ockham prior which penalizes complexity in parametric signal hypotheses is derived from Bayesian fundamentals. Novel quantitative definitions of complexity are deduced under the procedure. This improves on current variants of the coding theoretic Minimum Message Length (MML) criterion where complexity definitions are imposed as heuristics. It is shown that the Ockham prior arises naturally in marginal Bayesian inference, but is excluded if joint inference is adopted. 1. I N T R O D U C T I O N : BAYESIAN SYSTEMATIC HYPOTHESES The Signal Identification problem arises whenever a systematic hypothesis is adopted to explain an observed data set. Let d = (d l , . . . , d ~ ) ~ be a finite set of one-dimensional observations. The prior hypothesis, 2, asserts that 2. O C K H A M ’ S R A Z O R Systematic hypotheses (1) must be assessed in the context of Ockham’s Razor (i.e. the Desideratum of Simplicity) [3] which, for the purposes of time series analysis, states that randomness must not be fitted with determinism. Consider a set of hypotheses, 21~12, . . ., parameterized with p.p. sets, 0 1 , 0 2 , . . . respectively. The likelihood function (LF) for the kth hypothesis is 1(& I d, 2,) = p(d I &, I k ) . The classical approach is to identify a model for d by maximizing l ( . ) . The resulting point inference, arg . sup, . supel I(& I d, lk), identifies both a mode1 (solving the Model Selection problem) and a set of parameter values (solving the Parameter Estimation problem). In open-ended inference, we may posit models with increasing numbers of degrees of heedom (e.g. increasing model order). In this manner, llell is reduced (l), 11 . 11 being the norm implied by the p.d.f. of e [l]. Ultimately, the likelihood-based inference machine fails because of its insensitivity to Ockham’s Razor. d = s + e 2.1. Subjec t ive vs. Objec t ive Complexi ty s # 0 is the unknown deterministic component (i.e. the ‘signal’) and e is the vector of unknown, non-systematic residuals. If the hypothesis is parametric, then s = s(8). The validity of the decomposition (1) must be assessed for a particular d, since all subsequent steps in the inference taskmodel selection, parameter estimationdepend upon it. 1.1. Probabi l i ty as Belief Calculus The fundamental concept underlying the Bayesian Paradigm is that the beliefs associated with inductive inference are uniquely quantified as probabilities and are consistently manipulated using the Probability Calculus [l]. The equivalence of the Belief Calculus and the Probability Calculus has been deduced from fundamentals [’I. Any unknowneither fixed or random-in a hypothesis has a domain, R, of possible values. The distribution of beliefs across R is expressed by a p.d.f. This constitutes tlie definition of a probabilistic parameter (p.p.) [ I ] which is tlie appropriate Bayesian extension of the random variable (r.v.) concept of orthodox inference. The two definitions merge when the uiikiiown is i~iherently random. The problem of admitting Ockham to an inference tends to be treated in two distinct stages [3-71: (i) an acceptable (quantifiable) definition of complexity is proposed; (ii) an inference procedure which includes a monotonically decreasing function of this complexity is adopted. Jeffreys [3] sought to specify the rate of ezchange between simplicity and closeness of fit to observations. His ‘Simplicity Postulate” stated that ‘simpler laws have the greater prior probabilities”. This subjective approach overlooks the prospect of engendering the principle objectively.
一个客观的奥卡姆先验,惩罚复杂性的参数信号假设是由贝叶斯基础推导出来的。在此过程中推导出新的复杂性的定量定义。这改进了编码理论最小消息长度(MML)标准的当前变体,其中复杂性定义作为启发式强加。结果表明,在边际贝叶斯推理中,奥卡姆先验是自然产生的,但在联合推理中,奥卡姆先验是不存在的。1. 当采用系统假设来解释观测数据集时,信号识别问题就出现了。令d = (d1,…), d ~) ~是一维观测值的有限集合。先验假设2表明。系统假设(1)必须在奥卡姆剃刀(即简单的愿望)[3]的背景下进行评估,为了时间序列分析的目的,它表明随机性不能与决定论相适应。考虑一组假设,21~12,…,用p.p.集参数化,0,1,0,2,…。分别。第k个假设的似然函数(LF)是1(& I d, 2,) = p(d I &, I k)。经典的方法是通过最大化l()来确定d的模型。. 由此产生的点推理,如。吃晚饭。supel I(& I d,如)标识一个模型1(解决模型选择问题)和一组参数值(解决参数估计问题)。在开放式推理中,我们可以假设自由度不断增加的模型(例如,模型阶数不断增加)。以这种方式,ll被还原为(1),11。11为e的P.D.F.所隐含的规范[1]。最终,基于可能性的推理机失败了,因为它对奥卡姆剃刀不敏感。D = s + e主体与客体之间的关系复杂度为0是未知的确定性成分(即“信号”),e是未知的、非系统残差的向量。如果假设是参数化的,则s = s(8)。分解(1)的有效性必须对特定的d进行评估,因为推理任务模型选择、参数估计的所有后续步骤都依赖于它。1.1. 贝叶斯范式的基本概念是,与归纳推理相关的信念被唯一地量化为概率,并始终使用概率演算进行操作[1]。从基本原理[1]推导出了信念演算与概率演算的等价性。假设中的任何未知数,无论是固定的还是随机的,都有一个可能值的域R。信念在R上的分布由p.d.f表示。这构成了概率参数(p.p.)的定义[I],这是正统推理的随机变量(r.v.)概念的适当贝叶斯扩展。这两种定义合并在一起时,其本身是随机的。承认奥卡姆推理的问题往往分为两个不同的阶段来处理[3-71]:(i)提出一个可接受的(可量化的)复杂性定义;(ii)采用包含该复杂度的单调递减函数的推理过程。Jeffreys[3]试图确定简单性和拟合接近度之间的变化率。他的“简单性假设”指出,“更简单的定律有更大的先验概率”。这种主观的做法忽视了客观地产生原则的前景。