首页 > 最新文献

Proceedings of 1994 Workshop on Information Theory and Statistics最新文献

英文 中文
Markov random fields on graphs for natural languages 自然语言图上的马尔可夫随机场
Pub Date : 1994-10-27 DOI: 10.1109/WITS.1994.513880
J. O’Sullivan, K. Mark, M. Miller
The use of model-based methods for data compression for English dates back at least to Shannon's Markov chain (n-gram) models, where the probability of the next word given all previous words equals the probability of the next word given the previous n-1 words. A second approach seeks to model the hierarchical nature of language via tree graph structures arising from a context-free language (CFL). Neither the n-gram nor the CFL models approach the data compression predicted by the entropy of English as estimated by Shannon and Cover and King. This paper presents two models that incorporate the benefits of both the n-gram model and the tree-based models. In either case the neighborhood structure on the syntactic variables is determined by the tree while the neighborhood structure of the words is determined by the n-gram and the parent syntactic variable (preterminal) in the tree, Having both types of neighbors for the words should yield decreased entropy of the model and hence fewer bits per word in data compression. To motivate estimation of model parameters, some results in estimating parameters for random branching processes is reviewed.
基于模型的英语数据压缩方法的使用至少可以追溯到Shannon的马尔可夫链(n-gram)模型,在该模型中,给定前面所有单词的下一个单词的概率等于给定前面n-1个单词的下一个单词的概率。第二种方法寻求通过由上下文无关语言(CFL)产生的树状图结构来建模语言的层次本质。n-gram模型和CFL模型都无法达到Shannon、Cover和King所估计的英语熵所预测的数据压缩。本文提出了两个模型,结合了n-gram模型和基于树的模型的优点。在任何一种情况下,句法变量的邻域结构都是由树决定的,而单词的邻域结构是由n-gram和树中的父句法变量(前缀)决定的。拥有两种类型的单词邻居应该会降低模型的熵,从而减少数据压缩中每个单词的比特数。为了激励模型参数的估计,综述了随机分支过程参数估计的一些结果。
{"title":"Markov random fields on graphs for natural languages","authors":"J. O’Sullivan, K. Mark, M. Miller","doi":"10.1109/WITS.1994.513880","DOIUrl":"https://doi.org/10.1109/WITS.1994.513880","url":null,"abstract":"The use of model-based methods for data compression for English dates back at least to Shannon's Markov chain (n-gram) models, where the probability of the next word given all previous words equals the probability of the next word given the previous n-1 words. A second approach seeks to model the hierarchical nature of language via tree graph structures arising from a context-free language (CFL). Neither the n-gram nor the CFL models approach the data compression predicted by the entropy of English as estimated by Shannon and Cover and King. This paper presents two models that incorporate the benefits of both the n-gram model and the tree-based models. In either case the neighborhood structure on the syntactic variables is determined by the tree while the neighborhood structure of the words is determined by the n-gram and the parent syntactic variable (preterminal) in the tree, Having both types of neighbors for the words should yield decreased entropy of the model and hence fewer bits per word in data compression. To motivate estimation of model parameters, some results in estimating parameters for random branching processes is reviewed.","PeriodicalId":423518,"journal":{"name":"Proceedings of 1994 Workshop on Information Theory and Statistics","volume":"97 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127157496","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Consistency and rates of convergence of k/sub n/ nearest neighbor estimation under arbitrary sampling 任意抽样下k/sub n/最近邻估计的一致性和收敛速度
Pub Date : 1994-10-27 DOI: 10.1109/WITS.1994.513901
S. Posner, S. R. Kulkarni
Consistency and rates of convergence of the k/sub n/-NN estimator are established in the general case in which samples are chosen arbitrarily from a compact metric space.
在从紧度量空间中任意选取样本的一般情况下,建立了k/sub n/-NN估计量的一致性和收敛率。
{"title":"Consistency and rates of convergence of k/sub n/ nearest neighbor estimation under arbitrary sampling","authors":"S. Posner, S. R. Kulkarni","doi":"10.1109/WITS.1994.513901","DOIUrl":"https://doi.org/10.1109/WITS.1994.513901","url":null,"abstract":"Consistency and rates of convergence of the k/sub n/-NN estimator are established in the general case in which samples are chosen arbitrarily from a compact metric space.","PeriodicalId":423518,"journal":{"name":"Proceedings of 1994 Workshop on Information Theory and Statistics","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126887135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Tracking long-range dependencies with wavelets 用小波跟踪远程依赖关系
Pub Date : 1994-10-27 DOI: 10.1109/WITS.1994.513885
P. Flandrin, P. Abry
Long-range dependent processes exhibit features, such as 1/f spectra, for which wavelets offer versatile tools and provide a unifying framework. This efficiency is demonstrated on both continuous processes, point processes and filtered point processes. The fractal shot noise model is also considered.
远程依赖过程表现出特征,如1/f谱,小波为其提供了通用工具并提供了统一的框架。这种效率在连续过程、点过程和滤波点过程中都得到了证明。同时考虑了分形散粒噪声模型。
{"title":"Tracking long-range dependencies with wavelets","authors":"P. Flandrin, P. Abry","doi":"10.1109/WITS.1994.513885","DOIUrl":"https://doi.org/10.1109/WITS.1994.513885","url":null,"abstract":"Long-range dependent processes exhibit features, such as 1/f spectra, for which wavelets offer versatile tools and provide a unifying framework. This efficiency is demonstrated on both continuous processes, point processes and filtered point processes. The fractal shot noise model is also considered.","PeriodicalId":423518,"journal":{"name":"Proceedings of 1994 Workshop on Information Theory and Statistics","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115860351","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Identification via compressed data 通过压缩数据进行识别
Pub Date : 1994-10-27 DOI: 10.1109/WITS.1994.513869
R. Ahlswede, E. Yang, Zhen Zhang
A combined problem of source coding and identification is considered. To put the problem in perspective, the authors first review the traditional problem in source coding theory.
考虑了源编码与识别相结合的问题。为了正确地看待这个问题,作者首先回顾了源编码理论中的传统问题。
{"title":"Identification via compressed data","authors":"R. Ahlswede, E. Yang, Zhen Zhang","doi":"10.1109/WITS.1994.513869","DOIUrl":"https://doi.org/10.1109/WITS.1994.513869","url":null,"abstract":"A combined problem of source coding and identification is considered. To put the problem in perspective, the authors first review the traditional problem in source coding theory.","PeriodicalId":423518,"journal":{"name":"Proceedings of 1994 Workshop on Information Theory and Statistics","volume":"723 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114903243","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
The finite-sample risk of the k-nearest-neighbor classifier under the L/sub p/ metric 在L/sub p/度量下k-近邻分类器的有限样本风险
Pub Date : 1994-10-27 DOI: 10.1109/WITS.1994.513925
R. Snapp, S. S. Venkatesh
The finite-sample risk of the k-nearest neighbor classifier that uses an L/sub 2/ distance function is examined. For a family of classification problems with smooth distributions in R/sup n/, the risk can be represented as an asymptotic expansion in inverse powers of the n-th root of the reference-sample size. The leading coefficients of this expansion suggest that the Euclidean or L/sub 2/ distance function minimizes the risk for sufficiently large reference samples.
研究了使用L/sub 2/距离函数的k近邻分类器的有限样本风险。对于一类光滑分布在R/sup n/中的分类问题,其风险可以表示为参考样本量n次方根的反幂的渐近展开式。该展开式的主要系数表明,对于足够大的参考样本,欧几里得或L/sub 2/距离函数将风险最小化。
{"title":"The finite-sample risk of the k-nearest-neighbor classifier under the L/sub p/ metric","authors":"R. Snapp, S. S. Venkatesh","doi":"10.1109/WITS.1994.513925","DOIUrl":"https://doi.org/10.1109/WITS.1994.513925","url":null,"abstract":"The finite-sample risk of the k-nearest neighbor classifier that uses an L/sub 2/ distance function is examined. For a family of classification problems with smooth distributions in R/sup n/, the risk can be represented as an asymptotic expansion in inverse powers of the n-th root of the reference-sample size. The leading coefficients of this expansion suggest that the Euclidean or L/sub 2/ distance function minimizes the risk for sufficiently large reference samples.","PeriodicalId":423518,"journal":{"name":"Proceedings of 1994 Workshop on Information Theory and Statistics","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124836170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
New distortion measures for speech processing
Pub Date : 1994-10-27 DOI: 10.1109/WITS.1994.513919
T.-H. Li, J. Gibson
New distortion measures are derived from a recently proposed characterization function of stationary time series and are shown to be more robust than some commonly-used distortion measures such as the Kullback-Leibler spectral divergence in speech processing.
新的失真度量来源于最近提出的平稳时间序列的表征函数,并被证明比语音处理中常用的一些失真度量(如Kullback-Leibler谱散度)更具鲁棒性。
{"title":"New distortion measures for speech processing","authors":"T.-H. Li, J. Gibson","doi":"10.1109/WITS.1994.513919","DOIUrl":"https://doi.org/10.1109/WITS.1994.513919","url":null,"abstract":"New distortion measures are derived from a recently proposed characterization function of stationary time series and are shown to be more robust than some commonly-used distortion measures such as the Kullback-Leibler spectral divergence in speech processing.","PeriodicalId":423518,"journal":{"name":"Proceedings of 1994 Workshop on Information Theory and Statistics","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124853875","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
On the theory and application of universal classification to signal detection 通用分类在信号检测中的理论与应用
Pub Date : 1994-10-27 DOI: 10.1109/WITS.1994.513908
N. Warke, G.C. Orsak
The authors apply methods of universal classification to the problem of classifying one of M deterministic signals in the presence of dependent non-Gaussian noise.
本文应用广义分类方法对存在非高斯相关噪声的M个确定性信号中的一个进行分类。
{"title":"On the theory and application of universal classification to signal detection","authors":"N. Warke, G.C. Orsak","doi":"10.1109/WITS.1994.513908","DOIUrl":"https://doi.org/10.1109/WITS.1994.513908","url":null,"abstract":"The authors apply methods of universal classification to the problem of classifying one of M deterministic signals in the presence of dependent non-Gaussian noise.","PeriodicalId":423518,"journal":{"name":"Proceedings of 1994 Workshop on Information Theory and Statistics","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127717343","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Some estimation problems in infinite dimensional Gaussian white noise 无限维高斯白噪声中的若干估计问题
Pub Date : 1994-10-27 DOI: 10.1109/WITS.1994.513872
I. Ibragimov, R. Khasminskii
Methods of the information theory and approximation theory are used to obtain the conditions for the existence of consistent estimators for the observations in a Gaussian white noise in a Hilbert space.
利用信息论和逼近论的方法,得到了希尔伯特空间中高斯白噪声观测值相合估计量存在的条件。
{"title":"Some estimation problems in infinite dimensional Gaussian white noise","authors":"I. Ibragimov, R. Khasminskii","doi":"10.1109/WITS.1994.513872","DOIUrl":"https://doi.org/10.1109/WITS.1994.513872","url":null,"abstract":"Methods of the information theory and approximation theory are used to obtain the conditions for the existence of consistent estimators for the observations in a Gaussian white noise in a Hilbert space.","PeriodicalId":423518,"journal":{"name":"Proceedings of 1994 Workshop on Information Theory and Statistics","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121068666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Minimal randomness and information theory 最小随机性和信息论
Pub Date : 1994-10-27 DOI: 10.1109/WITS.1994.513867
S. Verdú
This is a tutorial survey of recent information theoretic results dealing with the minimal randomness necessary for the generation of random processes with prescribed distributions.
这是最近的信息论研究结果的一个指导调查,这些研究结果处理了产生具有规定分布的随机过程所必需的最小随机性。
{"title":"Minimal randomness and information theory","authors":"S. Verdú","doi":"10.1109/WITS.1994.513867","DOIUrl":"https://doi.org/10.1109/WITS.1994.513867","url":null,"abstract":"This is a tutorial survey of recent information theoretic results dealing with the minimal randomness necessary for the generation of random processes with prescribed distributions.","PeriodicalId":423518,"journal":{"name":"Proceedings of 1994 Workshop on Information Theory and Statistics","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121144548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Non-parametric discriminatory power 非参数歧视力
Pub Date : 1994-10-27 DOI: 10.1109/WITS.1994.513894
H.J. Holz, M. Loew
Discriminatory power is the relative usefulness of a feature for classification. Traditionally feature-selection techniques have defined discriminatory power in terms of a particular classifier. Non-parametric discriminately power allows feature selection to be based on the structure of the data rather than on the requirements of any one classifier. In previous research, we have defined a metric for non-parametric discriminatory power called relative feature importance (RFI). In this work, we explore the construction of RFI through closed-form analysis and experimentation. The behavior of RFI is also compared to traditional techniques.
区分力是一个特征对分类的相对有用性。传统的特征选择技术根据一个特定的分类器来定义区分权。非参数判别能力允许特征选择基于数据的结构,而不是基于任何一个分类器的要求。在之前的研究中,我们定义了一个度量非参数歧视能力的指标,称为相对特征重要性(RFI)。在这项工作中,我们通过封闭形式的分析和实验来探索RFI的构建。RFI的行为也与传统技术进行了比较。
{"title":"Non-parametric discriminatory power","authors":"H.J. Holz, M. Loew","doi":"10.1109/WITS.1994.513894","DOIUrl":"https://doi.org/10.1109/WITS.1994.513894","url":null,"abstract":"Discriminatory power is the relative usefulness of a feature for classification. Traditionally feature-selection techniques have defined discriminatory power in terms of a particular classifier. Non-parametric discriminately power allows feature selection to be based on the structure of the data rather than on the requirements of any one classifier. In previous research, we have defined a metric for non-parametric discriminatory power called relative feature importance (RFI). In this work, we explore the construction of RFI through closed-form analysis and experimentation. The behavior of RFI is also compared to traditional techniques.","PeriodicalId":423518,"journal":{"name":"Proceedings of 1994 Workshop on Information Theory and Statistics","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126077531","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Proceedings of 1994 Workshop on Information Theory and Statistics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1