首页 > 最新文献

Comput. Linguistics最新文献

英文 中文
Inheritance and Complementation: a Case Study of Easy Adjectives and Related Nouns 继承与互补:以简单形容词和相关名词为例
Pub Date : 1992-09-01 DOI: 10.22028/D291-24826
D. Flickinger, J. Nerbonne
Mechanisms for representing lexically the bulk of syntactic and semantic information for a language have been under active development, as is evident in the recent studies contained in this volume. Our study serves to highlight some of themost useful tools available for structured lexical representation, in particular (multiple) inheritance, default specification, and lexical rules. It then illustrates the value of these mechanisms in illuminating one corner of the lexicon involving an unusual kind of complementation among a group of adjectives exemplified by easy. The virtues of the structured lexicon are its succinctness and its tendency to highlight significant clusters of linguistic properties. From its succinctness follow two practical advantages, namely its ease of maintenance and modification. In order to suggest how important these may be practically, we extend the analysis of adjectival complementation in several directions. These further illustrate how the use of inheritance in lexical representation permits exact and explicit characterizations of phenomena in the language under study. We demonstrate how the use of the mechanisms employed in the analysis of easy enables us to give a unified account of related phenomena featuring nouns such as pleasure, and even the adverbs (adjectival specifiers) too and enough. Along the way we motivate some elaborations of the HPSG (head-driven phrase structure grammar) framework in which we couch our analysis, and offer several avenues for further study of this part of the English lexicon.
从本卷所载的最新研究中可以明显看出,一种语言在词汇上表示大量语法和语义信息的机制一直处于积极发展之中。我们的研究强调了结构化词法表示的一些最有用的工具,特别是(多重)继承、默认规范和词法规则。然后,它说明了这些机制的价值,在解释词典的一个角落,涉及一组形容词中的一种不寻常的补语,例如easy。结构化词汇的优点是简洁明了,并倾向于突出语言特性的重要集群。它的简洁遵循两个实用的优点,即易于维护和修改。为了说明它们在实践中可能有多重要,我们从几个方面扩展了对形容词补语的分析。这些进一步说明了在词汇表示中使用继承如何允许对所研究语言中的现象进行精确和明确的描述。我们展示了在easy的分析中使用的机制如何使我们能够对相关现象给出一个统一的解释,这些现象包括名词,如pleasure,甚至副词(形容词说明符)。在此过程中,我们激发了HPSG(头部驱动短语结构语法)框架的一些阐述,我们在这个框架中进行了分析,并为进一步研究这部分英语词汇提供了几种途径。
{"title":"Inheritance and Complementation: a Case Study of Easy Adjectives and Related Nouns","authors":"D. Flickinger, J. Nerbonne","doi":"10.22028/D291-24826","DOIUrl":"https://doi.org/10.22028/D291-24826","url":null,"abstract":"Mechanisms for representing lexically the bulk of syntactic and semantic information for a language have been under active development, as is evident in the recent studies contained in this volume. Our study serves to highlight some of themost useful tools available for structured lexical representation, in particular (multiple) inheritance, default specification, and lexical rules. It then illustrates the value of these mechanisms in illuminating one corner of the lexicon involving an unusual kind of complementation among a group of adjectives exemplified by easy. The virtues of the structured lexicon are its succinctness and its tendency to highlight significant clusters of linguistic properties. From its succinctness follow two practical advantages, namely its ease of maintenance and modification. In order to suggest how important these may be practically, we extend the analysis of adjectival complementation in several directions. These further illustrate how the use of inheritance in lexical representation permits exact and explicit characterizations of phenomena in the language under study. We demonstrate how the use of the mechanisms employed in the analysis of easy enables us to give a unified account of related phenomena featuring nouns such as pleasure, and even the adverbs (adjectival specifiers) too and enough. Along the way we motivate some elaborations of the HPSG (head-driven phrase structure grammar) framework in which we couch our analysis, and offer several avenues for further study of this part of the English lexicon.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1992-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127600487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 62
A Practical Approach to Multiple Default Inheritance for Unification-Based Lexicons 基于统一的词典多重默认继承的实用方法
Pub Date : 1992-09-01 DOI: 10.1017/CBO9780511663642.008
G. Russell, Afzal Ballim, John A. Carroll, Susan Warwick-Armstrong
This paper describes a unification-based lexicon system for NLP applications that incorporates mechanisms for multiple default inheritance. Such systems are intractable in the general case---the approach adopted here places a number of restrictions on the inheritance hierarchy in order to remove some of the sources of complexity while retaining more desirable properties. Implications of the design choices are discussed, comparisons are drawn with related work in computational linguistics and AI, and illustrative examples from the lexicons of German and English are given.
本文描述了一个基于统一的NLP应用词典系统,该系统包含多个默认继承机制。这样的系统在一般情况下是难以处理的——这里采用的方法对继承层次结构施加了许多限制,以便在保留更理想的属性的同时消除一些复杂性的来源。讨论了设计选择的含义,与计算语言学和人工智能的相关工作进行了比较,并给出了德语和英语词汇的说明性示例。
{"title":"A Practical Approach to Multiple Default Inheritance for Unification-Based Lexicons","authors":"G. Russell, Afzal Ballim, John A. Carroll, Susan Warwick-Armstrong","doi":"10.1017/CBO9780511663642.008","DOIUrl":"https://doi.org/10.1017/CBO9780511663642.008","url":null,"abstract":"This paper describes a unification-based lexicon system for NLP applications that incorporates mechanisms for multiple default inheritance. Such systems are intractable in the general case---the approach adopted here places a number of restrictions on the inheritance hierarchy in order to remove some of the sources of complexity while retaining more desirable properties. Implications of the design choices are discussed, comparisons are drawn with related work in computational linguistics and AI, and illustrative examples from the lexicons of German and English are given.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1992-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121465701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 59
Incremental Processing and the Hierarchical Lexicon 增量处理和分层词典
Pub Date : 1992-06-01 DOI: 10.5555/142235.142245
E. V. D. Linden
Hierarchical lexicon structures are not only of great importance for the nonredundant representation of lexical information, they may also contribute to the efficiency of the actual processing of natural language. Two parsing techniques that render the parsing process efficient are presented. Windowing is a technique for incrementally accessing the hierarchical lexicon. Lexical preferencing implements preferences within the parsing process as a natural consequence of the hierarchical structure of the lexicon. Within a proof-theoretic approach to Categorial Grammar it is possible to implement these techniques in a formal and principled way. Special attention is paid to idiomatic expressions.
分层词汇结构不仅对词汇信息的无冗余表示具有重要意义,而且有助于提高自然语言实际处理的效率。提出了两种解析技术,以提高解析过程的效率。窗口化是一种增量访问分层词典的技术。词法首选项在解析过程中实现首选项,这是词法层次结构的自然结果。在范畴语法的证明理论方法中,有可能以形式化和原则性的方式实现这些技术。特别注意习语表达。
{"title":"Incremental Processing and the Hierarchical Lexicon","authors":"E. V. D. Linden","doi":"10.5555/142235.142245","DOIUrl":"https://doi.org/10.5555/142235.142245","url":null,"abstract":"Hierarchical lexicon structures are not only of great importance for the nonredundant representation of lexical information, they may also contribute to the efficiency of the actual processing of natural language. Two parsing techniques that render the parsing process efficient are presented. Windowing is a technique for incrementally accessing the hierarchical lexicon. Lexical preferencing implements preferences within the parsing process as a natural consequence of the hierarchical structure of the lexicon. Within a proof-theoretic approach to Categorial Grammar it is possible to implement these techniques in a formal and principled way. Special attention is paid to idiomatic expressions.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1992-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130221394","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
A Program for Aligning Sentences in Bilingual Corpora 双语语料库句子对齐程序
Pub Date : 1991-06-18 DOI: 10.3115/981344.981367
W. Gale, Kenneth Ward Church
Researchers in both machine translation (e.g., Brown et al. 1990) and bilingual lexicography (e.g., Klavans and Tzoukermann 1990) have recently become interested in studying bilingual corpora, bodies of text such as the Canadian Hansards (parliamentary proceedings), which are available in multiple languages (such as French and English). One useful step is to align the sentences, that is, to identify correspondences between sentences in one language and sentences in the other language.This paper will describe a method and a program (align) for aligning sentences based on a simple statistical model of character lengths. The program uses the fact that longer sentences in one language tend to be translated into longer sentences in the other language, and that shorter sentences tend to be translated into shorter sentences. A probabilistic score is assigned to each proposed correspondence of sentences, based on the scaled difference of lengths of the two sentences (in characters) and the variance of this difference. This probabilistic score is used in a dynamic programming framework to find the maximum likelihood alignment of sentences.It is remarkable that such a simple approach works as well as it does. An evaluation was performed based on a trilingual corpus of economic reports issued by the Union Bank of Switzerland (UBS) in English, French, and German. The method correctly aligned all but 4% of the sentences. Moreover, it is possible to extract a large subcorpus that has a much smaller error rate. By selecting the best-scoring 80% of the alignments, the error rate is reduced from 4% to 0.7%. There were more errors on the English-French subcorpus than on the English-German subcorpus, showing that error rates will depend on the corpus considered; however, both were small enough to hope that the method will be useful for many language pairs.To further research on bilingual corpora, a much larger sample of Canadian Hansards (approximately 90 million words, half in English and and half in French) has been aligned with the align program and will be available through the Data Collection Initiative of the Association for Computational Linguistics (ACL/DCI). In addition, in order to facilitate replication of the align program, an appendix is provided with detailed c-code of the more difficult core of the align program.
机器翻译(如Brown et al. 1990)和双语词典编纂(如Klavans and Tzoukermann 1990)的研究人员最近开始对双语语料库的研究感兴趣,双语语料库是指以多种语言(如法语和英语)提供的加拿大议事录(议会会议记录)等文本体。一个有用的步骤是对齐句子,即识别一种语言中的句子与另一种语言中的句子之间的对应关系。本文将描述一种基于简单的字符长度统计模型的句子对齐方法和程序(align)。该程序利用这样一个事实:一种语言中的较长句子往往会被翻译成另一种语言中的较长句子,而较短的句子往往会被翻译成较短的句子。根据两个句子(以字符为单位)长度的比例差异和该差异的方差,为每个提议的句子对应分配一个概率分数。在动态规划框架中使用这个概率分数来找到句子的最大似然对齐。值得注意的是,这样一个简单的方法能发挥如此大的作用。根据瑞士联合银行(UBS)以英语、法语和德语发行的三语经济报告语料库进行了评估。除了4%的句子外,该方法正确排列了所有的句子。此外,还可以提取错误率小得多的大型子语料库。通过选择得分最高的80%的对齐,错误率从4%降低到0.7%。英语-法语子语料库的错误率高于英语-德语子语料库,这表明错误率取决于所考虑的语料库;然而,这两种方法都很小,希望该方法对许多语言对都有用。为了进一步研究双语语料库,一个更大的加拿大语料库样本(大约9000万字,一半是英语,一半是法语)已经与align程序对齐,并将通过计算语言学协会的数据收集倡议(ACL/DCI)提供。此外,为了便于align程序的复制,在附录中提供了align程序中较困难的核心的详细c代码。
{"title":"A Program for Aligning Sentences in Bilingual Corpora","authors":"W. Gale, Kenneth Ward Church","doi":"10.3115/981344.981367","DOIUrl":"https://doi.org/10.3115/981344.981367","url":null,"abstract":"Researchers in both machine translation (e.g., Brown et al. 1990) and bilingual lexicography (e.g., Klavans and Tzoukermann 1990) have recently become interested in studying bilingual corpora, bodies of text such as the Canadian Hansards (parliamentary proceedings), which are available in multiple languages (such as French and English). One useful step is to align the sentences, that is, to identify correspondences between sentences in one language and sentences in the other language.This paper will describe a method and a program (align) for aligning sentences based on a simple statistical model of character lengths. The program uses the fact that longer sentences in one language tend to be translated into longer sentences in the other language, and that shorter sentences tend to be translated into shorter sentences. A probabilistic score is assigned to each proposed correspondence of sentences, based on the scaled difference of lengths of the two sentences (in characters) and the variance of this difference. This probabilistic score is used in a dynamic programming framework to find the maximum likelihood alignment of sentences.It is remarkable that such a simple approach works as well as it does. An evaluation was performed based on a trilingual corpus of economic reports issued by the Union Bank of Switzerland (UBS) in English, French, and German. The method correctly aligned all but 4% of the sentences. Moreover, it is possible to extract a large subcorpus that has a much smaller error rate. By selecting the best-scoring 80% of the alignments, the error rate is reduced from 4% to 0.7%. There were more errors on the English-French subcorpus than on the English-German subcorpus, showing that error rates will depend on the corpus considered; however, both were small enough to hope that the method will be useful for many language pairs.To further research on bilingual corpora, a much larger sample of Canadian Hansards (approximately 90 million words, half in English and and half in French) has been aligned with the align program and will be available through the Data Collection Initiative of the Association for Computational Linguistics (ACL/DCI). In addition, in order to facilitate replication of the align program, an appendix is provided with detailed c-code of the more difficult core of the align program.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1991-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123602899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1340
A Statistical Approach to Machine Translation 机器翻译的统计方法
Pub Date : 1990-06-01 DOI: 10.7551/mitpress/5779.003.0039
P. Brown, J. Cocke, S. D. Pietra, V. D. Pietra, F. Jelinek, J. Lafferty, R. Mercer, P. Roossin
In this paper, we present a statistical approach to machine translation. We describe the application of our approach to translation from French to English and give preliminary results.
在本文中,我们提出了一种机器翻译的统计方法。我们描述了我们的方法在法语英译中的应用,并给出了初步的结果。
{"title":"A Statistical Approach to Machine Translation","authors":"P. Brown, J. Cocke, S. D. Pietra, V. D. Pietra, F. Jelinek, J. Lafferty, R. Mercer, P. Roossin","doi":"10.7551/mitpress/5779.003.0039","DOIUrl":"https://doi.org/10.7551/mitpress/5779.003.0039","url":null,"abstract":"In this paper, we present a statistical approach to machine translation. We describe the application of our approach to translation from French to English and give preliminary results.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"80 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128533176","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1981
Tense, Quantifiers, and Contexts 时态、量词和语境
Pub Date : 1988-06-01 DOI: 10.5555/55056.55057
E. Hinrichs
This paper describes a compositional semantics for temporal expressions as part of the meaning representation language (MRL) of the JANUS system, a natural language understanding and generation system under joint development by BBN Labs and ISI. The analysis is based on a higher-order intensional logic described in detail in Hinrichs (1987a). Temporal expressions of English are translated into this language as quantifiers over times that bind temporal indices on predicates. The semantic evaluation of time-dependent predicates is defined relative to a set of discourse contexts, which, following Reichenbach (1947), include the parameters of speech time and reference time. The resulting context-dependent and multi-indexed interpretation of temporal expressions solves a set of well-known problems that arise when traditional systems of tense logic are applied to natural language semantics. Based on the principle of rule-to-rule translation, the compositional nature of the analysis provides a straightforward and well-defined interface between the parsing component and the semantic-interpretation component of JANUS.
本文描述了一种时间表达式的组合语义,作为JANUS系统的意义表示语言(MRL)的一部分,JANUS系统是BBN实验室和ISI联合开发的自然语言理解和生成系统。该分析基于Hinrichs (1987a)详细描述的高阶内涵逻辑。随着时间的推移,英语的时间表达式作为量词被翻译成这种语言,这些量词在谓词上绑定了时间索引。时间依赖谓词的语义评价是相对于一组话语语境来定义的,在Reichenbach(1947)之后,这些话语语境包括言语时间和参考时间的参数。由此产生的时间表达式的上下文依赖和多索引解释解决了传统时态逻辑系统应用于自然语言语义时出现的一系列众所周知的问题。基于规则到规则转换的原则,分析的组合特性在解析组件和JANUS的语义解释组件之间提供了一个简单且定义良好的接口。
{"title":"Tense, Quantifiers, and Contexts","authors":"E. Hinrichs","doi":"10.5555/55056.55057","DOIUrl":"https://doi.org/10.5555/55056.55057","url":null,"abstract":"This paper describes a compositional semantics for temporal expressions as part of the meaning representation language (MRL) of the JANUS system, a natural language understanding and generation system under joint development by BBN Labs and ISI. The analysis is based on a higher-order intensional logic described in detail in Hinrichs (1987a). Temporal expressions of English are translated into this language as quantifiers over times that bind temporal indices on predicates. The semantic evaluation of time-dependent predicates is defined relative to a set of discourse contexts, which, following Reichenbach (1947), include the parameters of speech time and reference time. The resulting context-dependent and multi-indexed interpretation of temporal expressions solves a set of well-known problems that arise when traditional systems of tense logic are applied to natural language semantics. Based on the principle of rule-to-rule translation, the compositional nature of the analysis provides a straightforward and well-defined interface between the parsing component and the semantic-interpretation component of JANUS.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1988-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134488996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 45
Commonsense Metaphysics and Lexical Semantics 常识形而上学与词汇语义学
Pub Date : 1986-05-01 DOI: 10.3115/981131.981163
Jerry R. Hobbs, W. Bruce Croft, Todd R. Davies, Douglas Edwards, K. Laws
In the TACITUS project for using commonsense knowledge in the understanding of texts about mechanical devices and their failures, we have been developing various commonsense theories that are needed to mediate between the way we talk about the behavior of such devices and causal models of their operation. Of central importance in this effort is the axiomatization of what might be called "commonsense metaphysics". This includes a number of areas that figure in virtually every domain of discourse, such as scalar notions, granularity, time, space, material, physical objects, causality, functionality, force, and shape. Our approach to lexical semantics is then to construct core theories of each of these areas, and then to define, or at least characterize, a large number of lexical items in terms provided by the core theories. In the TACITUS system, processes for solving pragmatics problems posed by a text will use the knowledge base consisting of these theories in conjunction with the logical forms of the sentences in the text to produce an interpretation. In this paper we do not stress these interpretation processes; this is another, important aspect of the TACITUS project, and it will be described in subsequent papers.
在塔西佗的项目中,运用常识知识来理解关于机械设备及其故障的文本,我们已经发展了各种常识性理论,这些理论需要在我们谈论这些设备行为的方式和它们运行的因果模型之间进行调解。在这一努力中,最重要的是可以被称为“常识形而上学”的公理化。这包括了几乎在每个话语领域中都存在的一些领域,比如标量概念、粒度、时间、空间、材料、物理对象、因果关系、功能、力和形状。我们研究词汇语义学的方法是构建这些领域的核心理论,然后根据这些核心理论提供的术语定义或至少描述大量词汇项目。在塔西佗体系中,解决文本所提出的语用学问题的过程将使用由这些理论组成的知识库,并结合文本中句子的逻辑形式来产生解释。在本文中,我们不强调这些解释过程;这是TACITUS计划的另一个重要方面,将在后续的论文中描述。
{"title":"Commonsense Metaphysics and Lexical Semantics","authors":"Jerry R. Hobbs, W. Bruce Croft, Todd R. Davies, Douglas Edwards, K. Laws","doi":"10.3115/981131.981163","DOIUrl":"https://doi.org/10.3115/981131.981163","url":null,"abstract":"In the TACITUS project for using commonsense knowledge in the understanding of texts about mechanical devices and their failures, we have been developing various commonsense theories that are needed to mediate between the way we talk about the behavior of such devices and causal models of their operation. Of central importance in this effort is the axiomatization of what might be called \"commonsense metaphysics\". This includes a number of areas that figure in virtually every domain of discourse, such as scalar notions, granularity, time, space, material, physical objects, causality, functionality, force, and shape. Our approach to lexical semantics is then to construct core theories of each of these areas, and then to define, or at least characterize, a large number of lexical items in terms provided by the core theories. In the TACITUS system, processes for solving pragmatics problems posed by a text will use the knowledge base consisting of these theories in conjunction with the logical forms of the sentences in the text to produce an interpretation. In this paper we do not stress these interpretation processes; this is another, important aspect of the TACITUS project, and it will be described in subsequent papers.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1986-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128643311","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 107
PHRED: A Generator for Natural Language Interfaces PHRED:自然语言接口的生成器
Pub Date : 1985-10-01 DOI: 10.1007/978-1-4612-3846-1_9
P. Jacobs
{"title":"PHRED: A Generator for Natural Language Interfaces","authors":"P. Jacobs","doi":"10.1007/978-1-4612-3846-1_9","DOIUrl":"https://doi.org/10.1007/978-1-4612-3846-1_9","url":null,"abstract":"","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1985-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126433087","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 90
Strong Generative Capacity, Weak Generative Capacity, and Modern Linguistic Theories 强生成能力、弱生成能力与现代语言学理论
Pub Date : 1984-07-01 DOI: 10.5555/970170.970175
R. Berwick
What makes a language a natural language? A longstanding tradition in generative grammar holds that a language is natural just in case it is learnable under a constellation of auxiliary assumptions about input evidence available to children. Yet another approach seeks some key mathematical property that distinguishes the natural languages from all possible symbol-systems. With some exceptions for example, Chomsky's demonstration that a complete characterization of our grammatical knowledge lies beyond the power of finite state languages the mathematical approach has not provided clear-cut results. For example, for a variety of reasons we cannot say that the predicate is context-free characterizes all and only the natural languages. Still another use of mathematical analysis in linguistics has been to diagnose a proposed grammatical formalism as too powerful (allowing too many grammars or languages) rather than as too weak. Such a diagnosis was supposed by some to follow from Peters and Ritchie's demonstration that the theory of transformational grammar as described in Chomsky's Aspects of the Theory of Syntax could specify grammars to generate any recursively enumerable set. For some this demonstration marked a watershed in the formal analysis transformational grammar. One general reaction (not prompted by the Peters and Ritchie result alone) was to turn to other theories of grammar designed to explicitly avoid the problems of a theory that could specify an arbitrary Turing machine computation. The proposals for generalized phrase structure grammar (GPSG) and lexical-functional grammar (LFG) have explicitly emphasized this point. GPSG aims for grammars that generate context-free languages (though there is some recent wavering on this point; see Pullum 1984); LFG, for languages that are at worst context-sensitive. Whatever the merits of the arguments for this restriction in terms of weak generative capacity and they are far from obvious, as discussed at length in Berwick and Weinberg (1983) one point remains: the switch was prompted by criticism of the nearly two-decades old Aspects theory. Much has changed in transformational grammar in twenty years. Modern transformational grammars no longer contain swarms of individual rules such as Passive, Raising, or Dative. The modern government-binding (GB) theory does not reconstruct a "deep structure", does not contain powerful deletion rules, and has introduced a whole host of new constraints. Given these sweeping changes, it would seem appropriate, then, to re-examine the Peters and Ritchie result, and compare the power of the newer GB-style theories to these other current linguistic theories. That is the aim of this paper. The basic points to be made are these:
是什么使一门语言成为自然语言?生成语法的一个长期传统认为,一种语言是自然的,只要它是在一系列关于儿童可用的输入证据的辅助假设下可学习的。然而,另一种方法寻求一些关键的数学性质,将自然语言与所有可能的符号系统区分开来。除了一些例外,例如,乔姆斯基证明,我们的语法知识的完整表征超出了有限状态语言的能力,数学方法并没有提供明确的结果。例如,由于各种原因,我们不能说谓词是上下文无关的,它只能描述所有的自然语言。数学分析在语言学中的另一个用途是诊断一种拟议的语法形式主义是过于强大(允许太多的语法或语言),而不是太弱。这种诊断被一些人认为是从彼得斯和里奇的论证中得出的,乔姆斯基的《语法理论的各个方面》中描述的转换语法理论可以指定语法来生成任何递归可枚举集合。对于一些人来说,这个演示标志着形式分析转换语法的分水岭。一个普遍的反应(不仅仅是由彼得斯和里奇的结果引起的)是转向其他语法理论,这些理论旨在明确地避免可以指定任意图灵机计算的理论的问题。广义短语结构语法(GPSG)和词汇功能语法(LFG)的提出都明确强调了这一点。GPSG的目标是生成与上下文无关的语言的语法(尽管最近在这一点上有一些动摇;参见Pullum 1984);LFG,用于最坏的情况下对上下文敏感的语言。无论从弱生产能力的角度来看,这种限制的论点有什么优点,而且它们远不明显,正如Berwick和Weinberg(1983)中详细讨论的那样,有一点仍然存在:这种转变是由对近20年前的Aspects理论的批评引起的。二十年来,变换语法发生了很大的变化。现代转换语法不再包含大量的单独规则,如被动、提升或与格。现代政府约束(GB)理论没有重建一个“深层结构”,没有包含强大的删除规则,并且引入了一大堆新的约束。考虑到这些巨大的变化,重新审视彼得斯和里奇的结果,并将新的gb风格理论的力量与其他当前的语言理论进行比较,似乎是合适的。这就是本文的目的。要提出的基本观点如下:
{"title":"Strong Generative Capacity, Weak Generative Capacity, and Modern Linguistic Theories","authors":"R. Berwick","doi":"10.5555/970170.970175","DOIUrl":"https://doi.org/10.5555/970170.970175","url":null,"abstract":"What makes a language a natural language? A longstanding tradition in generative grammar holds that a language is natural just in case it is learnable under a constellation of auxiliary assumptions about input evidence available to children. Yet another approach seeks some key mathematical property that distinguishes the natural languages from all possible symbol-systems. With some exceptions for example, Chomsky's demonstration that a complete characterization of our grammatical knowledge lies beyond the power of finite state languages the mathematical approach has not provided clear-cut results. For example, for a variety of reasons we cannot say that the predicate is context-free characterizes all and only the natural languages. Still another use of mathematical analysis in linguistics has been to diagnose a proposed grammatical formalism as too powerful (allowing too many grammars or languages) rather than as too weak. Such a diagnosis was supposed by some to follow from Peters and Ritchie's demonstration that the theory of transformational grammar as described in Chomsky's Aspects of the Theory of Syntax could specify grammars to generate any recursively enumerable set. For some this demonstration marked a watershed in the formal analysis transformational grammar. One general reaction (not prompted by the Peters and Ritchie result alone) was to turn to other theories of grammar designed to explicitly avoid the problems of a theory that could specify an arbitrary Turing machine computation. The proposals for generalized phrase structure grammar (GPSG) and lexical-functional grammar (LFG) have explicitly emphasized this point. GPSG aims for grammars that generate context-free languages (though there is some recent wavering on this point; see Pullum 1984); LFG, for languages that are at worst context-sensitive. Whatever the merits of the arguments for this restriction in terms of weak generative capacity and they are far from obvious, as discussed at length in Berwick and Weinberg (1983) one point remains: the switch was prompted by criticism of the nearly two-decades old Aspects theory. Much has changed in transformational grammar in twenty years. Modern transformational grammars no longer contain swarms of individual rules such as Passive, Raising, or Dative. The modern government-binding (GB) theory does not reconstruct a \"deep structure\", does not contain powerful deletion rules, and has introduced a whole host of new constraints. Given these sweeping changes, it would seem appropriate, then, to re-examine the Peters and Ritchie result, and compare the power of the newer GB-style theories to these other current linguistic theories. That is the aim of this paper. The basic points to be made are these:","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"302 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1984-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124316034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
The Correction of Ill-Formed Input Using History-Based Expectation with Applications to Speech Understanding 基于历史期望的错误输入纠正及其在语音理解中的应用
Pub Date : 1900-01-01 DOI: 10.21236/ada162222
Pamela E. Fink, A. Biermann
A method for error correction of ill-formed input is described that acquires dialogue patterns in typical usage and uses these patterns to predict new inputs. Error correction is done by strongly biasing parsing toward expected meanings unless clear evidence from the input shows the current sentence is not expected. A dialogue acquisition and tracking algorithm is presented along with a description of its implementation in a voice interactive system. A series of tests are described that show the power of the error correction methodology when stereotypic dialogue occurs.
描述了一种错误格式输入的纠错方法,该方法获取典型用法中的对话模式,并使用这些模式预测新的输入。错误纠正是通过强烈偏向预期意义的解析来完成的,除非来自输入的明确证据表明当前句子不是预期的。提出了一种对话采集和跟踪算法,并描述了该算法在语音交互系统中的实现。本文描述了一系列测试,这些测试显示了在出现刻板印象对话时纠错方法的威力。
{"title":"The Correction of Ill-Formed Input Using History-Based Expectation with Applications to Speech Understanding","authors":"Pamela E. Fink, A. Biermann","doi":"10.21236/ada162222","DOIUrl":"https://doi.org/10.21236/ada162222","url":null,"abstract":"A method for error correction of ill-formed input is described that acquires dialogue patterns in typical usage and uses these patterns to predict new inputs. Error correction is done by strongly biasing parsing toward expected meanings unless clear evidence from the input shows the current sentence is not expected. A dialogue acquisition and tracking algorithm is presented along with a description of its implementation in a voice interactive system. A series of tests are described that show the power of the error correction methodology when stereotypic dialogue occurs.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125082802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
期刊
Comput. Linguistics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1