首页 > 最新文献

Comput. Linguistics最新文献

英文 中文
Foundations of Intensional Semantics 内涵语义学基础
Pub Date : 2006-06-01 DOI: 10.1162/coli.2006.32.2.291
From the combination of knowledge and actions, someone can improve their skill and ability. It will lead them to live and work much better. This is why, the students, workers, or even employers should have reading habit for books. Any book will give certain knowledge to take all benefits. This is what this foundations of intensional semantics tells you. It will add more knowledge of you to life and work better. Try it and prove it.
从知识和行动的结合,一个人可以提高他们的技能和能力。这将使他们生活和工作得更好。这就是为什么学生,工人,甚至雇主都应该有读书的习惯。任何一本书都会给予一定的知识,使人受益无穷。这就是内涵语义的基础告诉你的。它会为你的生活和工作增添更多的知识。尝试并证明它。
{"title":"Foundations of Intensional Semantics","authors":"","doi":"10.1162/coli.2006.32.2.291","DOIUrl":"https://doi.org/10.1162/coli.2006.32.2.291","url":null,"abstract":"From the combination of knowledge and actions, someone can improve their skill and ability. It will lead them to live and work much better. This is why, the students, workers, or even employers should have reading habit for books. Any book will give certain knowledge to take all benefits. This is what this foundations of intensional semantics tells you. It will add more knowledge of you to life and work better. Try it and prove it.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"141 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117092055","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
Generating Natural Language Summaries from Multiple On-Line Sources 从多个在线资源生成自然语言摘要
Pub Date : 1998-09-01 DOI: 10.7916/D83N2B7J
Dragomir R. Radev, K. McKeown
We present a methodology for summarization of news about current events in the form of briefings that include appropriate background (historical) information. The system that we developed, SUMMONS, uses the output of systems developed for the DARPA Message Understanding Conferences to generate summaries of multiple documents on the same or related events, presenting similarities and differences, contradictions, and generalizations among sources of information. We describe the various components of the system, showing how information from multiple articles is combined, organized into a paragraph, and finally, realized as English sentences. A feature of our work is the extraction of descriptions of entities such as people and places for reuse to enhance a briefing.
我们提出了一种方法,以简报的形式总结当前事件的新闻,其中包括适当的背景(历史)信息。我们开发的系统,传票,使用为DARPA消息理解会议开发的系统的输出来生成关于相同或相关事件的多个文档的摘要,呈现信息来源之间的异同、矛盾和概括。我们描述了系统的各个组成部分,展示了如何将来自多个文章的信息组合起来,组织成一个段落,最后实现为英语句子。我们工作的一个特点是提取实体的描述,如人员和地点,以便重用,以增强简报。
{"title":"Generating Natural Language Summaries from Multiple On-Line Sources","authors":"Dragomir R. Radev, K. McKeown","doi":"10.7916/D83N2B7J","DOIUrl":"https://doi.org/10.7916/D83N2B7J","url":null,"abstract":"We present a methodology for summarization of news about current events in the form of briefings that include appropriate background (historical) information. The system that we developed, SUMMONS, uses the output of systems developed for the DARPA Message Understanding Conferences to generate summaries of multiple documents on the same or related events, presenting similarities and differences, contradictions, and generalizations among sources of information. We describe the various components of the system, showing how information from multiple articles is combined, organized into a paragraph, and finally, realized as English sentences. A feature of our work is the extraction of descriptions of entities such as people and places for reuse to enhance a briefing.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129817622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 484
Floating Constraints in Lexical Choice 词汇选择中的浮动约束
Pub Date : 1997-06-01 DOI: 10.7916/D82V2S0J
Michael Elhadad, K. McKeown, J. Robin
Lexical choice is a computationally complex task, requiring a generation system to consider a potentially large number of mappings between concepts and words. Constraints that aid in determining which word is best come from a wide variety of sources, including syntax, semantics, pragmatics, the lexicon, and the underlying domain. Furthermore, in some situations, different constraints come into play early on, while in others, they apply much later. This makes it difficult to determine a systematic ordering in which to apply constraints. In this paper, we present a general approach to lexical choice that can handle multiple, interacting constraints. We focus on the problem of floating constraints, semantic or pragmatic constraints that float, appearing at a variety of different syntactic ranks, often merged with other semantic constraints. This means that multiple content units can be realized by a single surface element, and conversely, that a single content unit can be realized by a variety of surface elements. Our approach uses the Functional Unification Formalism (FUF) to represent a generation lexicon, allowing for declarative and compositional representation of individual constraints.
词汇选择是一项计算复杂的任务,需要生成系统考虑概念和单词之间潜在的大量映射。帮助确定哪个词最好的约束来自各种各样的来源,包括语法、语义、语用学、词典和底层领域。此外,在某些情况下,不同的约束在早期发挥作用,而在其他情况下,它们应用得很晚。这使得确定应用约束的系统顺序变得困难。在本文中,我们提出了一种通用的词汇选择方法,可以处理多个相互作用的约束。我们关注浮动约束的问题,浮动的语义或语用约束,出现在各种不同的句法级别,经常与其他语义约束合并。这意味着可以通过单个表面元素来实现多个内容单元,反过来,可以通过多种表面元素来实现单个内容单元。我们的方法使用功能统一形式(Functional unified Formalism, FUF)来表示生成词典,允许对单个约束进行声明性和组合性表示。
{"title":"Floating Constraints in Lexical Choice","authors":"Michael Elhadad, K. McKeown, J. Robin","doi":"10.7916/D82V2S0J","DOIUrl":"https://doi.org/10.7916/D82V2S0J","url":null,"abstract":"Lexical choice is a computationally complex task, requiring a generation system to consider a potentially large number of mappings between concepts and words. Constraints that aid in determining which word is best come from a wide variety of sources, including syntax, semantics, pragmatics, the lexicon, and the underlying domain. Furthermore, in some situations, different constraints come into play early on, while in others, they apply much later. This makes it difficult to determine a systematic ordering in which to apply constraints. In this paper, we present a general approach to lexical choice that can handle multiple, interacting constraints. We focus on the problem of floating constraints, semantic or pragmatic constraints that float, appearing at a variety of different syntactic ranks, often merged with other semantic constraints. This means that multiple content units can be realized by a single surface element, and conversely, that a single content unit can be realized by a variety of surface elements. Our approach uses the Functional Unification Formalism (FUF) to represent a generation lexicon, allowing for declarative and compositional representation of individual constraints.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115007248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 66
Machine Transliteration 机音译
Pub Date : 1997-04-14 DOI: 10.3115/976909.979634
Kevin Knight, Jonathan Graehl
It is challenging to translate names and technical terms across languages with different alphabets and sound inventories. These items are commonly transliterated, i.e., replaced with approximate phonetic equivalents. For example, computer in English comes out as (konpyuutaa) in Japanese. Translating such items from Japanese back to English is even more challenging, and of practical interest, as transliterated items make up the bulk of text phrases not found in bilingual dictionaries. We describe and evaluate a method for performing backwards transliterations by machine. This method uses a generative model, incorporating several distinct stages in the transliteration process.
在不同字母和发音的语言之间翻译名字和专业术语是一项挑战。这些词通常被音译,即用近似的音标代替。例如,英语中的computer在日语中是konpyuutaa。将这些词从日语翻译回英语更具挑战性,而且具有实际意义,因为音译词构成了双语词典中找不到的大部分文本短语。我们描述并评估了一种用机器进行倒音译的方法。这种方法使用生成模型,在音译过程中包含几个不同的阶段。
{"title":"Machine Transliteration","authors":"Kevin Knight, Jonathan Graehl","doi":"10.3115/976909.979634","DOIUrl":"https://doi.org/10.3115/976909.979634","url":null,"abstract":"It is challenging to translate names and technical terms across languages with different alphabets and sound inventories. These items are commonly transliterated, i.e., replaced with approximate phonetic equivalents. For example, computer in English comes out as (konpyuutaa) in Japanese. Translating such items from Japanese back to English is even more challenging, and of practical interest, as transliterated items make up the bulk of text phrases not found in bilingual dictionaries. We describe and evaluate a method for performing backwards transliterations by machine. This method uses a generative model, incorporating several distinct stages in the transliteration process.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-04-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124679940","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 690
Discourse Segmentation by Human and Automated Means 人工和自动化的话语分割方法
Pub Date : 1997-03-01 DOI: 10.5555/972684.972689
R. Passonneau, D. Litman
The need to model the relation between discourse structure and linguistic features of utterances is almost universally acknowledged in the literature on discourse. However, there is only weak consensus on what the units of discourse structure are, or the criteria for recognizing and generating them. We present quantitative results of a two-part study using a corpus of spontaneous, narrative monologues. The first part of our paper presents a method for empirically validating multitutterance units referred to as discourse segments. We report highly significant results of segmentations performed by naive subjects, where a commonsense notion of speaker intention is the segmentation criterion. In the second part of our study, data abstracted from the subjects' segmentations serve as a target for evaluating two sets of algorithms that use utterance features to perform segmentation. On the first algorithm set, we evaluate and compare the correlation of discourse segmentation with three types of linguistic cues (referential noun phrases, cue words, and pauses). We then develop a second set using two methods: error analysis and machine learning. Testing the new algorithms on a new data set shows that when multiple sources of linguistic knowledge are used concurrently, algorithm performance improves.
对话语结构和话语语言特征之间的关系进行建模的必要性在语篇研究中几乎是公认的。然而,对于话语结构的单位是什么,以及识别和产生话语结构的标准,人们只有微弱的共识。我们目前的定量结果两部分的研究使用自发的语料库,叙述独白。本文的第一部分提出了一种经验验证被称为话语片段的多话语单位的方法。我们报告了由幼稚受试者进行的高度显著的分割结果,其中说话人意图的常识性概念是分割标准。在我们研究的第二部分中,从受试者的分割中提取的数据作为评估两组使用话语特征进行分割的算法的目标。在第一个算法集上,我们评估并比较了三种类型的语言线索(指代名词短语、提示词和停顿)对语篇分割的相关性。然后,我们使用两种方法开发第二组:误差分析和机器学习。在新数据集上的测试表明,当多个语言知识源同时使用时,算法的性能得到了提高。
{"title":"Discourse Segmentation by Human and Automated Means","authors":"R. Passonneau, D. Litman","doi":"10.5555/972684.972689","DOIUrl":"https://doi.org/10.5555/972684.972689","url":null,"abstract":"The need to model the relation between discourse structure and linguistic features of utterances is almost universally acknowledged in the literature on discourse. However, there is only weak consensus on what the units of discourse structure are, or the criteria for recognizing and generating them. We present quantitative results of a two-part study using a corpus of spontaneous, narrative monologues. The first part of our paper presents a method for empirically validating multitutterance units referred to as discourse segments. We report highly significant results of segmentations performed by naive subjects, where a commonsense notion of speaker intention is the segmentation criterion. In the second part of our study, data abstracted from the subjects' segmentations serve as a target for evaluating two sets of algorithms that use utterance features to perform segmentation. On the first algorithm set, we evaluate and compare the correlation of discourse segmentation with three types of linguistic cues (referential noun phrases, cue words, and pauses). We then develop a second set using two methods: error analysis and machine learning. Testing the new algorithms on a new data set shows that when multiple sources of linguistic knowledge are used concurrently, algorithm performance improves.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115581260","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 264
Translating Collocations for Bilingual Lexicons: A Statistical Approach 双语词汇搭配翻译:统计学方法
Pub Date : 1996-03-01 DOI: 10.7916/D8C82M3R
Frank Smadja, K. McKeown, V. Hatzivassiloglou
Collocations are notoriously difficult for non-native speakers to translate, primarily because they are opaque and cannot be translated on a word-by-word basis. We describe a program named Champollion which, given a pair of parallel corpora in two different languages and a list of collocations in one of them, automatically produces their translations. Our goal is to provide a tool for compiling bilingual lexical information above the word level in multiple languages, for different domains. The algorithm we use is based on statistical methods and produces p-word translations of n-word collocations in which n and p need not be the same. For example, Champollion translates make...decision, employment equity, and stock market into prendre...decision, equite en matiere d'emploi, and bourse respectively. Testing Champollion on three years' worth of the Hansards corpus yielded the French translations of 300 collocations for each year, evaluated at 73% accuracy on average. In this paper, we describe the statistical measures used, the algorithm, and the implementation of Champollion, presenting our results and evaluation.
众所周知,搭配对于非母语人士来说很难翻译,主要是因为它们不透明,不能逐字翻译。我们描述了一个名为Champollion的程序,给定两种不同语言的一对平行语料库和其中一种语言的搭配列表,该程序自动生成它们的翻译。我们的目标是提供一种工具,用于在多种语言的不同领域中编译单词级别以上的双语词汇信息。我们使用的算法基于统计方法,并产生n个单词搭配的p个单词翻译,其中n和p不必相同。例如,商博良将make…decision、employment equity和stock market分别翻译成prenre…decision、equite en matiere d'emploi和bourse。商博良用三年的汉莎语料库进行测试,每年得出300个搭配的法语翻译,平均准确率为73%。在本文中,我们描述了使用的统计度量、算法和商博良的实现,并给出了我们的结果和评价。
{"title":"Translating Collocations for Bilingual Lexicons: A Statistical Approach","authors":"Frank Smadja, K. McKeown, V. Hatzivassiloglou","doi":"10.7916/D8C82M3R","DOIUrl":"https://doi.org/10.7916/D8C82M3R","url":null,"abstract":"Collocations are notoriously difficult for non-native speakers to translate, primarily because they are opaque and cannot be translated on a word-by-word basis. We describe a program named Champollion which, given a pair of parallel corpora in two different languages and a list of collocations in one of them, automatically produces their translations. Our goal is to provide a tool for compiling bilingual lexical information above the word level in multiple languages, for different domains. The algorithm we use is based on statistical methods and produces p-word translations of n-word collocations in which n and p need not be the same. For example, Champollion translates make...decision, employment equity, and stock market into prendre...decision, equite en matiere d'emploi, and bourse respectively. Testing Champollion on three years' worth of the Hansards corpus yielded the French translations of 300 collocations for each year, evaluated at 73% accuracy on average. In this paper, we describe the statistical measures used, the algorithm, and the implementation of Champollion, presenting our results and evaluation.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1996-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131178322","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 578
The Generative Lexicon 生成式词典
Pub Date : 1995-10-23 DOI: 10.2307/415891
J. Pustejovsky
In this paper, I will discuss four major topics relating to current research in lexical semantics: methodology, descriptive coverage, adequacy of the representation, and the computational usefulness of representations. In addressing these issues, I will discuss what I think are some of the central problems facing the lexical semantics community, and suggest ways of best approaching these issues. Then, I will provide a method for the decomposition of lexical categories and outline a theory of lexical semantics embodying a notion of cocompositionality and type coercion, as well as several levels of semantic description, where the semantic load is spread more evenly throughout the lexicon. I argue that lexical decomposition is possible if it is performed generatively. Rather than assuming a fixed set of primitives. I will assume a fixed number of generative devices that can be seen as constructing semantic expressions. I develop a theory of Qualia Structure, a representation language for lexical items, which renders much lexical ambiguity in the lexicon unnecessary, while still explaining the systematic polysemy that words carry. Finally, I discuss how individual lexical structures can be integrated into the larger lexical knowledge base through a theory of lexical inheritance. This provides us with the necessary principles of global organization for the lexicon, enabling us to fully integrate our natural language lexicon into a conceptual whole.
在本文中,我将讨论与当前词汇语义研究相关的四个主要主题:方法论、描述覆盖、表示的充分性和表示的计算有用性。在解决这些问题时,我将讨论我认为词汇语义社区面临的一些核心问题,并提出解决这些问题的最佳方法。然后,我将提供一种分解词汇类别的方法,并概述一个包含共组合性和类型强制概念的词汇语义理论,以及语义描述的几个层次,其中语义负载在整个词汇中更均匀地分布。我认为,如果词法分解是生成的,那么它是可能的。而不是假设一组固定的原语。我将假设有固定数量的生成设备,它们可以被视为构建语义表达式。我发展了一种质质结构理论,这是一种词汇项目的表征语言,它使词汇中的许多词汇歧义变得不必要,同时仍然解释了单词所携带的系统的多义性。最后,我讨论了如何通过词汇继承理论将单个词汇结构整合到更大的词汇知识库中。这为我们提供了词汇全局组织的必要原则,使我们能够将自然语言词汇充分整合为一个概念整体。
{"title":"The Generative Lexicon","authors":"J. Pustejovsky","doi":"10.2307/415891","DOIUrl":"https://doi.org/10.2307/415891","url":null,"abstract":"In this paper, I will discuss four major topics relating to current research in lexical semantics: methodology, descriptive coverage, adequacy of the representation, and the computational usefulness of representations. In addressing these issues, I will discuss what I think are some of the central problems facing the lexical semantics community, and suggest ways of best approaching these issues. Then, I will provide a method for the decomposition of lexical categories and outline a theory of lexical semantics embodying a notion of cocompositionality and type coercion, as well as several levels of semantic description, where the semantic load is spread more evenly throughout the lexicon. I argue that lexical decomposition is possible if it is performed generatively. Rather than assuming a fixed set of primitives. I will assume a fixed number of generative devices that can be seen as constructing semantic expressions. I develop a theory of Qualia Structure, a representation language for lexical items, which renders much lexical ambiguity in the lexicon unnecessary, while still explaining the systematic polysemy that words carry. Finally, I discuss how individual lexical structures can be integrated into the larger lexical knowledge base through a theory of lexical inheritance. This provides us with the necessary principles of global organization for the lexicon, enabling us to fully integrate our natural language lexicon into a conceptual whole.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"114 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124546863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3928
Indexical Expressions in the Scope of Attitude Verbs 态度动词范围内的引语表达
Pub Date : 1993-12-01 DOI: 10.5555/972500.972503
A. Haas
A sentential theory of attitudes holds that propositions (the things that agents believe and know) are sentences of a representation language. Given such a theory, it is natural to suggest that the proposition expressed by an utterance of natural language is also a sentence of a representation language. This leads to a straightforward account of the semantics of attitude verbs. However, this kind of theory encounters problems in dealing with indexicals: expressions such as "I," "here," and "now." It is hard to explain how an indexical in the scope of an attitude verb can be opaque. This paper suggests that while the propositions that agents believe and know are sentences, the propositions expressed by utterances are not sentences: they are singular propositions, of the type used in Kaplan's theory of direct reference. Drawing on recent work in situation semantics, this paper shows how such a theory can describe the semantics of attitude verbs and account for the opacity of indexicals in the scope of these verbs.
态度的句子理论认为命题(主体相信和知道的事物)是表征语言的句子。有了这样的理论,自然语言的话语所表达的命题也是表征语言的句子,这是很自然的。这导致了对态度动词语义的直接描述。然而,这种理论在处理索引时遇到了问题:诸如“我”、“这里”和“现在”之类的表达式。很难解释为什么态度动词范围内的索引是不透明的。本文认为,虽然主体相信和知道的命题是句子,但话语表达的命题不是句子:它们是单数命题,属于卡普兰直接指称理论中使用的类型。根据最近在情境语义学方面的工作,本文展示了这种理论如何描述态度动词的语义,并解释了这些动词范围内索引的不透明性。
{"title":"Indexical Expressions in the Scope of Attitude Verbs","authors":"A. Haas","doi":"10.5555/972500.972503","DOIUrl":"https://doi.org/10.5555/972500.972503","url":null,"abstract":"A sentential theory of attitudes holds that propositions (the things that agents believe and know) are sentences of a representation language. Given such a theory, it is natural to suggest that the proposition expressed by an utterance of natural language is also a sentence of a representation language. This leads to a straightforward account of the semantics of attitude verbs. However, this kind of theory encounters problems in dealing with indexicals: expressions such as \"I,\" \"here,\" and \"now.\" It is hard to explain how an indexical in the scope of an attitude verb can be opaque. This paper suggests that while the propositions that agents believe and know are sentences, the propositions expressed by utterances are not sentences: they are singular propositions, of the type used in Kaplan's theory of direct reference. Drawing on recent work in situation semantics, this paper shows how such a theory can describe the semantics of attitude verbs and account for the opacity of indexicals in the scope of these verbs.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1993-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132461217","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Empirical Studies on the Disambiguation of Cue Phrases 暗示短语消歧的实证研究
Pub Date : 1993-09-01 DOI: 10.7916/D8PV6TGB
Julia Hirschberg, D. Litman
Cue phrases are linguistic expressions such as now and well that function as explicit indicators of the structure of a discourse. For example, now may signal the beginning of a subtopic or a return to a previous topic, while well may mark subsequent material as a response to prior material, or as an explanatory comment. However, while cue phrases may convey discourse structure, each also has one or more alternate uses. While incidentally may be used sententially as an adverbial, for example, the discourse use initiates a digression. Although distinguishing discourse and sentential uses of cue phrases is critical to the interpretation and generation of discourse, the question of how speakers and hearers accomplish this disambiguation is rarely addressed.This paper reports results of empirical studies on discourse and sentential uses of cue phrases, in which both text-based and prosodic features were examined for disambiguating power. Based on these studies, it is proposed that discourse versus sentential usage may be distinguished by intonational features, specifically, pitch accent and prosodic phrasing. A prosodic model that characterizes these distinctions is identified. This model is associated with features identifiable from text analysis, including orthography and part of speech, to permit the application of the results of the prosodic analysis to the generation of appropriate intonational features for discourse and sentential uses of cue phrases in synthetic speech.
提示短语是语言表达,如now和well,其功能是话语结构的明确指示。例如,now可能标志着副标题的开始或回到上一个主题,而well可能将后续材料标记为对先前材料的回应,或作为解释性评论。然而,虽然提示短语可以传达话语结构,但每个提示短语也有一个或多个替代用途。例如,虽然它可以在句子中用作副词,但它在语篇中的使用会引发离题。虽然区分语篇和提示短语的句子使用对语篇的解释和生成至关重要,但说话者和听者如何完成这种消歧的问题却很少得到解决。本文报告了线索短语语篇和句子使用的实证研究结果,其中基于文本的和韵律的两种特征都考察了线索短语的消歧能力。在这些研究的基础上,我们提出语篇和句子的使用可以通过语调特征来区分,特别是音高重音和韵律短语。确定了表征这些区别的韵律模型。该模型与文本分析中可识别的特征相关联,包括正字法和词性,以允许将韵律分析的结果应用于合成语音中提示短语的话语和句子使用的适当语调特征的生成。
{"title":"Empirical Studies on the Disambiguation of Cue Phrases","authors":"Julia Hirschberg, D. Litman","doi":"10.7916/D8PV6TGB","DOIUrl":"https://doi.org/10.7916/D8PV6TGB","url":null,"abstract":"Cue phrases are linguistic expressions such as now and well that function as explicit indicators of the structure of a discourse. For example, now may signal the beginning of a subtopic or a return to a previous topic, while well may mark subsequent material as a response to prior material, or as an explanatory comment. However, while cue phrases may convey discourse structure, each also has one or more alternate uses. While incidentally may be used sententially as an adverbial, for example, the discourse use initiates a digression. Although distinguishing discourse and sentential uses of cue phrases is critical to the interpretation and generation of discourse, the question of how speakers and hearers accomplish this disambiguation is rarely addressed.This paper reports results of empirical studies on discourse and sentential uses of cue phrases, in which both text-based and prosodic features were examined for disambiguating power. Based on these studies, it is proposed that discourse versus sentential usage may be distinguished by intonational features, specifically, pitch accent and prosodic phrasing. A prosodic model that characterizes these distinctions is identified. This model is associated with features identifiable from text analysis, including orthography and part of speech, to permit the application of the results of the prosodic analysis to the generation of appropriate intonational features for discourse and sentential uses of cue phrases in synthetic speech.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1993-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128635709","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 366
Building a Large Annotated Corpus of English: The Penn Treebank 构建大型带注释的英语语料库:宾州树库
Pub Date : 1993-06-01 DOI: 10.21236/ada273556
Mitchell P. Marcus, Beatrice Santorini, Mary Ann Marcinkiewicz
Abstract : As a result of this grant, the researchers have now published oil CDROM a corpus of over 4 million words of running text annotated with part-of- speech (POS) tags, with over 3 million words of that material assigned skeletal grammatical structure. This material now includes a fully hand-parsed version of the classic Brown corpus. About one half of the papers at the ACL Workshop on Using Large Text Corpora this past summer were based on the materials generated by this grant.
摘要:由于这项资助,研究人员现在已经发表了一个超过400万字的带有词性标注(POS)标签的运行文本语料库,其中超过300万字的语料库被分配了骨架语法结构。该材料现在包括经典布朗语料库的完全手工解析版本。在刚刚过去的这个夏天,在使用大文本语料库的ACL研讨会上,大约有一半的论文是基于这项资助产生的材料。
{"title":"Building a Large Annotated Corpus of English: The Penn Treebank","authors":"Mitchell P. Marcus, Beatrice Santorini, Mary Ann Marcinkiewicz","doi":"10.21236/ada273556","DOIUrl":"https://doi.org/10.21236/ada273556","url":null,"abstract":"Abstract : As a result of this grant, the researchers have now published oil CDROM a corpus of over 4 million words of running text annotated with part-of- speech (POS) tags, with over 3 million words of that material assigned skeletal grammatical structure. This material now includes a fully hand-parsed version of the classic Brown corpus. About one half of the papers at the ACL Workshop on Using Large Text Corpora this past summer were based on the materials generated by this grant.","PeriodicalId":360119,"journal":{"name":"Comput. Linguistics","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1993-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121284573","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8621
期刊
Comput. Linguistics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1