首页 > 最新文献

VS@HLT-NAACL最新文献

英文 中文
Dependency Link Embeddings: Continuous Representations of Syntactic Substructures 依赖链接嵌入:语法子结构的连续表示
Pub Date : 2015-06-01 DOI: 10.3115/v1/W15-1514
Mohit Bansal
We present a simple method to learn continuous representations of dependency substructures (links), with the motivation of directly working with higher-order, structured embeddings and their hidden relationships, and also to avoid the millions of sparse, template-based word-cluster features in dependency parsing. These link embeddings allow a significantly smaller and simpler set of unary features for dependency parsing, while maintaining improvements similar to state-of-the-art, n-ary word-cluster features, and also stacking over them. Moreover, these link vectors (made publicly available) are directly portable as offthe-shelf, dense, syntactic features in various NLP tasks. As one example, we incorporate them into constituent parse reranking, where their small feature set again matches the performance of standard non-local, manuallydefined features, and also stacks over them.
我们提出了一种简单的方法来学习依赖子结构(链接)的连续表示,其动机是直接处理高阶结构化嵌入及其隐藏关系,并避免依赖解析中数百万个稀疏的、基于模板的词簇特征。这些链接嵌入允许更小、更简单的一元特性集用于依赖项解析,同时保持类似于最先进的n元词簇特性的改进,并在它们之上进行叠加。此外,这些链接向量(公开可用)可以直接移植为各种NLP任务中的现成的、密集的语法特征。作为一个例子,我们将它们合并到组成解析重新排序中,其中它们的小特征集再次与标准的非局部、手动定义的特征的性能相匹配,并且还叠加在它们之上。
{"title":"Dependency Link Embeddings: Continuous Representations of Syntactic Substructures","authors":"Mohit Bansal","doi":"10.3115/v1/W15-1514","DOIUrl":"https://doi.org/10.3115/v1/W15-1514","url":null,"abstract":"We present a simple method to learn continuous representations of dependency substructures (links), with the motivation of directly working with higher-order, structured embeddings and their hidden relationships, and also to avoid the millions of sparse, template-based word-cluster features in dependency parsing. These link embeddings allow a significantly smaller and simpler set of unary features for dependency parsing, while maintaining improvements similar to state-of-the-art, n-ary word-cluster features, and also stacking over them. Moreover, these link vectors (made publicly available) are directly portable as offthe-shelf, dense, syntactic features in various NLP tasks. As one example, we incorporate them into constituent parse reranking, where their small feature set again matches the performance of standard non-local, manuallydefined features, and also stacks over them.","PeriodicalId":299646,"journal":{"name":"VS@HLT-NAACL","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121857893","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Named Entity Recognition for Arabic Social Media 阿拉伯社交媒体的命名实体识别
Pub Date : 2015-06-01 DOI: 10.3115/v1/W15-1524
Ayah Zirikly, Mona T. Diab
The majority of research on Arabic Named Entity Recognition (NER) addresses the the task for newswire genre, where the language used is Modern Standard Arabic (MSA), however, the need to study this task in social media is becoming more vital. Social media is characterized by the use of both MSA and Dialectal Arabic (DA), with often code switching between the two language varieties. Despite some common characteristics between MSA and DA, there are significant differences between which result in poor performance when MSA targeting systems are applied for NER in DA. Additionally, most NER systems rely primarily on gazetteers, which can be more challenging in a social media processing context due to an inherent low coverage. In this paper, we present a gazetteers-free NER system for Dialectal data that yields an F1 score of 72.68% which is an absolute improvement of 2 3% over a comparable state-ofthe-art gazetteer based DA-NER system.
大多数关于阿拉伯语命名实体识别(NER)的研究解决了新闻专线类型的任务,其中使用的语言是现代标准阿拉伯语(MSA),然而,在社交媒体中研究这一任务的需求变得越来越重要。社交媒体的特点是同时使用MSA和方言阿拉伯语(DA),并经常在两种语言之间进行代码转换。尽管MSA和DA之间存在一些共同的特征,但它们之间存在显著差异,导致MSA靶向系统应用于数据处理中的NER时性能不佳。此外,大多数NER系统主要依赖于地名词典,由于固有的低覆盖率,这在社交媒体处理上下文中可能更具挑战性。在本文中,我们提出了一个方言数据的无地名词典NER系统,该系统产生了72.68%的F1分数,与同类的最先进的基于地名词典的DA-NER系统相比,绝对提高了2.3%。
{"title":"Named Entity Recognition for Arabic Social Media","authors":"Ayah Zirikly, Mona T. Diab","doi":"10.3115/v1/W15-1524","DOIUrl":"https://doi.org/10.3115/v1/W15-1524","url":null,"abstract":"The majority of research on Arabic Named Entity Recognition (NER) addresses the the task for newswire genre, where the language used is Modern Standard Arabic (MSA), however, the need to study this task in social media is becoming more vital. Social media is characterized by the use of both MSA and Dialectal Arabic (DA), with often code switching between the two language varieties. Despite some common characteristics between MSA and DA, there are significant differences between which result in poor performance when MSA targeting systems are applied for NER in DA. Additionally, most NER systems rely primarily on gazetteers, which can be more challenging in a social media processing context due to an inherent low coverage. In this paper, we present a gazetteers-free NER system for Dialectal data that yields an F1 score of 72.68% which is an absolute improvement of 2 3% over a comparable state-ofthe-art gazetteer based DA-NER system.","PeriodicalId":299646,"journal":{"name":"VS@HLT-NAACL","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125131906","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 57
Towards Combined Matrix and Tensor Factorization for Universal Schema Relation Extraction 面向通用模式关系提取的组合矩阵和张量分解
Pub Date : 2015-06-01 DOI: 10.3115/v1/W15-1519
Sameer Singh, Tim Rocktäschel, S. Riedel
Matrix factorization of knowledge bases in universal schema has facilitated accurate distantlysupervised relation extraction. This factorization encodes dependencies between textual patterns and structured relations using lowdimensional vectors defined for each entity pair; although these factors are effective at combining evidence for an entity pair, they are inaccurate on rare pairs, or for relations that depend crucially on the entity types. On the other hand, tensor factorization is able to overcome these shortcomings when applied to link prediction by maintaining entity-wise factors. However these models have been unsuitable for universal schema. In this paper we first present an illustration on synthetic data that explains the unsuitability of tensor factorization to relation extraction with universal schemas. Since the benefits of tensor and matrix factorization are complementary, we then investigate two hybrid methods that combine the benefits of the two paradigms. We show that the combination can be fruitful: we handle ambiguously phrased relations, achieve gains in accuracy on real-world relations, and demonstrate that entity embeddings encode entity types.
通用模式中知识库的矩阵分解有助于精确的远程监督关系提取。这种分解使用为每个实体对定义的低维向量编码文本模式和结构化关系之间的依赖关系;尽管这些因素在组合实体对的证据方面是有效的,但对于稀有对或关键依赖于实体类型的关系,它们是不准确的。另一方面,张量分解可以通过保持实体因子来克服这些缺点。然而,这些模型并不适合通用模式。在本文中,我们首先给出了一个合成数据的例子,解释了张量分解在通用模式下关系提取中的不适用性。由于张量分解和矩阵分解的优点是互补的,因此我们研究了结合两种范式优点的两种混合方法。我们证明了这种组合是有成效的:我们处理了措辞含糊的关系,在现实世界的关系上获得了准确性的提高,并证明了实体嵌入对实体类型进行编码。
{"title":"Towards Combined Matrix and Tensor Factorization for Universal Schema Relation Extraction","authors":"Sameer Singh, Tim Rocktäschel, S. Riedel","doi":"10.3115/v1/W15-1519","DOIUrl":"https://doi.org/10.3115/v1/W15-1519","url":null,"abstract":"Matrix factorization of knowledge bases in universal schema has facilitated accurate distantlysupervised relation extraction. This factorization encodes dependencies between textual patterns and structured relations using lowdimensional vectors defined for each entity pair; although these factors are effective at combining evidence for an entity pair, they are inaccurate on rare pairs, or for relations that depend crucially on the entity types. On the other hand, tensor factorization is able to overcome these shortcomings when applied to link prediction by maintaining entity-wise factors. However these models have been unsuitable for universal schema. In this paper we first present an illustration on synthetic data that explains the unsuitability of tensor factorization to relation extraction with universal schemas. Since the benefits of tensor and matrix factorization are complementary, we then investigate two hybrid methods that combine the benefits of the two paradigms. We show that the combination can be fruitful: we handle ambiguously phrased relations, achieve gains in accuracy on real-world relations, and demonstrate that entity embeddings encode entity types.","PeriodicalId":299646,"journal":{"name":"VS@HLT-NAACL","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123999065","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
DeepNL: a Deep Learning NLP pipeline DeepNL:一个深度学习NLP管道
Pub Date : 2015-06-01 DOI: 10.3115/v1/W15-1515
Giuseppe Attardi
We present the architecture of a deep learning pipeline for natural language processing. Based on this architecture we built a set of tools both for creating distributional vector representations and for performing specific NLP tasks. Three methods are available for creating embeddings: feedforward neural network, sentiment specific embeddings and embeddings based on counts and Hellinger PCA. Two methods are provided for training a network to perform sequence tagging, a window approach and a convolutional approach. The window approach is used for implementing a POS tagger and a NER tagger, the convolutional network is used for Semantic Role Labeling. The library is implemented in Python with core numerical processing written in C++ using parallel linear algebra library for efficiency and scalability.
我们提出了一个用于自然语言处理的深度学习管道的架构。基于这个架构,我们构建了一组工具,用于创建分布向量表示和执行特定的NLP任务。有三种方法可用于创建嵌入:前馈神经网络、情感特定嵌入和基于计数和海灵格主成分分析的嵌入。提供了两种方法来训练网络执行序列标记,窗口方法和卷积方法。使用窗口方法实现POS标注器和NER标注器,使用卷积网络实现语义角色标注。该库是用Python实现的,核心数值处理用c++编写,使用并行线性代数库来提高效率和可扩展性。
{"title":"DeepNL: a Deep Learning NLP pipeline","authors":"Giuseppe Attardi","doi":"10.3115/v1/W15-1515","DOIUrl":"https://doi.org/10.3115/v1/W15-1515","url":null,"abstract":"We present the architecture of a deep learning pipeline for natural language processing. Based on this architecture we built a set of tools both for creating distributional vector representations and for performing specific NLP tasks. Three methods are available for creating embeddings: feedforward neural network, sentiment specific embeddings and embeddings based on counts and Hellinger PCA. Two methods are provided for training a network to perform sequence tagging, a window approach and a convolutional approach. The window approach is used for implementing a POS tagger and a NER tagger, the convolutional network is used for Semantic Role Labeling. The library is implemented in Python with core numerical processing written in C++ using parallel linear algebra library for efficiency and scalability.","PeriodicalId":299646,"journal":{"name":"VS@HLT-NAACL","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128333730","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Neural word embeddings with multiplicative feature interactions for tensor-based compositions 基于张量的组合中具有乘法特征交互的神经词嵌入
Pub Date : 2015-06-01 DOI: 10.3115/v1/W15-1520
Joo-Kyung Kim, M. Marneffe, E. Fosler-Lussier
Categorical compositional distributional models unify compositional formal semantic models and distributional models by composing phrases with tensor-based methods from vector representations. For the tensor-based compositions, Milajevs et al. (2014) showed that word vectors obtained from the continuous bag-of-words (CBOW) model are competitive with those from co-occurrence based models. However, because word vectors from the CBOW model are trained assuming additive interactions between context words, the word composition used for the training mismatches to the tensor-based methods used for evaluating the actual compositions including pointwise multiplication and tensor product of context vectors. In this work, we show whether the word embeddings from extended CBOW models using multiplication or tensor product between context words, reflecting the actual composition methods, can show better performance than those from the baseline CBOW model in actual tasks of compositions with multiplication or tensor-based methods.
范畴组合分布模型将组合形式语义模型和分布模型统一起来,利用基于张量的向量表示方法组合短语。对于基于张量的组合,Milajevs等人(2014)表明,从连续词袋(CBOW)模型获得的词向量与基于共现模型获得的词向量具有竞争力。然而,由于来自CBOW模型的词向量是假设上下文词之间的加性相互作用来训练的,因此用于训练的词组成与用于评估实际组成的基于张量的方法(包括上下文向量的点乘法和张量积)不匹配。在这项工作中,我们展示了使用上下文词之间的乘法或张量积的扩展CBOW模型的词嵌入是否能够在使用乘法或基于张量的方法的组合的实际任务中表现出比基线CBOW模型更好的性能。
{"title":"Neural word embeddings with multiplicative feature interactions for tensor-based compositions","authors":"Joo-Kyung Kim, M. Marneffe, E. Fosler-Lussier","doi":"10.3115/v1/W15-1520","DOIUrl":"https://doi.org/10.3115/v1/W15-1520","url":null,"abstract":"Categorical compositional distributional models unify compositional formal semantic models and distributional models by composing phrases with tensor-based methods from vector representations. For the tensor-based compositions, Milajevs et al. (2014) showed that word vectors obtained from the continuous bag-of-words (CBOW) model are competitive with those from co-occurrence based models. However, because word vectors from the CBOW model are trained assuming additive interactions between context words, the word composition used for the training mismatches to the tensor-based methods used for evaluating the actual compositions including pointwise multiplication and tensor product of context vectors. In this work, we show whether the word embeddings from extended CBOW models using multiplication or tensor product between context words, reflecting the actual composition methods, can show better performance than those from the baseline CBOW model in actual tasks of compositions with multiplication or tensor-based methods.","PeriodicalId":299646,"journal":{"name":"VS@HLT-NAACL","volume":"151 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127011759","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A Word-Embedding-based Sense Index for Regular Polysemy Representation 一种基于词嵌入的规则多义表示语义索引
Pub Date : 2015-06-01 DOI: 10.3115/v1/W15-1510
Marco Del Tredici, Núria Bel
Comunicacio presentada a: 1st Workshop on Vector Space Modeling for Natural Language Processing, celebrada a Colorado, United States of America, del 31 de maig al 5 de juny de 2015.
Comunicacio presentada a:1st Workshop on Vector Space Modeling for Natural Language Processing, celebrada a a Colorado, United States of America, del 31 de maig al 5 de juny de 2015.
{"title":"A Word-Embedding-based Sense Index for Regular Polysemy Representation","authors":"Marco Del Tredici, Núria Bel","doi":"10.3115/v1/W15-1510","DOIUrl":"https://doi.org/10.3115/v1/W15-1510","url":null,"abstract":"Comunicacio presentada a: 1st Workshop on Vector Space Modeling for Natural Language Processing, celebrada a Colorado, United States of America, del 31 de maig al 5 de juny de 2015.","PeriodicalId":299646,"journal":{"name":"VS@HLT-NAACL","volume":"91 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131432008","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Distributional Representations of Words for Short Text Classification 用于短文本分类的词的分布表示
Pub Date : 2015-06-01 DOI: 10.3115/v1/W15-1505
Chenglong Ma, Weiqun Xu, Peijia Li, Yonghong Yan
Traditional supervised learning approaches to common NLP tasks depend heavily on manual annotation, which is labor intensive and time consuming, and often suffer from data sparseness. In this paper we show how to mitigate the problems in short text classification (STC) through word embeddings ‐ distributional representations of words learned from large unlabeled data. The word embeddings are trained from the entire English Wikipedia text. We assume that a short text document is a specific sample of one distribution in a Bayesian framework. A Gaussian process approach is used to model the distribution of words. The task of classification becomes a simple problem of selecting the most probable Gaussian distribution. This approach is compared with those based on the classical maximum entropy (MaxEnt) model and the Latent Dirichlet Allocation (LDA) approach. Our approach achieved better performance and also showed advantages in dealing with unseen words.
传统的监督学习方法在很大程度上依赖于人工标注,这是劳动密集型和耗时的,并且经常受到数据稀疏性的影响。在本文中,我们展示了如何通过词嵌入-从大量未标记数据中学习的词的分布表示来缓解短文本分类(STC)中的问题。单词嵌入是从整个英文维基百科文本中训练的。我们假设一个短文本文档是贝叶斯框架中一个分布的特定样本。使用高斯过程方法对单词的分布进行建模。分类任务变成了选择最可能的高斯分布的简单问题。将该方法与基于经典最大熵(MaxEnt)模型和潜在狄利克雷分配(LDA)方法进行了比较。我们的方法取得了更好的性能,并且在处理未见词方面也显示出优势。
{"title":"Distributional Representations of Words for Short Text Classification","authors":"Chenglong Ma, Weiqun Xu, Peijia Li, Yonghong Yan","doi":"10.3115/v1/W15-1505","DOIUrl":"https://doi.org/10.3115/v1/W15-1505","url":null,"abstract":"Traditional supervised learning approaches to common NLP tasks depend heavily on manual annotation, which is labor intensive and time consuming, and often suffer from data sparseness. In this paper we show how to mitigate the problems in short text classification (STC) through word embeddings ‐ distributional representations of words learned from large unlabeled data. The word embeddings are trained from the entire English Wikipedia text. We assume that a short text document is a specific sample of one distribution in a Bayesian framework. A Gaussian process approach is used to model the distribution of words. The task of classification becomes a simple problem of selecting the most probable Gaussian distribution. This approach is compared with those based on the classical maximum entropy (MaxEnt) model and the Latent Dirichlet Allocation (LDA) approach. Our approach achieved better performance and also showed advantages in dealing with unseen words.","PeriodicalId":299646,"journal":{"name":"VS@HLT-NAACL","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126921918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
Learning Distributed Representations for Multilingual Text Sequences 学习多语言文本序列的分布式表示
Pub Date : 2015-06-01 DOI: 10.3115/v1/W15-1512
Hieu Pham, Thang Luong, Christopher D. Manning
We propose a novel approach to learning distributed representations of variable-length text sequences in multiple languages simultaneously. Unlike previous work which often derive representations of multi-word sequences as weighted sums of individual word vectors, our model learns distributed representations for phrases and sentences as a whole. Our work is similar in spirit to the recent paragraph vector approach but extends to the bilingual context so as to efficiently encode meaning-equivalent text sequences of multiple languages in the same semantic space. Our learned embeddings achieve state-of-theart performance in the often used crosslingual document classification task (CLDC) with an accuracy of 92.7 for English to German and 91.5 for German to English. By learning text sequence representations as a whole, our model performs equally well in both classification directions in the CLDC task in which past work did not achieve.
我们提出了一种同时学习多种语言的变长文本序列的分布式表示的新方法。与以往的工作不同,我们的模型通常将多词序列的表示作为单个词向量的加权和,而将短语和句子作为一个整体来学习分布式表示。我们的工作在精神上与最近的段落向量方法相似,但扩展到双语上下文,以便在同一语义空间中有效地编码多种语言的意义等效文本序列。我们学习的嵌入在常用的跨语言文档分类任务(CLDC)中实现了最先进的性能,英语到德语的准确率为92.7,德语到英语的准确率为91.5。通过整体学习文本序列表示,我们的模型在CLDC任务的两个分类方向上都表现良好,这是过去的工作无法实现的。
{"title":"Learning Distributed Representations for Multilingual Text Sequences","authors":"Hieu Pham, Thang Luong, Christopher D. Manning","doi":"10.3115/v1/W15-1512","DOIUrl":"https://doi.org/10.3115/v1/W15-1512","url":null,"abstract":"We propose a novel approach to learning distributed representations of variable-length text sequences in multiple languages simultaneously. Unlike previous work which often derive representations of multi-word sequences as weighted sums of individual word vectors, our model learns distributed representations for phrases and sentences as a whole. Our work is similar in spirit to the recent paragraph vector approach but extends to the bilingual context so as to efficiently encode meaning-equivalent text sequences of multiple languages in the same semantic space. Our learned embeddings achieve state-of-theart performance in the often used crosslingual document classification task (CLDC) with an accuracy of 92.7 for English to German and 91.5 for German to English. By learning text sequence representations as a whole, our model performs equally well in both classification directions in the CLDC task in which past work did not achieve.","PeriodicalId":299646,"journal":{"name":"VS@HLT-NAACL","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114296315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 61
A Simple Word Embedding Model for Lexical Substitution 一个简单的词汇替换词嵌入模型
Pub Date : 2015-06-01 DOI: 10.3115/v1/W15-1501
Oren Melamud, Omer Levy, Ido Dagan
The lexical substitution task requires identifying meaning-preserving substitutes for a target word instance in a given sentential context. Since its introduction in SemEval-2007, various models addressed this challenge, mostly in an unsupervised setting. In this work we propose a simple model for lexical substitution, which is based on the popular skip-gram word embedding model. The novelty of our approach is in leveraging explicitly the context embeddings generated within the skip-gram model, which were so far considered only as an internal component of the learning process. Our model is efficient, very simple to implement, and at the same time achieves state-ofthe-art results on lexical substitution tasks in an unsupervised setting.
词汇替换任务需要在给定的句子上下文中为目标单词实例识别保留意义的替代品。自SemEval-2007中引入以来,各种模型解决了这一挑战,主要是在无监督的环境中。在这项工作中,我们提出了一个简单的词汇替换模型,该模型基于流行的skip-gram词嵌入模型。我们方法的新颖之处在于明确地利用了skip-gram模型中生成的上下文嵌入,到目前为止,上下文嵌入只被认为是学习过程的内部组成部分。我们的模型非常高效,实现起来非常简单,同时在无监督设置下的词汇替换任务上实现了最先进的结果。
{"title":"A Simple Word Embedding Model for Lexical Substitution","authors":"Oren Melamud, Omer Levy, Ido Dagan","doi":"10.3115/v1/W15-1501","DOIUrl":"https://doi.org/10.3115/v1/W15-1501","url":null,"abstract":"The lexical substitution task requires identifying meaning-preserving substitutes for a target word instance in a given sentential context. Since its introduction in SemEval-2007, various models addressed this challenge, mostly in an unsupervised setting. In this work we propose a simple model for lexical substitution, which is based on the popular skip-gram word embedding model. The novelty of our approach is in leveraging explicitly the context embeddings generated within the skip-gram model, which were so far considered only as an internal component of the learning process. Our model is efficient, very simple to implement, and at the same time achieves state-ofthe-art results on lexical substitution tasks in an unsupervised setting.","PeriodicalId":299646,"journal":{"name":"VS@HLT-NAACL","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127006929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 110
A Deep Architecture for Non-Projective Dependency Parsing 非投射依赖解析的深层体系结构
Pub Date : 2015-06-01 DOI: 10.3115/v1/W15-1508
E. Fonseca, S. Aluísio
Graph-based dependency parsing algorithms commonly employ features up to third order in an attempt to capture richer syntactic relations. However, each level and each feature combination must be defined manually. Besides that, input features are usually represented as huge, sparse binary vectors, offering limited generalization. In this work, we present a deep architecture for dependency parsing based on a convolutional neural network. It can examine the whole sentence structure before scoring each head/modifier candidate pair, and uses dense embeddings as input. Our model is still under ongoing work, achieving 91.6% unlabeled attachment score in the Penn Treebank.
基于图的依赖解析算法通常使用三阶特征,试图捕获更丰富的语法关系。但是,每个级别和每个特征组合必须手动定义。除此之外,输入特征通常被表示为巨大的、稀疏的二值向量,提供有限的泛化。在这项工作中,我们提出了一个基于卷积神经网络的依赖解析的深度架构。它可以在对每个头/修饰语候选对评分之前检查整个句子结构,并使用密集嵌入作为输入。我们的模型仍在进行中,在Penn Treebank中获得了91.6%的未标记附件分数。
{"title":"A Deep Architecture for Non-Projective Dependency Parsing","authors":"E. Fonseca, S. Aluísio","doi":"10.3115/v1/W15-1508","DOIUrl":"https://doi.org/10.3115/v1/W15-1508","url":null,"abstract":"Graph-based dependency parsing algorithms commonly employ features up to third order in an attempt to capture richer syntactic relations. However, each level and each feature combination must be defined manually. Besides that, input features are usually represented as huge, sparse binary vectors, offering limited generalization. In this work, we present a deep architecture for dependency parsing based on a convolutional neural network. It can examine the whole sentence structure before scoring each head/modifier candidate pair, and uses dense embeddings as input. Our model is still under ongoing work, achieving 91.6% unlabeled attachment score in the Penn Treebank.","PeriodicalId":299646,"journal":{"name":"VS@HLT-NAACL","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115896194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
VS@HLT-NAACL
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1