首页 > 最新文献

Special Interest Group on Computational Morphology and Phonology Workshop最新文献

英文 中文
Getting the ##life out of living: How Adequate Are Word-Pieces for Modelling Complex Morphology? 从生活中获得生命:对复杂形态建模的词块有多充分?
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.sigmorphon-1.24
Stav Klein, Reut Tsarfaty
This work investigates the most basic units that underlie contextualized word embeddings, such as BERT — the so-called word pieces. In Morphologically-Rich Languages (MRLs) which exhibit morphological fusion and non-concatenative morphology, the different units of meaning within a word may be fused, intertwined, and cannot be separated linearly. Therefore, when using word-pieces in MRLs, we must consider that: (1) a linear segmentation into sub-word units might not capture the full morphological complexity of words; and (2) representations that leave morphological knowledge on sub-word units inaccessible might negatively affect performance. Here we empirically examine the capacity of word-pieces to capture morphology by investigating the task of multi-tagging in Modern Hebrew, as a proxy to evaluate the underlying segmentation. Our results show that, while models trained to predict multi-tags for complete words outperform models tuned to predict the distinct tags of WPs, we can improve the WPs tag prediction by purposefully constraining the word-pieces to reflect their internal functions. We suggest that linguistically-informed word-pieces schemes, that make the morphological structure explicit, might boost performance for MRLs.
这项工作调查了构成语境化词嵌入的最基本单位,如BERT——即所谓的词块。在形态丰富的语言中,表现为形态融合和非连接形态,一个词中的不同意义单位可能是融合的,相互交织的,不能线性分离。因此,在MRLs中使用词块时,我们必须考虑到:(1)对子词单元的线性分割可能无法捕捉到词的全部形态复杂性;(2)子词单位的形态知识不可访问的表征可能会对性能产生负面影响。在这里,我们通过调查现代希伯来语中的多重标记任务,作为评估潜在分词的代理,实证研究了词块捕捉形态学的能力。我们的研究结果表明,虽然训练用于预测完整词的多标签的模型优于调整用于预测WPs的单独标签的模型,但我们可以通过有目的地约束词块来反映其内部功能来改进WPs标签预测。我们认为,使词形结构明确的语言知情词块方案可能会提高核磁共振成像的表现。
{"title":"Getting the ##life out of living: How Adequate Are Word-Pieces for Modelling Complex Morphology?","authors":"Stav Klein, Reut Tsarfaty","doi":"10.18653/v1/2020.sigmorphon-1.24","DOIUrl":"https://doi.org/10.18653/v1/2020.sigmorphon-1.24","url":null,"abstract":"This work investigates the most basic units that underlie contextualized word embeddings, such as BERT — the so-called word pieces. In Morphologically-Rich Languages (MRLs) which exhibit morphological fusion and non-concatenative morphology, the different units of meaning within a word may be fused, intertwined, and cannot be separated linearly. Therefore, when using word-pieces in MRLs, we must consider that: (1) a linear segmentation into sub-word units might not capture the full morphological complexity of words; and (2) representations that leave morphological knowledge on sub-word units inaccessible might negatively affect performance. Here we empirically examine the capacity of word-pieces to capture morphology by investigating the task of multi-tagging in Modern Hebrew, as a proxy to evaluate the underlying segmentation. Our results show that, while models trained to predict multi-tags for complete words outperform models tuned to predict the distinct tags of WPs, we can improve the WPs tag prediction by purposefully constraining the word-pieces to reflect their internal functions. We suggest that linguistically-informed word-pieces schemes, that make the morphological structure explicit, might boost performance for MRLs.","PeriodicalId":186158,"journal":{"name":"Special Interest Group on Computational Morphology and Phonology Workshop","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127885696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
Data Augmentation for Transformer-based G2P 基于变压器的G2P数据增强
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.sigmorphon-1.21
Zach Ryan, Mans Hulden
The Transformer model has been shown to outperform other neural seq2seq models in several character-level tasks. It is unclear, however, if the Transformer would benefit as much as other seq2seq models from data augmentation strategies in the low-resource setting. In this paper we explore strategies for data augmentation in the g2p task together with the Transformer model. Our results show that a relatively simple alignment-based strategy of identifying consistent input-output subsequences in grapheme-phoneme data coupled together with a subsequent splicing together of such pieces to generate hallucinated data works well in the low-resource setting, often delivering substantial performance improvement over a standard Transformer model.
Transformer模型已被证明在几个字符级任务中优于其他神经seq2seq模型。然而,在低资源环境下,Transformer是否能像其他seq2seq模型一样从数据增强策略中获益还不清楚。在本文中,我们将与Transformer模型一起探讨g2p任务中的数据增强策略。我们的研究结果表明,一种相对简单的基于对齐的策略,即在字形-音素数据中识别一致的输入-输出子序列,然后将这些片段拼接在一起以产生幻觉数据,在低资源环境中效果很好,通常比标准Transformer模型提供实质性的性能改进。
{"title":"Data Augmentation for Transformer-based G2P","authors":"Zach Ryan, Mans Hulden","doi":"10.18653/v1/2020.sigmorphon-1.21","DOIUrl":"https://doi.org/10.18653/v1/2020.sigmorphon-1.21","url":null,"abstract":"The Transformer model has been shown to outperform other neural seq2seq models in several character-level tasks. It is unclear, however, if the Transformer would benefit as much as other seq2seq models from data augmentation strategies in the low-resource setting. In this paper we explore strategies for data augmentation in the g2p task together with the Transformer model. Our results show that a relatively simple alignment-based strategy of identifying consistent input-output subsequences in grapheme-phoneme data coupled together with a subsequent splicing together of such pieces to generate hallucinated data works well in the low-resource setting, often delivering substantial performance improvement over a standard Transformer model.","PeriodicalId":186158,"journal":{"name":"Special Interest Group on Computational Morphology and Phonology Workshop","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122478523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
The UniMelb Submission to the SIGMORPHON 2020 Shared Task 0: Typologically Diverse Morphological Inflection UniMelb提交给SIGMORPHON 2020共享任务0:类型学上多样的形态变化
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.sigmorphon-1.20
Andreas Scherbakov
The paper describes the University of Melbourne’s submission to the SIGMORPHON 2020 Shared Task 0: Typologically Diverse Morphological Inflection. Our team submitted three systems in total, two neural and one non-neural. Our analysis of systems’ performance shows positive effects of newly introduced data hallucination technique that we employed in one of neural systems, especially in low-resource scenarios. A non-neural system based on observed inflection patterns shows optimistic results even in its simple implementation (>75% accuracy for 50% of languages). With possible improvement within the same modeling principle, accuracy might grow to values above 90%.
这篇论文描述了墨尔本大学提交给SIGMORPHON 2020共享任务0:类型学上多样的形态变化。我们团队总共提交了三个系统,两个神经系统和一个非神经系统。我们对系统性能的分析表明,我们在一个神经系统中采用的新引入的数据幻觉技术具有积极的效果,特别是在资源匮乏的情况下。基于观察到的屈折模式的非神经系统即使在简单的实现中也显示出乐观的结果(50%的语言有75%的准确率)。在相同的建模原则下进行可能的改进,准确度可能会增长到90%以上的值。
{"title":"The UniMelb Submission to the SIGMORPHON 2020 Shared Task 0: Typologically Diverse Morphological Inflection","authors":"Andreas Scherbakov","doi":"10.18653/v1/2020.sigmorphon-1.20","DOIUrl":"https://doi.org/10.18653/v1/2020.sigmorphon-1.20","url":null,"abstract":"The paper describes the University of Melbourne’s submission to the SIGMORPHON 2020 Shared Task 0: Typologically Diverse Morphological Inflection. Our team submitted three systems in total, two neural and one non-neural. Our analysis of systems’ performance shows positive effects of newly introduced data hallucination technique that we employed in one of neural systems, especially in low-resource scenarios. A non-neural system based on observed inflection patterns shows optimistic results even in its simple implementation (>75% accuracy for 50% of languages). With possible improvement within the same modeling principle, accuracy might grow to values above 90%.","PeriodicalId":186158,"journal":{"name":"Special Interest Group on Computational Morphology and Phonology Workshop","volume":"160 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127315658","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
SIGMORPHON 2020 Task 0 System Description: ETH Zürich Team SIGMORPHON 2020 Task 0系统描述:ETH z<e:1> rich Team
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.sigmorphon-1.10
Martina Forster, Clara Meister
This paper presents our system for the SIGMORPHON 2020 Shared Task. We build off of the baseline systems, performing exact inference on models trained on language family data. Our systems return the globally best solution under these models. Our two systems achieve 80.9% and 75.6% accuracy on the test set. We ultimately find that, in this setting, exact inference does not seem to help or hinder the performance of morphological inflection generators, which stands in contrast to its affect on Neural Machine Translation (NMT) models.
本文介绍了我们的SIGMORPHON 2020共享任务系统。我们在基线系统的基础上构建,对经过语言族数据训练的模型进行精确的推断。我们的系统在这些模型下返回全球最佳解决方案。我们的两个系统在测试集上的准确率分别达到80.9%和75.6%。我们最终发现,在这种情况下,精确的推理似乎不会帮助或阻碍形态屈折生成器的性能,这与它对神经机器翻译(NMT)模型的影响形成鲜明对比。
{"title":"SIGMORPHON 2020 Task 0 System Description: ETH Zürich Team","authors":"Martina Forster, Clara Meister","doi":"10.18653/v1/2020.sigmorphon-1.10","DOIUrl":"https://doi.org/10.18653/v1/2020.sigmorphon-1.10","url":null,"abstract":"This paper presents our system for the SIGMORPHON 2020 Shared Task. We build off of the baseline systems, performing exact inference on models trained on language family data. Our systems return the globally best solution under these models. Our two systems achieve 80.9% and 75.6% accuracy on the test set. We ultimately find that, in this setting, exact inference does not seem to help or hinder the performance of morphological inflection generators, which stands in contrast to its affect on Neural Machine Translation (NMT) models.","PeriodicalId":186158,"journal":{"name":"Special Interest Group on Computational Morphology and Phonology Workshop","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114902020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Leveraging Principal Parts for Morphological Inflection 利用主成分进行形态变化
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.sigmorphon-1.17
L. Liu, Mans Hulden
This paper presents the submission by the CU Ling team from the University of Colorado to SIGMORPHON 2020 shared task 0 on morphological inflection. The task is to generate the target inflected word form given a lemma form and a target morphosyntactic description. Our system uses the Transformer architecture. Our overall approach is to treat the morphological inflection task as a paradigm cell filling problem and to design the system to leverage principal parts information for better morphological inflection when the training data is limited. We train one model for each language separately without external data. The overall average performance of our submission ranks the first in both average accuracy and Levenshtein distance from the gold inflection among all submissions including those using external resources.
本文介绍了科罗拉多大学的CU Ling团队向SIGMORPHON 2020提交的关于形态变化的共享任务0。任务是在给定引理形式和目标形态句法描述的情况下生成目标屈折词形。我们的系统使用Transformer架构。我们的总体方法是将形态学变形任务视为范式细胞填充问题,并设计系统在训练数据有限的情况下利用主成分信息进行更好的形态学变形。我们在没有外部数据的情况下为每种语言单独训练一个模型。我们提交的整体平均表现在所有提交(包括使用外部资源的提交)中,无论是平均准确率还是与黄金拐点的Levenshtein距离都排名第一。
{"title":"Leveraging Principal Parts for Morphological Inflection","authors":"L. Liu, Mans Hulden","doi":"10.18653/v1/2020.sigmorphon-1.17","DOIUrl":"https://doi.org/10.18653/v1/2020.sigmorphon-1.17","url":null,"abstract":"This paper presents the submission by the CU Ling team from the University of Colorado to SIGMORPHON 2020 shared task 0 on morphological inflection. The task is to generate the target inflected word form given a lemma form and a target morphosyntactic description. Our system uses the Transformer architecture. Our overall approach is to treat the morphological inflection task as a paradigm cell filling problem and to design the system to leverage principal parts information for better morphological inflection when the training data is limited. We train one model for each language separately without external data. The overall average performance of our submission ranks the first in both average accuracy and Levenshtein distance from the gold inflection among all submissions including those using external resources.","PeriodicalId":186158,"journal":{"name":"Special Interest Group on Computational Morphology and Phonology Workshop","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130785107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Multi-Tiered Strictly Local Functions 多层严格局部函数
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.sigmorphon-1.29
Phillip Burness, Kevin McMullin
Tier-based Strictly Local functions, as they have so far been defined, are equipped with just a single tier. In light of this fact, they are currently incapable of modelling simultaneous phonological processes that would require different tiers. In this paper we consider whether and how we can allow a single function to operate over more than one tier. We conclude that multiple tiers can and should be permitted, but that the relationships between them must be restricted in some way to avoid overgeneration. The particular restriction that we propose comes in two parts. First, each input element is associated with a set of tiers that on their own can fully determine what the element is mapped to. Second, the set of tiers associated to a given input element must form a strict superset-subset hierarchy. In this way, we can track multiple, related sources of information when deciding how to process a particular input element. We demonstrate that doing so enables simple and intuitive analyses to otherwise challenging phonological phenomena.
迄今为止所定义的基于层的严格本地函数只配备了一个层。鉴于这一事实,他们目前无法模拟需要不同层次的同时语音过程。在本文中,我们考虑是否以及如何允许单个函数在多个层上运行。我们的结论是,多层可以而且应该被允许,但是它们之间的关系必须以某种方式加以限制,以避免过度生成。我们提出的具体限制包括两个部分。首先,每个输入元素都与一组层相关联,这些层本身可以完全确定元素映射到什么。第二,与给定输入元素相关联的层集必须形成严格的超集-子集层次结构。通过这种方式,在决定如何处理特定的输入元素时,我们可以跟踪多个相关的信息源。我们证明,这样做可以使简单和直观的分析,否则具有挑战性的语音现象。
{"title":"Multi-Tiered Strictly Local Functions","authors":"Phillip Burness, Kevin McMullin","doi":"10.18653/v1/2020.sigmorphon-1.29","DOIUrl":"https://doi.org/10.18653/v1/2020.sigmorphon-1.29","url":null,"abstract":"Tier-based Strictly Local functions, as they have so far been defined, are equipped with just a single tier. In light of this fact, they are currently incapable of modelling simultaneous phonological processes that would require different tiers. In this paper we consider whether and how we can allow a single function to operate over more than one tier. We conclude that multiple tiers can and should be permitted, but that the relationships between them must be restricted in some way to avoid overgeneration. The particular restriction that we propose comes in two parts. First, each input element is associated with a set of tiers that on their own can fully determine what the element is mapped to. Second, the set of tiers associated to a given input element must form a strict superset-subset hierarchy. In this way, we can track multiple, related sources of information when deciding how to process a particular input element. We demonstrate that doing so enables simple and intuitive analyses to otherwise challenging phonological phenomena.","PeriodicalId":186158,"journal":{"name":"Special Interest Group on Computational Morphology and Phonology Workshop","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128991090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
KU-CST at the SIGMORPHON 2020 Task 2 on Unsupervised Morphological Paradigm Completion KU-CST在SIGMORPHON 2020任务2中的无监督形态范式完成
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.sigmorphon-1.11
Manex Agirrezabal, Jürgen Wedekind
We present a model for the unsupervised dis- covery of morphological paradigms. The goal of this model is to induce morphological paradigms from the bible (raw text) and a list of lemmas. We have created a model that splits each lemma in a stem and a suffix, and then we try to create a plausible suffix list by con- sidering lemma pairs. Our model was not able to outperform the official baseline, and there is still room for improvement, but we believe that the ideas presented here are worth considering.
我们提出了一个形态学范式的无监督发现模型。这个模型的目标是从圣经(原始文本)和引理列表中归纳出词形范式。我们已经创建了一个模型,该模型将词干中的每个引理和一个后缀分开,然后我们尝试通过考虑引理对来创建一个可信的后缀列表。我们的模型不能超过官方的基线,还有改进的空间,但我们认为这里提出的想法值得考虑。
{"title":"KU-CST at the SIGMORPHON 2020 Task 2 on Unsupervised Morphological Paradigm Completion","authors":"Manex Agirrezabal, Jürgen Wedekind","doi":"10.18653/v1/2020.sigmorphon-1.11","DOIUrl":"https://doi.org/10.18653/v1/2020.sigmorphon-1.11","url":null,"abstract":"We present a model for the unsupervised dis- covery of morphological paradigms. The goal of this model is to induce morphological paradigms from the bible (raw text) and a list of lemmas. We have created a model that splits each lemma in a stem and a suffix, and then we try to create a plausible suffix list by con- sidering lemma pairs. Our model was not able to outperform the official baseline, and there is still room for improvement, but we believe that the ideas presented here are worth considering.","PeriodicalId":186158,"journal":{"name":"Special Interest Group on Computational Morphology and Phonology Workshop","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114136943","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Linguist vs. Machine: Rapid Development of Finite-State Morphological Grammars 语言学家与机器:有限状态形态语法的快速发展
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.sigmorphon-1.18
Sarah Beemer, Zak Boston, April Bukoski, Daniel Chen, P. Dickens, Andrew Gerlach, Torin Hopkins, Parth Anand Jawale, Chris Koski, Akanksha Malhotra, Piyush Mishra, S. Muradoglu, Lan Sang, Tyler Short, Sagarika Shreevastava, Eliza Spaulding, Testumichi Umada, Beilei Xiang, Changbing Yang, Mans Hulden
Sequence-to-sequence models have proven to be highly successful in learning morphological inflection from examples as the series of SIGMORPHON/CoNLL shared tasks have shown. It is usually assumed, however, that a linguist working with inflectional examples could in principle develop a gold standard-level morphological analyzer and generator that would surpass a trained neural network model in accuracy of predictions, but that it may require significant amounts of human labor. In this paper, we discuss an experiment where a group of people with some linguistic training develop 25+ grammars as part of the shared task and weigh the cost/benefit ratio of developing grammars by hand. We also present tools that can help linguists triage difficult complex morphophonological phenomena within a language and hypothesize inflectional class membership. We conclude that a significant development effort by trained linguists to analyze and model morphophonological patterns are required in order to surpass the accuracy of neural models.
序列到序列模型已被证明在从示例中学习形态变化方面非常成功,如SIGMORPHON/CoNLL共享任务系列所示。然而,人们通常认为,研究屈折例子的语言学家原则上可以开发出一种金标准级别的形态分析器和生成器,其预测的准确性将超过经过训练的神经网络模型,但这可能需要大量的人力。在本文中,我们讨论了一个实验,在这个实验中,一群受过一定语言训练的人开发了25个以上的语法,作为共享任务的一部分,并权衡了手工开发语法的成本/收益比。我们还提供了一些工具,可以帮助语言学家在语言中分类困难的复杂词形音素现象,并假设屈折词类隶属关系。我们得出的结论是,为了超越神经模型的准确性,需要训练有素的语言学家在分析和建模词音模式方面做出重大的发展努力。
{"title":"Linguist vs. Machine: Rapid Development of Finite-State Morphological Grammars","authors":"Sarah Beemer, Zak Boston, April Bukoski, Daniel Chen, P. Dickens, Andrew Gerlach, Torin Hopkins, Parth Anand Jawale, Chris Koski, Akanksha Malhotra, Piyush Mishra, S. Muradoglu, Lan Sang, Tyler Short, Sagarika Shreevastava, Eliza Spaulding, Testumichi Umada, Beilei Xiang, Changbing Yang, Mans Hulden","doi":"10.18653/v1/2020.sigmorphon-1.18","DOIUrl":"https://doi.org/10.18653/v1/2020.sigmorphon-1.18","url":null,"abstract":"Sequence-to-sequence models have proven to be highly successful in learning morphological inflection from examples as the series of SIGMORPHON/CoNLL shared tasks have shown. It is usually assumed, however, that a linguist working with inflectional examples could in principle develop a gold standard-level morphological analyzer and generator that would surpass a trained neural network model in accuracy of predictions, but that it may require significant amounts of human labor. In this paper, we discuss an experiment where a group of people with some linguistic training develop 25+ grammars as part of the shared task and weigh the cost/benefit ratio of developing grammars by hand. We also present tools that can help linguists triage difficult complex morphophonological phenomena within a language and hypothesize inflectional class membership. We conclude that a significant development effort by trained linguists to analyze and model morphophonological patterns are required in order to surpass the accuracy of neural models.","PeriodicalId":186158,"journal":{"name":"Special Interest Group on Computational Morphology and Phonology Workshop","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114215390","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Exploring Neural Architectures And Techniques For Typologically Diverse Morphological Inflection 探索类型学上多样形态变化的神经结构和技术
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.sigmorphon-1.14
P. Jayarao, Siddhanth Pillay, P. Thombre, Aditi Chaudhary
Morphological inflection in low resource languages is critical to augment existing corpora in Low Resource Languages, which can help develop several applications in these languages with very good social impact. We describe our attention-based encoder-decoder approach that we implement using LSTMs and Transformers as the base units. We also describe the ancillary techniques that we experimented with, such as hallucination, language vector injection, sparsemax loss and adversarial language network alongside our approach to select the related language(s) for training. We present the results we generated on the constrained as well as unconstrained SIGMORPHON 2020 dataset (CITATION). One of the primary goals of our paper was to study the contribution varied components described above towards the performance of our system and perform an analysis on the same.
低资源语言的形态屈折变化是扩充低资源语言现有语料库的关键,有助于开发具有良好社会影响的低资源语言应用。我们描述了我们使用lstm和transformer作为基本单元实现的基于注意力的编码器-解码器方法。我们还描述了我们实验的辅助技术,如幻觉、语言向量注入、sparsemax损失和对抗性语言网络,以及我们选择相关语言进行训练的方法。我们展示了我们在有约束和无约束的SIGMORPHON 2020数据集上生成的结果(引文)。本文的主要目标之一是研究上述不同组件对系统性能的贡献,并对其进行分析。
{"title":"Exploring Neural Architectures And Techniques For Typologically Diverse Morphological Inflection","authors":"P. Jayarao, Siddhanth Pillay, P. Thombre, Aditi Chaudhary","doi":"10.18653/v1/2020.sigmorphon-1.14","DOIUrl":"https://doi.org/10.18653/v1/2020.sigmorphon-1.14","url":null,"abstract":"Morphological inflection in low resource languages is critical to augment existing corpora in Low Resource Languages, which can help develop several applications in these languages with very good social impact. We describe our attention-based encoder-decoder approach that we implement using LSTMs and Transformers as the base units. We also describe the ancillary techniques that we experimented with, such as hallucination, language vector injection, sparsemax loss and adversarial language network alongside our approach to select the related language(s) for training. We present the results we generated on the constrained as well as unconstrained SIGMORPHON 2020 dataset (CITATION). One of the primary goals of our paper was to study the contribution varied components described above towards the performance of our system and perform an analysis on the same.","PeriodicalId":186158,"journal":{"name":"Special Interest Group on Computational Morphology and Phonology Workshop","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114444865","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
One-Size-Fits-All Multilingual Models 一刀切的多语言模型
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.sigmorphon-1.4
Ben Peters, André F. T. Martins
This paper presents DeepSPIN’s submissions to Tasks 0 and 1 of the SIGMORPHON 2020 Shared Task. For both tasks, we present multilingual models, training jointly on data in all languages. We perform no language-specific hyperparameter tuning – each of our submissions uses the same model for all languages. Our basic architecture is the sparse sequence-to-sequence model with entmax attention and loss, which allows our models to learn sparse, local alignments while still being trainable with gradient-based techniques. For Task 1, we achieve strong performance with both RNN- and transformer-based sparse models. For Task 0, we extend our RNN-based model to a multi-encoder set-up in which separate modules encode the lemma and inflection sequences. Despite our models’ lack of language-specific tuning, they tie for first in Task 0 and place third in Task 1.
本文介绍了DeepSPIN提交给SIGMORPHON 2020共享任务的任务0和1。对于这两个任务,我们提出了多语言模型,对所有语言的数据进行联合训练。我们不执行特定于语言的超参数调优——我们的每个提交都对所有语言使用相同的模型。我们的基本架构是具有entmax注意力和损失的稀疏序列到序列模型,它允许我们的模型学习稀疏的局部对齐,同时仍然可以使用基于梯度的技术进行训练。对于任务1,我们使用基于RNN和基于变压器的稀疏模型都获得了较强的性能。对于任务0,我们将基于rnn的模型扩展到多编码器设置,其中单独的模块对引理和变形序列进行编码。尽管我们的模型缺乏特定于语言的调优,但它们在任务0中并列第一,在任务1中排名第三。
{"title":"One-Size-Fits-All Multilingual Models","authors":"Ben Peters, André F. T. Martins","doi":"10.18653/v1/2020.sigmorphon-1.4","DOIUrl":"https://doi.org/10.18653/v1/2020.sigmorphon-1.4","url":null,"abstract":"This paper presents DeepSPIN’s submissions to Tasks 0 and 1 of the SIGMORPHON 2020 Shared Task. For both tasks, we present multilingual models, training jointly on data in all languages. We perform no language-specific hyperparameter tuning – each of our submissions uses the same model for all languages. Our basic architecture is the sparse sequence-to-sequence model with entmax attention and loss, which allows our models to learn sparse, local alignments while still being trainable with gradient-based techniques. For Task 1, we achieve strong performance with both RNN- and transformer-based sparse models. For Task 0, we extend our RNN-based model to a multi-encoder set-up in which separate modules encode the lemma and inflection sequences. Despite our models’ lack of language-specific tuning, they tie for first in Task 0 and place third in Task 1.","PeriodicalId":186158,"journal":{"name":"Special Interest Group on Computational Morphology and Phonology Workshop","volume":"181 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124541513","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
期刊
Special Interest Group on Computational Morphology and Phonology Workshop
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1