首页 > 最新文献

Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics最新文献

英文 中文
The Development of Abstract Concepts in Children’s Early Lexical Networks 儿童早期词汇网络中抽象概念的发展
Abdellah Fourtassi, Isaac Scheinfeld, Michael C. Frank
How do children learn abstract concepts such as animal vs. artifact? Previous research has suggested that such concepts can partly be derived using cues from the language children hear around them. Following this suggestion, we propose a model where we represent the children’ developing lexicon as an evolving network. The nodes of this network are based on vocabulary knowledge as reported by parents, and the edges between pairs of nodes are based on the probability of their co-occurrence in a corpus of child-directed speech. We found that several abstract categories can be identified as the dense regions in such networks. In addition, our simulations suggest that these categories develop simultaneously, rather than sequentially, thanks to the children’s word learning trajectory which favors the exploration of the global conceptual space.
孩子们是如何学习抽象概念的,比如动物和神器?先前的研究表明,这些概念可以部分地从儿童听到的语言中获得线索。根据这一建议,我们提出了一个模型,我们将儿童词汇的发展表示为一个不断发展的网络。该网络的节点基于父母报告的词汇知识,节点对之间的边基于它们在儿童导向语音语料库中共同出现的概率。我们发现一些抽象的类别可以被识别为这种网络中的密集区域。此外,我们的模拟表明,由于儿童的单词学习轨迹有利于探索全局概念空间,这些类别是同时发展的,而不是顺序发展的。
{"title":"The Development of Abstract Concepts in Children’s Early Lexical Networks","authors":"Abdellah Fourtassi, Isaac Scheinfeld, Michael C. Frank","doi":"10.18653/v1/W19-2914","DOIUrl":"https://doi.org/10.18653/v1/W19-2914","url":null,"abstract":"How do children learn abstract concepts such as animal vs. artifact? Previous research has suggested that such concepts can partly be derived using cues from the language children hear around them. Following this suggestion, we propose a model where we represent the children’ developing lexicon as an evolving network. The nodes of this network are based on vocabulary knowledge as reported by parents, and the edges between pairs of nodes are based on the probability of their co-occurrence in a corpus of child-directed speech. We found that several abstract categories can be identified as the dense regions in such networks. In addition, our simulations suggest that these categories develop simultaneously, rather than sequentially, thanks to the children’s word learning trajectory which favors the exploration of the global conceptual space.","PeriodicalId":428409,"journal":{"name":"Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127609322","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Simulating Spanish-English Code-Switching: El Modelo Está Generating Code-Switches 模拟西班牙语-英语代码转换:El Modelo est生成代码转换
Chara Tsoukala, S. Frank, A. V. D. Bosch, Jorge R. Valdés Kroff, M. Broersma
Multilingual speakers are able to switch from one language to the other (“code-switch”) between or within sentences. Because the underlying cognitive mechanisms are not well understood, in this study we use computational cognitive modeling to shed light on the process of code-switching. We employed the Bilingual Dual-path model, a Recurrent Neural Network of bilingual sentence production (Tsoukala et al., 2017), and simulated sentence production in simultaneous Spanish-English bilinguals. Our first goal was to investigate whether the model would code-switch without being exposed to code-switched training input. The model indeed produced code-switches even without any exposure to such input and the patterns of code-switches are in line with earlier linguistic work (Poplack,1980). The second goal of this study was to investigate an auxiliary phrase asymmetry that exists in Spanish-English code-switched production. Using this cognitive model, we examined a possible cause for this asymmetry. To our knowledge, this is the first computational cognitive model that aims to simulate code-switched sentence production.
多语言使用者能够在句子之间或句子内部从一种语言切换到另一种语言(“代码切换”)。由于潜在的认知机制尚未被很好地理解,在本研究中,我们使用计算认知模型来阐明代码转换的过程。我们采用双语双路径模型,双语句子生成的递归神经网络(Tsoukala et al., 2017),并模拟了西班牙语-英语双语者同时的句子生成。我们的第一个目标是调查模型是否会在不暴露于代码切换训练输入的情况下进行代码切换。即使没有任何输入,该模型也确实产生了代码转换,并且代码转换的模式与早期的语言学工作一致(Poplack,1980)。本研究的第二个目的是调查在西班牙语-英语语码转换生产中存在的辅助短语不对称。利用这种认知模型,我们研究了这种不对称的可能原因。据我们所知,这是第一个旨在模拟代码转换句子生成的计算认知模型。
{"title":"Simulating Spanish-English Code-Switching: El Modelo Está Generating Code-Switches","authors":"Chara Tsoukala, S. Frank, A. V. D. Bosch, Jorge R. Valdés Kroff, M. Broersma","doi":"10.18653/v1/W19-2903","DOIUrl":"https://doi.org/10.18653/v1/W19-2903","url":null,"abstract":"Multilingual speakers are able to switch from one language to the other (“code-switch”) between or within sentences. Because the underlying cognitive mechanisms are not well understood, in this study we use computational cognitive modeling to shed light on the process of code-switching. We employed the Bilingual Dual-path model, a Recurrent Neural Network of bilingual sentence production (Tsoukala et al., 2017), and simulated sentence production in simultaneous Spanish-English bilinguals. Our first goal was to investigate whether the model would code-switch without being exposed to code-switched training input. The model indeed produced code-switches even without any exposure to such input and the patterns of code-switches are in line with earlier linguistic work (Poplack,1980). The second goal of this study was to investigate an auxiliary phrase asymmetry that exists in Spanish-English code-switched production. Using this cognitive model, we examined a possible cause for this asymmetry. To our knowledge, this is the first computational cognitive model that aims to simulate code-switched sentence production.","PeriodicalId":428409,"journal":{"name":"Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-04-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132553570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Dependency Parsing with your Eyes: Dependency Structure Predicts Eye Regressions During Reading 用眼睛分析依赖关系:依赖关系结构预测阅读过程中的眼睛退化
A. Lopopolo, S. Frank, Antal van den Bosch, Roel M. Willems
Backward saccades during reading have been hypothesized to be involved in structural reanalysis, or to be related to the level of text difficulty. We test the hypothesis that backward saccades are involved in online syntactic analysis. If this is the case we expect that saccades will coincide, at least partially, with the edges of the relations computed by a dependency parser. In order to test this, we analyzed a large eye-tracking dataset collected while 102 participants read three short narrative texts. Our results show a relation between backward saccades and the syntactic structure of sentences.
阅读过程中的向后扫视被假设与结构再分析有关,或者与文本难度水平有关。我们检验了向后扫视参与在线句法分析的假设。如果是这种情况,我们预计扫视将至少部分地与依赖项解析器计算的关系边缘重合。为了验证这一点,我们分析了102名参与者在阅读三篇短篇叙事文本时收集的大型眼球追踪数据集。我们的研究结果显示了向后扫视与句子的句法结构之间的关系。
{"title":"Dependency Parsing with your Eyes: Dependency Structure Predicts Eye Regressions During Reading","authors":"A. Lopopolo, S. Frank, Antal van den Bosch, Roel M. Willems","doi":"10.18653/v1/W19-2909","DOIUrl":"https://doi.org/10.18653/v1/W19-2909","url":null,"abstract":"Backward saccades during reading have been hypothesized to be involved in structural reanalysis, or to be related to the level of text difficulty. We test the hypothesis that backward saccades are involved in online syntactic analysis. If this is the case we expect that saccades will coincide, at least partially, with the edges of the relations computed by a dependency parser. In order to test this, we analyzed a large eye-tracking dataset collected while 102 participants read three short narrative texts. Our results show a relation between backward saccades and the syntactic structure of sentences.","PeriodicalId":428409,"journal":{"name":"Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-04-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125836314","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Estimating word co-occurrence probabilities from pretrained static embeddings using a log-bilinear model 使用对数双线性模型估计预训练静态嵌入的词共现概率
Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.cmcl-1.6
Richard Futrell
We investigate how to use pretrained static word embeddings to deliver improved estimates of bilexical co-occurrence probabilities: conditional probabilities of one word given a single other word in a specific relationship. Such probabilities play important roles in psycholinguistics, corpus linguistics, and usage-based cognitive modeling of language more generally. We propose a log-bilinear model taking pretrained vector representations of the two words as input, enabling generalization based on the distributional information contained in both vectors. We show that this model outperforms baselines in estimating probabilities of adjectives given nouns that they attributively modify, and probabilities of nominal direct objects given their head verbs, given limited training data in Arabic, English, Korean, and Spanish.
我们研究了如何使用预训练的静态词嵌入来提供双字共现概率的改进估计:一个词在特定关系中给定单个另一个词的条件概率。这种概率在心理语言学、语料库语言学和基于使用的语言认知模型中发挥着重要作用。我们提出了一个对数双线性模型,将两个词的预训练向量表示作为输入,基于两个向量中包含的分布信息实现泛化。我们表明,在给定阿拉伯语、英语、韩语和西班牙语有限的训练数据的情况下,该模型在估计给定属性修饰名词的形容词概率和给定头部动词的名义直接对象概率方面优于基线。
{"title":"Estimating word co-occurrence probabilities from pretrained static embeddings using a log-bilinear model","authors":"Richard Futrell","doi":"10.18653/v1/2022.cmcl-1.6","DOIUrl":"https://doi.org/10.18653/v1/2022.cmcl-1.6","url":null,"abstract":"We investigate how to use pretrained static word embeddings to deliver improved estimates of bilexical co-occurrence probabilities: conditional probabilities of one word given a single other word in a specific relationship. Such probabilities play important roles in psycholinguistics, corpus linguistics, and usage-based cognitive modeling of language more generally. We propose a log-bilinear model taking pretrained vector representations of the two words as input, enabling generalization based on the distributional information contained in both vectors. We show that this model outperforms baselines in estimating probabilities of adjectives given nouns that they attributively modify, and probabilities of nominal direct objects given their head verbs, given limited training data in Arabic, English, Korean, and Spanish.","PeriodicalId":428409,"journal":{"name":"Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124995304","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Team DMG at CMCL 2022 Shared Task: Transformer Adapters for the Multi- and Cross-Lingual Prediction of Human Reading Behavior 团队DMG在CMCL 2022共享任务:用于人类阅读行为的多语言和跨语言预测的变压器适配器
Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.cmcl-1.16
Ece Takmaz
In this paper, we present the details of our approaches that attained the second place in the shared task of the ACL 2022 Cognitive Modeling and Computational Linguistics Workshop. The shared task is focused on multi- and cross-lingual prediction of eye movement features in human reading behavior, which could provide valuable information regarding language processing. To this end, we train ‘adapters’ inserted into the layers of frozen transformer-based pretrained language models. We find that multilingual models equipped with adapters perform well in predicting eye-tracking features. Our results suggest that utilizing language- and task-specific adapters is beneficial and translating test sets into similar languages that exist in the training set could help with zero-shot transferability in the prediction of human reading behavior.
在本文中,我们介绍了我们的方法的细节,这些方法在ACL 2022认知建模和计算语言学研讨会的共享任务中获得了第二名。该共享任务的重点是对人类阅读行为的多语言和跨语言眼动特征进行预测,为语言处理提供有价值的信息。为此,我们训练“适配器”插入到基于冷冻变压器的预训练语言模型层中。我们发现配备适配器的多语言模型在预测眼动追踪特征方面表现良好。我们的研究结果表明,使用特定于语言和任务的适配器是有益的,并且将测试集翻译成训练集中存在的类似语言可以帮助实现零射击可转移性,从而预测人类阅读行为。
{"title":"Team DMG at CMCL 2022 Shared Task: Transformer Adapters for the Multi- and Cross-Lingual Prediction of Human Reading Behavior","authors":"Ece Takmaz","doi":"10.18653/v1/2022.cmcl-1.16","DOIUrl":"https://doi.org/10.18653/v1/2022.cmcl-1.16","url":null,"abstract":"In this paper, we present the details of our approaches that attained the second place in the shared task of the ACL 2022 Cognitive Modeling and Computational Linguistics Workshop. The shared task is focused on multi- and cross-lingual prediction of eye movement features in human reading behavior, which could provide valuable information regarding language processing. To this end, we train ‘adapters’ inserted into the layers of frozen transformer-based pretrained language models. We find that multilingual models equipped with adapters perform well in predicting eye-tracking features. Our results suggest that utilizing language- and task-specific adapters is beneficial and translating test sets into similar languages that exist in the training set could help with zero-shot transferability in the prediction of human reading behavior.","PeriodicalId":428409,"journal":{"name":"Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128734838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Modeling the Relationship between Input Distributions and Learning Trajectories with the Tolerance Principle 用容差原则建模输入分布与学习轨迹之间的关系
Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.cmcl-1.7
Jordan Kodner
Child language learners develop with remarkable uniformity, both in their learning trajectories and ultimate outcomes, despite major differences in their learning environments. In this paper, we explore the role that the frequencies and distributions of irregular lexical items in the input plays in driving learning trajectories. We conclude that while the Tolerance Principle, a type-based model of productivity learning, accounts for inter-learner uniformity, it also interacts with input distributions to drive cross-linguistic variation in learning trajectories.
儿童语言学习者的发展具有显著的一致性,无论是学习轨迹还是最终结果,尽管他们的学习环境存在重大差异。在本文中,我们探讨了输入中不规则词汇的频率和分布在驱动学习轨迹中的作用。我们得出的结论是,宽容原则是一种基于类型的生产力学习模型,虽然它解释了学习者之间的一致性,但它也与输入分布相互作用,驱动学习轨迹的跨语言差异。
{"title":"Modeling the Relationship between Input Distributions and Learning Trajectories with the Tolerance Principle","authors":"Jordan Kodner","doi":"10.18653/v1/2022.cmcl-1.7","DOIUrl":"https://doi.org/10.18653/v1/2022.cmcl-1.7","url":null,"abstract":"Child language learners develop with remarkable uniformity, both in their learning trajectories and ultimate outcomes, despite major differences in their learning environments. In this paper, we explore the role that the frequencies and distributions of irregular lexical items in the input plays in driving learning trajectories. We conclude that while the Tolerance Principle, a type-based model of productivity learning, accounts for inter-learner uniformity, it also interacts with input distributions to drive cross-linguistic variation in learning trajectories.","PeriodicalId":428409,"journal":{"name":"Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics","volume":"332 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121685587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Modeling Hierarchical Syntactic Structures in Morphological Processing 形态处理中的分层句法结构建模
Yohei Oseki, Charles D. Yang, A. Marantz
Sentences are represented as hierarchical syntactic structures, which have been successfully modeled in sentence processing. In contrast, despite the theoretical agreement on hierarchical syntactic structures within words, words have been argued to be computationally less complex than sentences and implemented by finite-state models as linear strings of morphemes, and even the psychological reality of morphemes has been denied. In this paper, extending the computational models employed in sentence processing to morphological processing, we performed a computational simulation experiment where, given incremental surprisal as a linking hypothesis, five computational models with different representational assumptions were evaluated against human reaction times in visual lexical decision experiments available from the English Lexicon Project (ELP), a “shared task” in the morphological processing literature. The simulation experiment demonstrated that (i) “amorphous” models without morpheme units underperformed relative to “morphous” models, (ii) a computational model with hierarchical syntactic structures, Probabilistic Context-Free Grammar (PCFG), most accurately explained human reaction times, and (iii) this performance was achieved on top of surface frequency effects. These results strongly suggest that morphological processing tracks morphemes incrementally from left to right and parses them into hierarchical syntactic structures, contrary to “amorphous” and finite-state models of morphological processing.
句子被表示为层次句法结构,这种句法结构已经成功地在句子处理中建模。相比之下,尽管在理论上对词的分层句法结构达成了一致,但人们一直认为词在计算上不如句子复杂,并且通过有限状态模型作为线性语素串来实现,甚至语素的心理现实也被否认了。在本文中,我们将句子处理中的计算模型扩展到形态处理中,进行了一个计算模拟实验,在此实验中,我们将增量惊讶作为一个关联假设,对来自英语词典项目(ELP)的视觉词汇决策实验中人类的反应时间进行了评估,ELP是形态学处理文献中的一个“共享任务”。模拟实验表明:(i)没有语素单位的“无定形”模型相对于“形态”模型表现不佳,(ii)具有分层句法结构的计算模型,概率上下文无关语法(PCFG)最准确地解释了人类的反应时间,以及(iii)这种性能是在表面频率效应的基础上实现的。这些结果强烈表明,形态学处理从左到右逐渐跟踪语素,并将其解析为分层句法结构,这与形态学处理的“无定形”和有限状态模型相反。
{"title":"Modeling Hierarchical Syntactic Structures in Morphological Processing","authors":"Yohei Oseki, Charles D. Yang, A. Marantz","doi":"10.18653/v1/W19-2905","DOIUrl":"https://doi.org/10.18653/v1/W19-2905","url":null,"abstract":"Sentences are represented as hierarchical syntactic structures, which have been successfully modeled in sentence processing. In contrast, despite the theoretical agreement on hierarchical syntactic structures within words, words have been argued to be computationally less complex than sentences and implemented by finite-state models as linear strings of morphemes, and even the psychological reality of morphemes has been denied. In this paper, extending the computational models employed in sentence processing to morphological processing, we performed a computational simulation experiment where, given incremental surprisal as a linking hypothesis, five computational models with different representational assumptions were evaluated against human reaction times in visual lexical decision experiments available from the English Lexicon Project (ELP), a “shared task” in the morphological processing literature. The simulation experiment demonstrated that (i) “amorphous” models without morpheme units underperformed relative to “morphous” models, (ii) a computational model with hierarchical syntactic structures, Probabilistic Context-Free Grammar (PCFG), most accurately explained human reaction times, and (iii) this performance was achieved on top of surface frequency effects. These results strongly suggest that morphological processing tracks morphemes incrementally from left to right and parses them into hierarchical syntactic structures, contrary to “amorphous” and finite-state models of morphological processing.","PeriodicalId":428409,"journal":{"name":"Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics","volume":"151 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128973051","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
CMCL 2022 Shared Task on Multilingual and Crosslingual Prediction of Human Reading Behavior CMCL 2022人类阅读行为的多语言和跨语言预测共享任务
Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.cmcl-1.14
Nora Hollenstein, Emmanuele Chersoni, Cassandra L. Jacobs, Yohei Oseki, Laurent Prévot, Enrico Santus
We present the second shared task on eye-tracking data prediction of the Cognitive Modeling and Computational Linguistics Workshop (CMCL). Differently from the previous edition, participating teams are asked to predict eye-tracking features from multiple languages, including a surprise language for which there were no available training data. Moreover, the task also included the prediction of standard deviations of feature values in order to account for individual differences between readers.A total of six teams registered to the task. For the first subtask on multilingual prediction, the winning team proposed a regression model based on lexical features, while for the second subtask on cross-lingual prediction, the winning team used a hybrid model based on a multilingual transformer embeddings as well as statistical features.
我们提出了认知建模与计算语言学研讨会(CMCL)关于眼动追踪数据预测的第二个共享任务。与上一版不同的是,参赛团队被要求预测多种语言的眼球追踪特征,包括一种没有可用训练数据的意外语言。此外,该任务还包括预测特征值的标准差,以解释读者之间的个体差异。共有六个队报名参加了这项任务。对于多语言预测的第一个子任务,获胜团队提出了基于词汇特征的回归模型,而对于跨语言预测的第二个子任务,获胜团队使用了基于多语言转换嵌入和统计特征的混合模型。
{"title":"CMCL 2022 Shared Task on Multilingual and Crosslingual Prediction of Human Reading Behavior","authors":"Nora Hollenstein, Emmanuele Chersoni, Cassandra L. Jacobs, Yohei Oseki, Laurent Prévot, Enrico Santus","doi":"10.18653/v1/2022.cmcl-1.14","DOIUrl":"https://doi.org/10.18653/v1/2022.cmcl-1.14","url":null,"abstract":"We present the second shared task on eye-tracking data prediction of the Cognitive Modeling and Computational Linguistics Workshop (CMCL). Differently from the previous edition, participating teams are asked to predict eye-tracking features from multiple languages, including a surprise language for which there were no available training data. Moreover, the task also included the prediction of standard deviations of feature values in order to account for individual differences between readers.A total of six teams registered to the task. For the first subtask on multilingual prediction, the winning team proposed a regression model based on lexical features, while for the second subtask on cross-lingual prediction, the winning team used a hybrid model based on a multilingual transformer embeddings as well as statistical features.","PeriodicalId":428409,"journal":{"name":"Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129344125","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Toward a Computational Multidimensional Lexical Similarity Measure for Modeling Word Association Tasks in Psycholinguistics 面向心理语言学词关联任务建模的计算多维词汇相似度量
Bruno Gaume, L. Ho-Dac, Ludovic Tanguy, Cécile Fabre, Bénédicte Pierrejean, Nabil Hathout, Jérôme Farinas, J. Pinquier, Lola Danet, P. Péran, X. D. Boissezon, M. Jucla
This paper presents the first results of a multidisciplinary project, the “Evolex” project, gathering researchers in Psycholinguistics, Neuropsychology, Computer Science, Natural Language Processing and Linguistics. The Evolex project aims at proposing a new data-based inductive method for automatically characterising the relation between pairs of french words collected in psycholinguistics experiments on lexical access. This method takes advantage of several complementary computational measures of semantic similarity. We show that some measures are more correlated than others with the frequency of lexical associations, and that they also differ in the way they capture different semantic relations. This allows us to consider building a multidimensional lexical similarity to automate the classification of lexical associations.
这篇论文介绍了一个多学科项目的第一批成果,“Evolex”项目,汇集了心理语言学、神经心理学、计算机科学、自然语言处理和语言学的研究人员。Evolex项目旨在提出一种新的基于数据的归纳方法,用于自动描述在词汇获取的心理语言学实验中收集的法语单词对之间的关系。该方法利用了几种互补的语义相似度计算度量。我们表明,一些测量与词汇关联的频率比其他测量更相关,并且它们在捕获不同语义关系的方式上也有所不同。这允许我们考虑构建多维词汇相似度来自动分类词汇关联。
{"title":"Toward a Computational Multidimensional Lexical Similarity Measure for Modeling Word Association Tasks in Psycholinguistics","authors":"Bruno Gaume, L. Ho-Dac, Ludovic Tanguy, Cécile Fabre, Bénédicte Pierrejean, Nabil Hathout, Jérôme Farinas, J. Pinquier, Lola Danet, P. Péran, X. D. Boissezon, M. Jucla","doi":"10.18653/v1/W19-2908","DOIUrl":"https://doi.org/10.18653/v1/W19-2908","url":null,"abstract":"This paper presents the first results of a multidisciplinary project, the “Evolex” project, gathering researchers in Psycholinguistics, Neuropsychology, Computer Science, Natural Language Processing and Linguistics. The Evolex project aims at proposing a new data-based inductive method for automatically characterising the relation between pairs of french words collected in psycholinguistics experiments on lexical access. This method takes advantage of several complementary computational measures of semantic similarity. We show that some measures are more correlated than others with the frequency of lexical associations, and that they also differ in the way they capture different semantic relations. This allows us to consider building a multidimensional lexical similarity to automate the classification of lexical associations.","PeriodicalId":428409,"journal":{"name":"Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133466602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Neural Model for Compositional Word Embeddings and Sentence Processing 合成词嵌入与句子处理的神经模型
Pub Date : 1900-01-01 DOI: 10.18653/v1/2022.cmcl-1.2
Shalom Lappin, Jean-Philippe Bernardy
We propose a new neural model for word embeddings, which uses Unitary Matrices as the primary device for encoding lexical information. It uses simple matrix multiplication to derive matrices for large units, yielding a sentence processing model that is strictly compositional, does not lose information over time steps, and is transparent, in the sense that word embeddings can be analysed regardless of context. This model does not employ activation functions, and so the network is fully accessible to analysis by the methods of linear algebra at each point in its operation on an input sequence. We test it in two NLP agreement tasks and obtain rule like perfect accuracy, with greater stability than current state-of-the-art systems. Our proposed model goes some way towards offering a class of computationally powerful deep learning systems that can be fully understood and compared to human cognitive processes for natural language learning and representation.
本文提出了一种新的词嵌入神经网络模型,该模型使用酉矩阵作为编码词汇信息的主要设备。它使用简单的矩阵乘法来导出大单位的矩阵,从而产生一个严格组合的句子处理模型,不会随着时间的推移丢失信息,并且是透明的,从某种意义上说,无论上下文如何,都可以分析词嵌入。该模型不使用激活函数,因此网络在其对输入序列的操作中的每个点都可以通过线性代数方法进行分析。我们在两个NLP协议任务中测试了它,并获得了类似规则的完美精度,比目前最先进的系统具有更大的稳定性。我们提出的模型在某种程度上提供了一类计算能力强大的深度学习系统,这些系统可以完全理解,并与自然语言学习和表示的人类认知过程进行比较。
{"title":"A Neural Model for Compositional Word Embeddings and Sentence Processing","authors":"Shalom Lappin, Jean-Philippe Bernardy","doi":"10.18653/v1/2022.cmcl-1.2","DOIUrl":"https://doi.org/10.18653/v1/2022.cmcl-1.2","url":null,"abstract":"We propose a new neural model for word embeddings, which uses Unitary Matrices as the primary device for encoding lexical information. It uses simple matrix multiplication to derive matrices for large units, yielding a sentence processing model that is strictly compositional, does not lose information over time steps, and is transparent, in the sense that word embeddings can be analysed regardless of context. This model does not employ activation functions, and so the network is fully accessible to analysis by the methods of linear algebra at each point in its operation on an input sequence. We test it in two NLP agreement tasks and obtain rule like perfect accuracy, with greater stability than current state-of-the-art systems. Our proposed model goes some way towards offering a class of computationally powerful deep learning systems that can be fully understood and compared to human cognitive processes for natural language learning and representation.","PeriodicalId":428409,"journal":{"name":"Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128365912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1