首页 > 最新文献

Proceedings of the conference. Association for Computational Linguistics. Meeting最新文献

英文 中文
Predicting pragmatic discourse features in the language of adults with autism spectrum disorder. 自闭症谱系障碍成人语用语篇特征预测。
Pub Date : 2021-08-01 DOI: 10.18653/v1/2021.acl-srw.29
Christine Yang, Duanchen Liu, Qingyun Yang, Zoey Liu, Emily Prud'hommeaux

Individuals with autism spectrum disorder (ASD) experience difficulties in social aspects of communication, but the linguistic characteristics associated with deficits in discourse and pragmatic expression are often difficult to precisely identify and quantify. We are currently collecting a corpus of transcribed natural conversations produced in an experimental setting in which participants with and without ASD complete a number of collaborative tasks with their neurotypical peers. Using this dyadic conversational data, we investigate three pragmatic features - politeness, uncertainty, and informativeness - and present a dataset of utterances annotated for each of these features on a three-point scale. We then introduce ongoing work in developing and training neural models to automatically predict these features, with the goal of identifying the same between-groups differences that are observed using manual annotations. We find the best performing model for all three features is a feedforward neural network trained with BERT embeddings. Our models yield higher accuracy than ones used in previous approaches for deriving these features, with F1 exceeding 0.82 for all three pragmatic features.

自闭症谱系障碍(ASD)患者在社交方面存在沟通困难,但与话语和语用表达缺陷相关的语言特征往往难以精确识别和量化。我们目前正在收集在实验环境中产生的转录自然对话的语料库,在实验环境中,有和没有ASD的参与者与他们的神经正常的同伴完成了许多合作任务。使用这种二元会话数据,我们研究了三个语用特征——礼貌、不确定性和信息性——并给出了一个三分制的话语数据集,对这些特征中的每一个都进行了注释。然后,我们介绍了正在进行的开发和训练神经模型的工作,以自动预测这些特征,目标是识别使用手动注释观察到的相同组间差异。我们发现这三个特征的最佳表现模型是用BERT嵌入训练的前馈神经网络。我们的模型比以前用于导出这些特征的方法产生更高的精度,所有三个实用特征的F1都超过0.82。
{"title":"Predicting pragmatic discourse features in the language of adults with autism spectrum disorder.","authors":"Christine Yang,&nbsp;Duanchen Liu,&nbsp;Qingyun Yang,&nbsp;Zoey Liu,&nbsp;Emily Prud'hommeaux","doi":"10.18653/v1/2021.acl-srw.29","DOIUrl":"https://doi.org/10.18653/v1/2021.acl-srw.29","url":null,"abstract":"<p><p>Individuals with autism spectrum disorder (ASD) experience difficulties in social aspects of communication, but the linguistic characteristics associated with deficits in discourse and pragmatic expression are often difficult to precisely identify and quantify. We are currently collecting a corpus of transcribed natural conversations produced in an experimental setting in which participants with and without ASD complete a number of collaborative tasks with their neurotypical peers. Using this dyadic conversational data, we investigate three pragmatic features - politeness, uncertainty, and informativeness - and present a dataset of utterances annotated for each of these features on a three-point scale. We then introduce ongoing work in developing and training neural models to automatically predict these features, with the goal of identifying the same between-groups differences that are observed using manual annotations. We find the best performing model for all three features is a feedforward neural network trained with BERT embeddings. Our models yield higher accuracy than ones used in previous approaches for deriving these features, with F1 exceeding 0.82 for all three pragmatic features.</p>","PeriodicalId":74541,"journal":{"name":"Proceedings of the conference. Association for Computational Linguistics. Meeting","volume":"2021 ","pages":"284-291"},"PeriodicalIF":0.0,"publicationDate":"2021-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9633181/pdf/nihms-1846176.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"40669995","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Robust Knowledge Graph Completion with Stacked Convolutions and a Student Re-Ranking Network. 基于堆叠卷积的鲁棒知识图谱补全与学生重排序网络。
Pub Date : 2021-08-01 DOI: 10.18653/v1/2021.acl-long.82
Justin Lovelace, Denis Newman-Griffis, Shikhar Vashishth, Jill Fain Lehman, Carolyn Penstein Rosé

Knowledge Graph (KG) completion research usually focuses on densely connected benchmark datasets that are not representative of real KGs. We curate two KG datasets that include biomedical and encyclopedic knowledge and use an existing commonsense KG dataset to explore KG completion in the more realistic setting where dense connectivity is not guaranteed. We develop a deep convolutional network that utilizes textual entity representations and demonstrate that our model outperforms recent KG completion methods in this challenging setting. We find that our model's performance improvements stem primarily from its robustness to sparsity. We then distill the knowledge from the convolutional network into a student network that re-ranks promising candidate entities. This re-ranking stage leads to further improvements in performance and demonstrates the effectiveness of entity re-ranking for KG completion.

知识图谱(Knowledge Graph, KG)完井研究通常侧重于密集连接的基准数据集,这些数据集不能代表真实的知识图谱。我们整理了两个KG数据集,包括生物医学和百科知识,并使用现有的常识KG数据集来探索更现实的、不能保证密集连接的知识图谱完井。我们开发了一个利用文本实体表示的深度卷积网络,并证明我们的模型在这个具有挑战性的环境中优于最近的KG补全方法。我们发现我们的模型的性能改进主要源于它对稀疏性的鲁棒性。然后,我们将卷积网络中的知识提取到一个学生网络中,该网络对有希望的候选实体进行重新排序。这个重新排序阶段可以进一步提高性能,并证明了KG完井实体重新排序的有效性。
{"title":"Robust Knowledge Graph Completion with Stacked Convolutions and a Student Re-Ranking Network.","authors":"Justin Lovelace,&nbsp;Denis Newman-Griffis,&nbsp;Shikhar Vashishth,&nbsp;Jill Fain Lehman,&nbsp;Carolyn Penstein Rosé","doi":"10.18653/v1/2021.acl-long.82","DOIUrl":"https://doi.org/10.18653/v1/2021.acl-long.82","url":null,"abstract":"<p><p>Knowledge Graph (KG) completion research usually focuses on densely connected benchmark datasets that are not representative of real KGs. We curate two KG datasets that include biomedical and encyclopedic knowledge and use an existing commonsense KG dataset to explore KG completion in the more realistic setting where dense connectivity is not guaranteed. We develop a deep convolutional network that utilizes textual entity representations and demonstrate that our model outperforms recent KG completion methods in this challenging setting. We find that our model's performance improvements stem primarily from its robustness to sparsity. We then distill the knowledge from the convolutional network into a student network that re-ranks promising candidate entities. This re-ranking stage leads to further improvements in performance and demonstrates the effectiveness of entity re-ranking for KG completion.</p>","PeriodicalId":74541,"journal":{"name":"Proceedings of the conference. Association for Computational Linguistics. Meeting","volume":"2021 ","pages":"1016-1029"},"PeriodicalIF":0.0,"publicationDate":"2021-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9272461/pdf/nihms-1810978.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"40516342","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Topic-Based Measures of Conversation for Detecting Mild Cognitive Impairment. 基于话题的谈话方法检测轻度认知障碍。
Liu Chen, Hiroko H Dodge, Meysam Asgari

Conversation is a complex cognitive task that engages multiple aspects of cognitive functions to remember the discussed topics, monitor the semantic and linguistic elements, and recognize others' emotions. In this paper, we propose a computational method based on the lexical coherence of consecutive utterances to quantify topical variations in semi-structured conversations of older adults with cognitive impairments. Extracting the lexical knowledge of conversational utterances, our method generates a set of novel conversational measures that indicate underlying cognitive deficits among subjects with mild cognitive impairment (MCI). Our preliminary results verify the utility of the proposed conversation-based measures in distinguishing MCI from healthy controls.

对话是一项复杂的认知任务,涉及多个方面的认知功能,包括记住所讨论的话题,监控语义和语言元素,以及识别他人的情绪。在本文中,我们提出了一种基于连续话语的词汇连贯性的计算方法来量化认知障碍老年人半结构化对话中的话题变化。通过提取会话话语的词汇知识,我们的方法生成了一套新的会话测量方法,这些测量方法可以显示轻度认知障碍(MCI)受试者潜在的认知缺陷。我们的初步结果验证了所提出的基于对话的措施在区分MCI和健康对照方面的效用。
{"title":"Topic-Based Measures of Conversation for Detecting Mild Cognitive Impairment.","authors":"Liu Chen,&nbsp;Hiroko H Dodge,&nbsp;Meysam Asgari","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Conversation is a complex cognitive task that engages multiple aspects of cognitive functions to remember the discussed topics, monitor the semantic and linguistic elements, and recognize others' emotions. In this paper, we propose a computational method based on the lexical coherence of consecutive utterances to quantify topical variations in semi-structured conversations of older adults with cognitive impairments. Extracting the lexical knowledge of conversational utterances, our method generates a set of novel conversational measures that indicate underlying cognitive deficits among subjects with mild cognitive impairment (MCI). Our preliminary results verify the utility of the proposed conversation-based measures in distinguishing MCI from healthy controls.</p>","PeriodicalId":74541,"journal":{"name":"Proceedings of the conference. Association for Computational Linguistics. Meeting","volume":" ","pages":"63-67"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7909094/pdf/nihms-1670817.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25414585","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Benchmark and Best Practices for Biomedical Knowledge Graph Embeddings. 生物医学知识图嵌入的基准和最佳实践。
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.bionlp-1.18
David Chang, Ivana Balažević, Carl Allen, Daniel Chawla, Cynthia Brandt, Richard Andrew Taylor

Much of biomedical and healthcare data is encoded in discrete, symbolic form such as text and medical codes. There is a wealth of expert-curated biomedical domain knowledge stored in knowledge bases and ontologies, but the lack of reliable methods for learning knowledge representation has limited their usefulness in machine learning applications. While text-based representation learning has significantly improved in recent years through advances in natural language processing, attempts to learn biomedical concept embeddings so far have been lacking. A recent family of models called knowledge graph embeddings have shown promising results on general domain knowledge graphs, and we explore their capabilities in the biomedical domain. We train several state-of-the-art knowledge graph embedding models on the SNOMED-CT knowledge graph, provide a benchmark with comparison to existing methods and in-depth discussion on best practices, and make a case for the importance of leveraging the multi-relational nature of knowledge graphs for learning biomedical knowledge representation. The embeddings, code, and materials will be made available to the community.

许多生物医学和医疗保健数据以离散的符号形式编码,如文本和医疗代码。在知识库和本体中存储着丰富的专家管理的生物医学领域知识,但是缺乏可靠的学习知识表示方法限制了它们在机器学习应用中的实用性。近年来,随着自然语言处理的进步,基于文本的表示学习有了显著的改善,但迄今为止,学习生物医学概念嵌入的尝试还很缺乏。最近一组称为知识图嵌入的模型在一般领域知识图上显示了有希望的结果,我们探索了它们在生物医学领域的能力。我们在SNOMED-CT知识图上训练了几个最先进的知识图嵌入模型,提供了与现有方法比较的基准和对最佳实践的深入讨论,并说明了利用知识图的多关系性质学习生物医学知识表示的重要性。嵌入、代码和材料将提供给社区。
{"title":"Benchmark and Best Practices for Biomedical Knowledge Graph Embeddings.","authors":"David Chang,&nbsp;Ivana Balažević,&nbsp;Carl Allen,&nbsp;Daniel Chawla,&nbsp;Cynthia Brandt,&nbsp;Richard Andrew Taylor","doi":"10.18653/v1/2020.bionlp-1.18","DOIUrl":"https://doi.org/10.18653/v1/2020.bionlp-1.18","url":null,"abstract":"<p><p>Much of biomedical and healthcare data is encoded in discrete, symbolic form such as text and medical codes. There is a wealth of expert-curated biomedical domain knowledge stored in knowledge bases and ontologies, but the lack of reliable methods for learning knowledge representation has limited their usefulness in machine learning applications. While text-based representation learning has significantly improved in recent years through advances in natural language processing, attempts to learn biomedical concept embeddings so far have been lacking. A recent family of models called knowledge graph embeddings have shown promising results on general domain knowledge graphs, and we explore their capabilities in the biomedical domain. We train several state-of-the-art knowledge graph embedding models on the SNOMED-CT knowledge graph, provide a benchmark with comparison to existing methods and in-depth discussion on best practices, and make a case for the importance of leveraging the multi-relational nature of knowledge graphs for learning biomedical knowledge representation. The embeddings, code, and materials will be made available to the community.</p>","PeriodicalId":74541,"journal":{"name":"Proceedings of the conference. Association for Computational Linguistics. Meeting","volume":" ","pages":"167-176"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7971091/pdf/nihms-1676481.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25511788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Towards End-2-end Learning for Predicting Behavior Codes from Spoken Utterances in Psychotherapy Conversations. 从心理治疗对话口语中预测行为代码的终结-2 端学习。
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.acl-main.351
Karan Singla, Zhuohao Chen, David C Atkins, Shrikanth Narayanan

Spoken language understanding tasks usually rely on pipelines involving complex processing blocks such as voice activity detection, speaker diarization and Automatic speech recognition (ASR). We propose a novel framework for predicting utterance level labels directly from speech features, thus removing the dependency on first generating transcripts, and transcription free behavioral coding. Our classifier uses a pretrained Speech-2-Vector encoder as bottleneck to generate word-level representations from speech features. This pre-trained encoder learns to encode speech features for a word using an objective similar to Word2Vec. Our proposed approach just uses speech features and word segmentation information for predicting spoken utterance-level target labels. We show that our model achieves competitive results to other state-of-the-art approaches which use transcribed text for the task of predicting psychotherapy-relevant behavior codes.

口语理解任务通常依赖于涉及复杂处理模块的流水线,如语音活动检测、说话者日记化和自动语音识别(ASR)。我们提出了一个新颖的框架,可直接从语音特征预测语句级标签,从而消除了对首次生成转录和无转录行为编码的依赖。我们的分类器使用预训练的 Speech-2-Vector 编码器作为瓶颈,从语音特征生成词级表示。这种预先训练好的编码器通过类似于 Word2Vec 的目标来学习对单词的语音特征进行编码。我们提出的方法仅使用语音特征和单词分段信息来预测口语语段级目标标签。我们的研究表明,我们的模型与其他使用转录文本预测心理治疗相关行为代码的先进方法相比,取得了具有竞争力的结果。
{"title":"Towards End-2-end Learning for Predicting Behavior Codes from Spoken Utterances in Psychotherapy Conversations.","authors":"Karan Singla, Zhuohao Chen, David C Atkins, Shrikanth Narayanan","doi":"10.18653/v1/2020.acl-main.351","DOIUrl":"10.18653/v1/2020.acl-main.351","url":null,"abstract":"<p><p>Spoken language understanding tasks usually rely on pipelines involving complex processing blocks such as voice activity detection, speaker diarization and Automatic speech recognition (ASR). We propose a novel framework for predicting utterance level labels directly from speech features, thus removing the dependency on first generating transcripts, and transcription free behavioral coding. Our classifier uses a pretrained Speech-2-Vector encoder as bottleneck to generate word-level representations from speech features. This pre-trained encoder learns to encode speech features for a word using an objective similar to Word2Vec. Our proposed approach just uses speech features and word segmentation information for predicting spoken utterance-level target labels. We show that our model achieves competitive results to other state-of-the-art approaches which use transcribed text for the task of predicting psychotherapy-relevant behavior codes.</p>","PeriodicalId":74541,"journal":{"name":"Proceedings of the conference. Association for Computational Linguistics. Meeting","volume":"2020 ","pages":"3797-3803"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9901279/pdf/nihms-1858361.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9229146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards an ontology-based medication conversational agent for PrEP and PEP. 面向PrEP和PEP的基于本体的药物会话代理。
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.nlpmc-1.5
Muhammad Tuan Amith, Licong Cui, Kirk Roberts, Cui Tao

HIV (human immunodeficiency virus) can damage a human's immune system and cause Acquired Immunodeficiency Syndrome (AIDS) which could lead to severe outcomes, including death. While HIV infections have decreased over the last decade, there is still a significant population where the infection permeates. PrEP and PEP are two proven preventive measures introduced that involve periodic dosage to stop the onset of HIV infection. However, the adherence rates for this medication is low in part due to the lack of information about the medication. There exist several communication barriers that prevent patient-provider communication from happening. In this work, we present our ontology-based method for automating the communication of this medication that can be deployed for live conversational agents for PrEP and PEP. This method facilitates a model of automated conversation between the machine and user can also answer relevant questions.

HIV(人类免疫缺陷病毒)可以破坏人类的免疫系统,导致获得性免疫缺陷综合症(艾滋病),这可能导致严重的后果,包括死亡。虽然艾滋病毒感染在过去十年中有所下降,但仍有大量人口感染。PrEP和PEP是两种经过验证的预防措施,涉及定期给药以阻止艾滋病毒感染的发生。然而,这种药物的依从率很低,部分原因是缺乏有关药物的信息。存在一些阻碍医患沟通的沟通障碍。在这项工作中,我们提出了基于本体的方法来自动化这种药物的通信,该方法可以部署为PrEP和PEP的实时会话代理。这种方法促进了机器和用户之间的自动对话模型,也可以回答相关问题。
{"title":"Towards an ontology-based medication conversational agent for PrEP and PEP.","authors":"Muhammad Tuan Amith,&nbsp;Licong Cui,&nbsp;Kirk Roberts,&nbsp;Cui Tao","doi":"10.18653/v1/2020.nlpmc-1.5","DOIUrl":"https://doi.org/10.18653/v1/2020.nlpmc-1.5","url":null,"abstract":"<p><p>HIV (human immunodeficiency virus) can damage a human's immune system and cause Acquired Immunodeficiency Syndrome (AIDS) which could lead to severe outcomes, including death. While HIV infections have decreased over the last decade, there is still a significant population where the infection permeates. PrEP and PEP are two proven preventive measures introduced that involve periodic dosage to stop the onset of HIV infection. However, the adherence rates for this medication is low in part due to the lack of information about the medication. There exist several communication barriers that prevent patient-provider communication from happening. In this work, we present our ontology-based method for automating the communication of this medication that can be deployed for live conversational agents for PrEP and PEP. This method facilitates a model of automated conversation between the machine and user can also answer relevant questions.</p>","PeriodicalId":74541,"journal":{"name":"Proceedings of the conference. Association for Computational Linguistics. Meeting","volume":" ","pages":"31-40"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7680642/pdf/nihms-1642495.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38636472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Calibrating Structured Output Predictors for Natural Language Processing. 校准用于自然语言处理的结构化输出预测器。
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.acl-main.188
Abhyuday Jagannatha, Hong Yu

We address the problem of calibrating prediction confidence for output entities of interest in natural language processing (NLP) applications. It is important that NLP applications such as named entity recognition and question answering produce calibrated confidence scores for their predictions, especially if the applications are to be deployed in a safety-critical domain such as healthcare. However, the output space of such structured prediction models is often too large to adapt binary or multi-class calibration methods directly. In this study, we propose a general calibration scheme for output entities of interest in neural network based structured prediction models. Our proposed method can be used with any binary class calibration scheme and a neural network model. Additionally, we show that our calibration method can also be used as an uncertainty-aware, entity-specific decoding step to improve the performance of the underlying model at no additional training cost or data requirements. We show that our method outperforms current calibration techniques for named-entity-recognition, part-of-speech and question answering. We also improve our model's performance from our decoding step across several tasks and benchmark datasets. Our method improves the calibration and model performance on out-ofdomain test scenarios as well.

我们解决了校准自然语言处理(NLP)应用中感兴趣的输出实体的预测置信度的问题。重要的是,NLP应用程序(如命名实体识别和问答)为其预测生成校准的置信度分数,特别是如果应用程序要部署在安全关键领域(如医疗保健)中。然而,这种结构化预测模型的输出空间往往太大,无法直接适应二值或多类校准方法。在这项研究中,我们提出了一种基于神经网络的结构化预测模型中感兴趣的输出实体的通用校准方案。该方法可用于任何二值类标定方案和神经网络模型。此外,我们表明,我们的校准方法也可以用作不确定性感知,实体特定的解码步骤,以提高底层模型的性能,而不需要额外的训练成本或数据需求。我们表明,我们的方法优于当前的命名实体识别、词性和问答校准技术。我们还通过跨多个任务和基准数据集的解码步骤提高了模型的性能。该方法还提高了域外测试场景下的标定和模型性能。
{"title":"Calibrating Structured Output Predictors for Natural Language Processing.","authors":"Abhyuday Jagannatha,&nbsp;Hong Yu","doi":"10.18653/v1/2020.acl-main.188","DOIUrl":"https://doi.org/10.18653/v1/2020.acl-main.188","url":null,"abstract":"<p><p>We address the problem of calibrating prediction confidence for output entities of interest in natural language processing (NLP) applications. It is important that NLP applications such as named entity recognition and question answering produce calibrated confidence scores for their predictions, especially if the applications are to be deployed in a safety-critical domain such as healthcare. However, the output space of such structured prediction models is often too large to adapt binary or multi-class calibration methods directly. In this study, we propose a general calibration scheme for output entities of interest in neural network based structured prediction models. Our proposed method can be used with any binary class calibration scheme and a neural network model. Additionally, we show that our calibration method can also be used as an uncertainty-aware, entity-specific decoding step to improve the performance of the underlying model at no additional training cost or data requirements. We show that our method outperforms current calibration techniques for named-entity-recognition, part-of-speech and question answering. We also improve our model's performance from our decoding step across several tasks and benchmark datasets. Our method improves the calibration and model performance on out-ofdomain test scenarios as well.</p>","PeriodicalId":74541,"journal":{"name":"Proceedings of the conference. Association for Computational Linguistics. Meeting","volume":" ","pages":"2078-2092"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7890517/pdf/nihms-1661932.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25390283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Automated Scoring of Clinical Expressive Language Evaluation Tasks. 临床表达性语言评估任务的自动评分。
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.bea-1.18
Yiyi Wang, Emily Prud'hommeaux, Meysam Asgari, Jill Dolata

Many clinical assessment instruments used to diagnose language impairments in children include a task in which the subject must formulate a sentence to describe an image using a specific target word. Because producing sentences in this way requires the speaker to integrate syntactic and semantic knowledge in a complex manner, responses are typically evaluated on several different dimensions of appropriateness yielding a single composite score for each response. In this paper, we present a dataset consisting of non-clinically elicited responses for three related sentence formulation tasks, and we propose an approach for automatically evaluating their appropriateness. Using neural machine translation, we generate correct-incorrect sentence pairs to serve as synthetic data in order to increase the amount and diversity of training data for our scoring model. Our scoring model uses transfer learning to facilitate automatic sentence appropriateness evaluation. We further compare custom word embeddings with pre-trained contextualized embeddings serving as features for our scoring model. We find that transfer learning improves scoring accuracy, particularly when using pre-trained contextualized embeddings.

许多用于诊断儿童语言障碍的临床评估工具都包含一项任务,即受试者必须用特定的目标词造句来描述一幅图像。由于以这种方式造句需要说话者以复杂的方式整合句法和语义知识,因此通常会从几个不同的适当性维度对回答进行评估,从而为每个回答得出一个综合分数。在本文中,我们介绍了一个数据集,该数据集由三个相关造句任务的非临床诱导回答组成,我们还提出了一种自动评估其适当性的方法。通过使用神经机器翻译,我们生成了正确-不正确句子对作为合成数据,以增加评分模型训练数据的数量和多样性。我们的评分模型使用迁移学习来促进句子适当性的自动评估。我们进一步比较了自定义词嵌入和作为评分模型特征的预训练上下文嵌入。我们发现,迁移学习提高了评分的准确性,尤其是在使用预先训练好的上下文嵌入式时。
{"title":"Automated Scoring of Clinical Expressive Language Evaluation Tasks.","authors":"Yiyi Wang, Emily Prud'hommeaux, Meysam Asgari, Jill Dolata","doi":"10.18653/v1/2020.bea-1.18","DOIUrl":"10.18653/v1/2020.bea-1.18","url":null,"abstract":"<p><p>Many clinical assessment instruments used to diagnose language impairments in children include a task in which the subject must formulate a sentence to describe an image using a specific target word. Because producing sentences in this way requires the speaker to integrate syntactic and semantic knowledge in a complex manner, responses are typically evaluated on several different dimensions of appropriateness yielding a single composite score for each response. In this paper, we present a dataset consisting of non-clinically elicited responses for three related sentence formulation tasks, and we propose an approach for automatically evaluating their appropriateness. Using neural machine translation, we generate correct-incorrect sentence pairs to serve as synthetic data in order to increase the amount and diversity of training data for our scoring model. Our scoring model uses transfer learning to facilitate automatic sentence appropriateness evaluation. We further compare custom word embeddings with pre-trained contextualized embeddings serving as features for our scoring model. We find that transfer learning improves scoring accuracy, particularly when using pre-trained contextualized embeddings.</p>","PeriodicalId":74541,"journal":{"name":"Proceedings of the conference. Association for Computational Linguistics. Meeting","volume":" ","pages":"177-185"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7556318/pdf/nihms-1636235.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38497581","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integrating Multimodal Information in Large Pretrained Transformers. 大型预训练变压器的多模态信息集成。
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.acl-main.214
Wasifur Rahman, Md Kamrul Hasan, Sangwu Lee, Amir Zadeh, Chengfeng Mao, Louis-Philippe Morency, Ehsan Hoque

Recent Transformer-based contextual word representations, including BERT and XLNet, have shown state-of-the-art performance in multiple disciplines within NLP. Fine-tuning the trained contextual models on task-specific datasets has been the key to achieving superior performance downstream. While fine-tuning these pre-trained models is straight-forward for lexical applications (applications with only language modality), it is not trivial for multimodal language (a growing area in NLP focused on modeling face-to-face communication). Pre-trained models don't have the necessary components to accept two extra modalities of vision and acoustic. In this paper, we proposed an attachment to BERT and XLNet called Multimodal Adaptation Gate (MAG). MAG allows BERT and XLNet to accept multimodal nonverbal data during fine-tuning. It does so by generating a shift to internal representation of BERT and XLNet; a shift that is conditioned on the visual and acoustic modalities. In our experiments, we study the commonly used CMU-MOSI and CMU-MOSEI datasets for multimodal sentiment analysis. Fine-tuning MAG-BERT and MAG-XLNet significantly boosts the sentiment analysis performance over previous baselines as well as language-only fine-tuning of BERT and XLNet. On the CMU-MOSI dataset, MAG-XLNet achieves human-level multimodal sentiment analysis performance for the first time in the NLP community.

最近基于transformer的上下文词表示,包括BERT和XLNet,已经在NLP的多个学科中显示了最先进的性能。在任务特定数据集上对训练好的上下文模型进行微调是在下游实现卓越性能的关键。虽然对这些预训练的模型进行微调对于词汇应用程序(只有语言模态的应用程序)来说是很简单的,但对于多模态语言(NLP中一个不断发展的领域,专注于建模面对面的交流)来说,这并不容易。预先训练的模型没有必要的组件来接受视觉和听觉两种额外的模式。在本文中,我们提出了BERT和XLNet的附件,称为多模态自适应门(MAG)。MAG允许BERT和XLNet在微调期间接受多模态非语言数据。它通过生成BERT和XLNet的内部表示的转换来实现这一点;以视觉和听觉模式为条件的转变。在我们的实验中,我们研究了常用的CMU-MOSI和CMU-MOSEI数据集用于多模态情感分析。与之前的基线相比,对magg -BERT和magg -XLNet进行微调可以显著提高情感分析的性能,也可以对BERT和XLNet进行仅语言的微调。在CMU-MOSI数据集上,MAG-XLNet在NLP社区首次实现了人类水平的多模态情感分析性能。
{"title":"Integrating Multimodal Information in Large Pretrained Transformers.","authors":"Wasifur Rahman,&nbsp;Md Kamrul Hasan,&nbsp;Sangwu Lee,&nbsp;Amir Zadeh,&nbsp;Chengfeng Mao,&nbsp;Louis-Philippe Morency,&nbsp;Ehsan Hoque","doi":"10.18653/v1/2020.acl-main.214","DOIUrl":"https://doi.org/10.18653/v1/2020.acl-main.214","url":null,"abstract":"<p><p>Recent Transformer-based contextual word representations, including BERT and XLNet, have shown state-of-the-art performance in multiple disciplines within NLP. Fine-tuning the trained contextual models on task-specific datasets has been the key to achieving superior performance downstream. While fine-tuning these pre-trained models is straight-forward for lexical applications (applications with only language modality), it is not trivial for multimodal language (a growing area in NLP focused on modeling face-to-face communication). Pre-trained models don't have the necessary components to accept two extra modalities of vision and acoustic. In this paper, we proposed an attachment to BERT and XLNet called Multimodal Adaptation Gate (MAG). MAG allows BERT and XLNet to accept multimodal nonverbal data during fine-tuning. It does so by generating a shift to internal representation of BERT and XLNet; a shift that is conditioned on the visual and acoustic modalities. In our experiments, we study the commonly used CMU-MOSI and CMU-MOSEI datasets for multimodal sentiment analysis. Fine-tuning MAG-BERT and MAG-XLNet significantly boosts the sentiment analysis performance over previous baselines as well as language-only fine-tuning of BERT and XLNet. On the CMU-MOSI dataset, MAG-XLNet achieves human-level multimodal sentiment analysis performance for the first time in the NLP community.</p>","PeriodicalId":74541,"journal":{"name":"Proceedings of the conference. Association for Computational Linguistics. Meeting","volume":" ","pages":"2359-2369"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8005298/pdf/nihms-1680563.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25543966","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 229
Topic-Based Measures of Conversation for Detecting Mild CognitiveImpairment 基于话题的谈话方法检测轻度认知障碍
Pub Date : 2020-07-01 DOI: 10.18653/v1/2020.nlpmc-1.9
Meysam Asgari, Liu Chen, H. Dodge
Conversation is a complex cognitive task that engages multiple aspects of cognitive functions to remember the discussed topics, monitor the semantic and linguistic elements, and recognize others’ emotions. In this paper, we propose a computational method based on the lexical coherence of consecutive utterances to quantify topical variations in semi-structured conversations of older adults with cognitive impairments. Extracting the lexical knowledge of conversational utterances, our method generate a set of novel conversational measures that indicate underlying cognitive deficits among subjects with mild cognitive impairment (MCI). Our preliminary results verifies the utility of the proposed conversation-based measures in distinguishing MCI from healthy controls.
对话是一项复杂的认知任务,涉及多个方面的认知功能,包括记住所讨论的话题,监控语义和语言元素,以及识别他人的情绪。在本文中,我们提出了一种基于连续话语的词汇连贯性的计算方法来量化认知障碍老年人半结构化对话中的话题变化。通过提取会话话语的词汇知识,我们的方法生成了一套新的会话测量方法,这些测量方法可以显示轻度认知障碍(MCI)受试者潜在的认知缺陷。我们的初步结果验证了所提出的基于对话的措施在区分MCI和健康对照方面的实用性。
{"title":"Topic-Based Measures of Conversation for Detecting Mild CognitiveImpairment","authors":"Meysam Asgari, Liu Chen, H. Dodge","doi":"10.18653/v1/2020.nlpmc-1.9","DOIUrl":"https://doi.org/10.18653/v1/2020.nlpmc-1.9","url":null,"abstract":"Conversation is a complex cognitive task that engages multiple aspects of cognitive functions to remember the discussed topics, monitor the semantic and linguistic elements, and recognize others’ emotions. In this paper, we propose a computational method based on the lexical coherence of consecutive utterances to quantify topical variations in semi-structured conversations of older adults with cognitive impairments. Extracting the lexical knowledge of conversational utterances, our method generate a set of novel conversational measures that indicate underlying cognitive deficits among subjects with mild cognitive impairment (MCI). Our preliminary results verifies the utility of the proposed conversation-based measures in distinguishing MCI from healthy controls.","PeriodicalId":74541,"journal":{"name":"Proceedings of the conference. Association for Computational Linguistics. Meeting","volume":"54 1","pages":"63-67"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75754791","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
Proceedings of the conference. Association for Computational Linguistics. Meeting
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1