首页 > 最新文献

Proceedings of the 2nd Workshop on Deep Learning on Graphs for Natural Language Processing (DLG4NLP 2022)最新文献

英文 中文
LiGCN: Label-interpretable Graph Convolutional Networks for Multi-label Text Classification 用于多标签文本分类的标签可解释图卷积网络
Irene Li, Aosong Feng, Hao Wu, Tianxiao Li, T. Suzumura, Ruihai Dong
Multi-label text classification (MLTC) is an attractive and challenging task in natural language processing (NLP). Compared with single-label text classification, MLTC has a wider range of applications in practice. In this paper, we propose a label-interpretable graph convolutional network model to solve the MLTC problem by modeling tokens and labels as nodes in a heterogeneous graph. In this way, we are able to take into account multiple relationships including token-level relationships. Besides, the model allows better interpretability for predicted labels as the token-label edges are exposed. We evaluate our method on four real-world datasets and it achieves competitive scores against selected baseline methods. Specifically, this model achieves a gain of 0.14 on the F1 score in the small label set MLTC, and 0.07 in the large label set scenario.
多标签文本分类(MLTC)是自然语言处理(NLP)中一个极具挑战性的课题。与单标签文本分类相比,多标签文本分类在实践中有更广泛的应用。在本文中,我们提出了一个标签可解释的图卷积网络模型,通过将标记和标记建模为异构图中的节点来解决MLTC问题。通过这种方式,我们能够考虑多种关系,包括令牌级关系。此外,该模型允许更好的可解释性预测标签,因为令牌标签的边缘是公开的。我们在四个真实世界的数据集上评估了我们的方法,它与选定的基线方法相比获得了具有竞争力的分数。具体来说,该模型在小标签集MLTC场景下的F1得分增益为0.14,在大标签集场景下的F1得分增益为0.07。
{"title":"LiGCN: Label-interpretable Graph Convolutional Networks for Multi-label Text Classification","authors":"Irene Li, Aosong Feng, Hao Wu, Tianxiao Li, T. Suzumura, Ruihai Dong","doi":"10.18653/v1/2022.dlg4nlp-1.7","DOIUrl":"https://doi.org/10.18653/v1/2022.dlg4nlp-1.7","url":null,"abstract":"Multi-label text classification (MLTC) is an attractive and challenging task in natural language processing (NLP). Compared with single-label text classification, MLTC has a wider range of applications in practice. In this paper, we propose a label-interpretable graph convolutional network model to solve the MLTC problem by modeling tokens and labels as nodes in a heterogeneous graph. In this way, we are able to take into account multiple relationships including token-level relationships. Besides, the model allows better interpretability for predicted labels as the token-label edges are exposed. We evaluate our method on four real-world datasets and it achieves competitive scores against selected baseline methods. Specifically, this model achieves a gain of 0.14 on the F1 score in the small label set MLTC, and 0.07 in the large label set scenario.","PeriodicalId":367475,"journal":{"name":"Proceedings of the 2nd Workshop on Deep Learning on Graphs for Natural Language Processing (DLG4NLP 2022)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126545518","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Improving Neural Machine Translation with the Abstract Meaning Representation by Combining Graph and Sequence Transformers 结合图变换和序列变换的抽象意义表示改进神经机器翻译
Changmao Li, Jeffrey Flanigan
Previous studies have shown that the Abstract Meaning Representation (AMR) can improve Neural Machine Translation (NMT). However, there has been little work investigating incorporating AMR graphs into Transformer models. In this work, we propose a novel encoder-decoder architecture which augments the Transformer model with a Heterogeneous Graph Transformer (Yao et al., 2020) which encodes source sentence AMR graphs. Experimental results demonstrate the proposed model outperforms the Transformer model and previous non-Transformer based models on two different language pairs in both the high resource setting and low resource setting. Our source code, training corpus and released models are available at https://github.com/jlab-nlp/amr-nmt.
已有研究表明,抽象意义表示(AMR)可以提高神经机器翻译(NMT)的翻译效率。然而,很少有研究将AMR图合并到Transformer模型中。在这项工作中,我们提出了一种新的编码器-解码器架构,该架构使用异构图转换器(Yao et al., 2020)增强了Transformer模型,该模型编码源句子AMR图。实验结果表明,该模型在高资源环境和低资源环境下,在两种不同的语言对上都优于Transformer模型和以前的非Transformer模型。我们的源代码、训练语料库和发布的模型可在https://github.com/jlab-nlp/amr-nmt上获得。
{"title":"Improving Neural Machine Translation with the Abstract Meaning Representation by Combining Graph and Sequence Transformers","authors":"Changmao Li, Jeffrey Flanigan","doi":"10.18653/v1/2022.dlg4nlp-1.2","DOIUrl":"https://doi.org/10.18653/v1/2022.dlg4nlp-1.2","url":null,"abstract":"Previous studies have shown that the Abstract Meaning Representation (AMR) can improve Neural Machine Translation (NMT). However, there has been little work investigating incorporating AMR graphs into Transformer models. In this work, we propose a novel encoder-decoder architecture which augments the Transformer model with a Heterogeneous Graph Transformer (Yao et al., 2020) which encodes source sentence AMR graphs. Experimental results demonstrate the proposed model outperforms the Transformer model and previous non-Transformer based models on two different language pairs in both the high resource setting and low resource setting. Our source code, training corpus and released models are available at https://github.com/jlab-nlp/amr-nmt.","PeriodicalId":367475,"journal":{"name":"Proceedings of the 2nd Workshop on Deep Learning on Graphs for Natural Language Processing (DLG4NLP 2022)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127893194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Graph Neural Networks for Adapting Off-the-shelf General Domain Language Models to Low-Resource Specialised Domains 基于图神经网络的通用领域语言模型在低资源特殊领域中的应用
Mérième Bouhandi, E. Morin, Thierry Hamon
Language models encode linguistic proprieties and are used as input for more specific models. Using their word representations as-is for specialised and low-resource domains might be less efficient. Methods of adapting them exist, but these models often overlook global information about how words, terms, and concepts relate to each other in a corpus due to their strong reliance on attention. We consider that global information can influence the results of the downstream tasks, and combination with contextual information is performed using graph convolution networks or GCN built on vocabulary graphs. By outperforming baselines, we show that this architecture is profitable for domain-specific tasks.
语言模型对语言特性进行编码,并用作更具体模型的输入。对于专门的和低资源的领域,使用它们的单词表示可能效率较低。虽然存在调整它们的方法,但这些模型往往忽略了语料库中单词、术语和概念如何相互关联的全局信息,因为它们强烈依赖于注意力。我们考虑到全局信息可以影响下游任务的结果,并使用基于词汇图的图卷积网络或GCN与上下文信息进行组合。通过超越基线,我们表明该体系结构对于特定于领域的任务是有益的。
{"title":"Graph Neural Networks for Adapting Off-the-shelf General Domain Language Models to Low-Resource Specialised Domains","authors":"Mérième Bouhandi, E. Morin, Thierry Hamon","doi":"10.18653/v1/2022.dlg4nlp-1.5","DOIUrl":"https://doi.org/10.18653/v1/2022.dlg4nlp-1.5","url":null,"abstract":"Language models encode linguistic proprieties and are used as input for more specific models. Using their word representations as-is for specialised and low-resource domains might be less efficient. Methods of adapting them exist, but these models often overlook global information about how words, terms, and concepts relate to each other in a corpus due to their strong reliance on attention. We consider that global information can influence the results of the downstream tasks, and combination with contextual information is performed using graph convolution networks or GCN built on vocabulary graphs. By outperforming baselines, we show that this architecture is profitable for domain-specific tasks.","PeriodicalId":367475,"journal":{"name":"Proceedings of the 2nd Workshop on Deep Learning on Graphs for Natural Language Processing (DLG4NLP 2022)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115430412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Scene Graph Parsing via Abstract Meaning Representation in Pre-trained Language Models 基于预训练语言模型的抽象意义表示的场景图解析
Woo Suk Choi, Y. Heo, Dharani Punithan, Byoung-Tak Zhang
In this work, we propose the application of abstract meaning representation (AMR) based semantic parsing models to parse textual descriptions of a visual scene into scene graphs, which is the first work to the best of our knowledge. Previous works examined scene graph parsing from textual descriptions using dependency parsing and left the AMR parsing approach as future work since sophisticated methods are required to apply AMR. Hence, we use pre-trained AMR parsing models to parse the region descriptions of visual scenes (i.e. images) into AMR graphs and pre-trained language models (PLM), BART and T5, to parse AMR graphs into scene graphs. The experimental results show that our approach explicitly captures high-level semantics from textual descriptions of visual scenes, such as objects, attributes of objects, and relationships between objects. Our textual scene graph parsing approach outperforms the previous state-of-the-art results by 9.3% in the SPICE metric score.
在这项工作中,我们提出应用基于抽象意义表示(AMR)的语义分析模型将视觉场景的文本描述解析为场景图,这是我们所知的第一项工作。以前的工作使用依赖关系分析从文本描述中分析场景图,并将AMR分析方法作为未来的工作,因为需要复杂的方法来应用AMR。因此,我们使用预训练的AMR解析模型将视觉场景(即图像)的区域描述解析为AMR图,并使用预训练的语言模型(PLM) BART和T5将AMR图解析为场景图。实验结果表明,我们的方法明确地从视觉场景的文本描述中捕获高级语义,例如对象、对象的属性和对象之间的关系。我们的文本场景图解析方法在SPICE度量得分上比以前最先进的结果高出9.3%。
{"title":"Scene Graph Parsing via Abstract Meaning Representation in Pre-trained Language Models","authors":"Woo Suk Choi, Y. Heo, Dharani Punithan, Byoung-Tak Zhang","doi":"10.18653/v1/2022.dlg4nlp-1.4","DOIUrl":"https://doi.org/10.18653/v1/2022.dlg4nlp-1.4","url":null,"abstract":"In this work, we propose the application of abstract meaning representation (AMR) based semantic parsing models to parse textual descriptions of a visual scene into scene graphs, which is the first work to the best of our knowledge. Previous works examined scene graph parsing from textual descriptions using dependency parsing and left the AMR parsing approach as future work since sophisticated methods are required to apply AMR. Hence, we use pre-trained AMR parsing models to parse the region descriptions of visual scenes (i.e. images) into AMR graphs and pre-trained language models (PLM), BART and T5, to parse AMR graphs into scene graphs. The experimental results show that our approach explicitly captures high-level semantics from textual descriptions of visual scenes, such as objects, attributes of objects, and relationships between objects. Our textual scene graph parsing approach outperforms the previous state-of-the-art results by 9.3% in the SPICE metric score.","PeriodicalId":367475,"journal":{"name":"Proceedings of the 2nd Workshop on Deep Learning on Graphs for Natural Language Processing (DLG4NLP 2022)","volume":"47 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134260439","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Explicit Graph Reasoning Fusing Knowledge and Contextual Information for Multi-hop Question Answering 融合知识和上下文信息的多跳问答显式图推理
Zhenyun Deng, Yonghua Zhu, Qianqian Qi, M. Witbrock, Patricia J. Riddle
Current graph-neural-network-based (GNN-based) approaches to multi-hop questions integrate clues from scattered paragraphs in an entity graph, achieving implicit reasoning by synchronous update of graph node representations using information from neighbours; this is poorly suited for explaining how clues are passed through the graph in hops. In this paper, we describe a structured Knowledge and contextual Information Fusion GNN (KIFGraph) whose explicit multi-hop graph reasoning mimics human step by step reasoning. Specifically, we first integrate clues at multiple levels of granularity (question, paragraph, sentence, entity) as nodes in the graph, connected by edges derived using structured semantic knowledge, then use a contextual encoder to obtain the initial node representations, followed by step-by-step two-stage graph reasoning that asynchronously updates node representations. Each node can be related to its neighbour nodes through fused structured knowledge and contextual information, reliably integrating their answer clues. Moreover, a masked attention mechanism (MAM) filters out noisy or redundant nodes and edges, to avoid ineffective clue propagation in graph reasoning. Experimental results show performance competitive with published models on the HotpotQA dataset.
当前基于图神经网络(gnn)的多跳问题方法整合了实体图中分散段落的线索,通过使用邻居信息同步更新图节点表示来实现隐式推理;这并不适合解释线索是如何在图形中跳跃传递的。在本文中,我们描述了一个结构化的知识与上下文信息融合GNN (KIFGraph),其显式多跳图推理模仿人类逐步推理。具体来说,我们首先将多个粒度级别(问题、段落、句子、实体)的线索集成为图中的节点,通过使用结构化语义知识派生的边连接起来,然后使用上下文编码器获得初始节点表示,随后进行分步两阶段图推理,异步更新节点表示。每个节点可以通过融合结构化知识和上下文信息与相邻节点建立联系,可靠地整合它们的答案线索。此外,该算法还采用了一种屏蔽注意机制(MAM)来过滤掉有噪声或冗余的节点和边缘,以避免图推理中无效的线索传播。实验结果表明,在HotpotQA数据集上,该模型的性能与已发表的模型相当。
{"title":"Explicit Graph Reasoning Fusing Knowledge and Contextual Information for Multi-hop Question Answering","authors":"Zhenyun Deng, Yonghua Zhu, Qianqian Qi, M. Witbrock, Patricia J. Riddle","doi":"10.18653/v1/2022.dlg4nlp-1.8","DOIUrl":"https://doi.org/10.18653/v1/2022.dlg4nlp-1.8","url":null,"abstract":"Current graph-neural-network-based (GNN-based) approaches to multi-hop questions integrate clues from scattered paragraphs in an entity graph, achieving implicit reasoning by synchronous update of graph node representations using information from neighbours; this is poorly suited for explaining how clues are passed through the graph in hops. In this paper, we describe a structured Knowledge and contextual Information Fusion GNN (KIFGraph) whose explicit multi-hop graph reasoning mimics human step by step reasoning. Specifically, we first integrate clues at multiple levels of granularity (question, paragraph, sentence, entity) as nodes in the graph, connected by edges derived using structured semantic knowledge, then use a contextual encoder to obtain the initial node representations, followed by step-by-step two-stage graph reasoning that asynchronously updates node representations. Each node can be related to its neighbour nodes through fused structured knowledge and contextual information, reliably integrating their answer clues. Moreover, a masked attention mechanism (MAM) filters out noisy or redundant nodes and edges, to avoid ineffective clue propagation in graph reasoning. Experimental results show performance competitive with published models on the HotpotQA dataset.","PeriodicalId":367475,"journal":{"name":"Proceedings of the 2nd Workshop on Deep Learning on Graphs for Natural Language Processing (DLG4NLP 2022)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122390428","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the 2nd Workshop on Deep Learning on Graphs for Natural Language Processing (DLG4NLP 2022)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1