Zero-shot relation triplet extraction as Next-Sentence Prediction

IF 7.2 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Knowledge-Based Systems Pub Date : 2024-09-27 DOI:10.1016/j.knosys.2024.112507
{"title":"Zero-shot relation triplet extraction as Next-Sentence Prediction","authors":"","doi":"10.1016/j.knosys.2024.112507","DOIUrl":null,"url":null,"abstract":"<div><div>Zero-shot relation triplet extraction (ZeroRTE) endeavors to extract relation triplets from a test set using a model trained on a training set with disjoint relations from the test set. Current ZeroRTE approaches primarily rely on two strategies: 1) Combining pre-trained language models to generate additional training samples; 2) Adding a large number of parameters that require training from scratch on top of a pre-trained language model. However, the former approach does not ensure the quality of generated samples, and the latter often struggles to generalize to unseen relations in the test set, particularly when the training set is small. In this paper, we introduce a novel method, <u>N</u>ext <u>S</u>entence <u>P</u>rediction for <u>R</u>elation <u>T</u>riplet <u>E</u>xtraction (NSP-RTE), abstracting ZeroRTE as a higher-level next sentence prediction (NSP) task to enhance its generalization ability to unseen relation categories. NSP-RTE integrates modules for relation recognition, entity detection, and triplet classification, leveraging pre-trained BERT models with fewer parameters requiring training from scratch, while eliminating the need for additional sample generation. Our experiments on the FewRel and Wiki-ZSL datasets demonstrate that NSP-RTE, with its simple and efficient design, significantly outperforms previous methods.</div></div>","PeriodicalId":49939,"journal":{"name":"Knowledge-Based Systems","volume":null,"pages":null},"PeriodicalIF":7.2000,"publicationDate":"2024-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Knowledge-Based Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0950705124011419","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Zero-shot relation triplet extraction (ZeroRTE) endeavors to extract relation triplets from a test set using a model trained on a training set with disjoint relations from the test set. Current ZeroRTE approaches primarily rely on two strategies: 1) Combining pre-trained language models to generate additional training samples; 2) Adding a large number of parameters that require training from scratch on top of a pre-trained language model. However, the former approach does not ensure the quality of generated samples, and the latter often struggles to generalize to unseen relations in the test set, particularly when the training set is small. In this paper, we introduce a novel method, Next Sentence Prediction for Relation Triplet Extraction (NSP-RTE), abstracting ZeroRTE as a higher-level next sentence prediction (NSP) task to enhance its generalization ability to unseen relation categories. NSP-RTE integrates modules for relation recognition, entity detection, and triplet classification, leveraging pre-trained BERT models with fewer parameters requiring training from scratch, while eliminating the need for additional sample generation. Our experiments on the FewRel and Wiki-ZSL datasets demonstrate that NSP-RTE, with its simple and efficient design, significantly outperforms previous methods.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
作为下一句预测的零次关系三连抽取
零次关系三元组提取(ZeroRTE)是指使用在测试集的不相关关系训练集上训练的模型,从测试集中提取关系三元组。目前的 ZeroRTE 方法主要依赖两种策略:1) 结合预先训练好的语言模型来生成额外的训练样本;2) 在预先训练好的语言模型上添加大量需要从头开始训练的参数。然而,前一种方法无法确保生成样本的质量,而后一种方法往往难以泛化到测试集中未见的关系,尤其是当训练集较小时。在本文中,我们介绍了一种新方法--关系三重提取的下一句预测(NSP-RTE),它将 ZeroRTE 抽象为更高层次的下一句预测(NSP)任务,以增强其对未见关系类别的泛化能力。NSP-RTE 集成了关系识别、实体检测和三连音分类模块,利用预先训练的 BERT 模型,减少了需要从头开始训练的参数,同时消除了生成额外样本的需要。我们在 FewRel 和 Wikii-ZSL 数据集上的实验表明,NSP-RTE 凭借其简单高效的设计,显著超越了之前的方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Knowledge-Based Systems
Knowledge-Based Systems 工程技术-计算机:人工智能
CiteScore
14.80
自引率
12.50%
发文量
1245
审稿时长
7.8 months
期刊介绍: Knowledge-Based Systems, an international and interdisciplinary journal in artificial intelligence, publishes original, innovative, and creative research results in the field. It focuses on knowledge-based and other artificial intelligence techniques-based systems. The journal aims to support human prediction and decision-making through data science and computation techniques, provide a balanced coverage of theory and practical study, and encourage the development and implementation of knowledge-based intelligence models, methods, systems, and software tools. Applications in business, government, education, engineering, and healthcare are emphasized.
期刊最新文献
Automated grasp labeling and detection framework with pixel-level precision Semi-supervised intrusion detection system for in-vehicle networks based on variational autoencoder and adversarial reinforcement learning Bridging spatiotemporal feature gap for video salient object detection Privacy preservation in deep reinforcement learning: A training perspective Zero-shot relation triplet extraction as Next-Sentence Prediction
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1