首页 > 最新文献

2021 IEEE International Conference on Big Knowledge (ICBK)最新文献

英文 中文
A Character-Word Graph Attention Networks for Chinese Text Classification 中文文本分类的字词图注意网络
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00068
Shigang Yang, Yongguo Liu
Text classification is an important task in natural language processing. Different from English, Chinese text owns two representations, character-level and word-level. The former has abundant connotations and the latter owns specific meanings. Current researches often simply concatenated two-level features with little processing and failed to explore the affiliation relation-ship between Chinese characters and words. In this paper, we proposed a character-word graph attention network (CW-GAT) to explore the interactive information between characters and words for Chinese text classification. A graph attention network is adopted to capture the context of sentences and the interaction between characters and words. Extensive experiments on six real Chinese text datasets show that the proposed model outperforms the latest baseline methods.
文本分类是自然语言处理中的一项重要任务。与英语不同的是,汉语文本有两种表征:字符层和词层。前者具有丰富的内涵,后者具有特定的意义。目前的研究往往是简单地将两层特征拼接起来,处理较少,未能探究汉字与词之间的隶属关系。本文提出了一种字符-词图关注网络(CW-GAT)来探索汉字与词之间的交互信息,用于中文文本分类。采用图注意网络来捕捉句子的语境和字词之间的相互作用。在六个真实中文文本数据集上的大量实验表明,该模型优于最新的基线方法。
{"title":"A Character-Word Graph Attention Networks for Chinese Text Classification","authors":"Shigang Yang, Yongguo Liu","doi":"10.1109/ICKG52313.2021.00068","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00068","url":null,"abstract":"Text classification is an important task in natural language processing. Different from English, Chinese text owns two representations, character-level and word-level. The former has abundant connotations and the latter owns specific meanings. Current researches often simply concatenated two-level features with little processing and failed to explore the affiliation relation-ship between Chinese characters and words. In this paper, we proposed a character-word graph attention network (CW-GAT) to explore the interactive information between characters and words for Chinese text classification. A graph attention network is adopted to capture the context of sentences and the interaction between characters and words. Extensive experiments on six real Chinese text datasets show that the proposed model outperforms the latest baseline methods.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128326508","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Semi-supervised Bilingual Lexicon Induction Method for Distant Language Pairs Based on Bidirectional Adversarial Model 基于双向对抗模型的远程语言对半监督双语词典归纳方法
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00046
Wenwu Zhi, Yuhong Zhang
Bilingual lexicon induction (BLI) can transfer knowledgefrom well- to under- resourced language, and has been widelyapplied to various NLP tasks. Recent work on BLI is projection-based that learns a mapping to connect source and target embedding spaces, with the isomorphism assumption. Unfortunately, the isomorphism assumption doesn't hold gener-ally, especially in typologically distant language pairs. Moreover, without supervised signals guiding, the training will further com-plicates BLI, making the performance of unsupervised methods unsatisfactory. To broke the restrict of isomorphism, we propose a semi-supervised method for distant BLI tasks, named A Semi-supervised Bilingual Lexicon Induction method in Latent Space based on Bidirectional Adversarial Model. First, two latent spaces are learned by two autoencoders for source and target domain independently to weaken the constraint of isomorphism in the embedding spaces. Then we add a few pairs of dictionary to learn the initial mapping to connect the Latent Space. Last, based on initial mapping, Cycle-Consistency is combined with Distance constraint constraint to maintain the geometry structure of both embedding spaces stable in the learning of bi-direction mapping based on adversarial model. By conducting extensive experiments, our method gets state-of-the-art results on most language pairs, especially with significant improvements on distant language pairs.
双语词汇归纳(BLI)可以将知识从资源丰富的语言转移到资源不足的语言,并已广泛应用于各种自然语言处理任务。最近的BLI研究是基于投影的,它在同构假设下学习映射来连接源和目标嵌入空间。不幸的是,同构假设并不普遍成立,特别是在类型学上距离较远的语言对中。此外,如果没有监督信号的引导,训练将使BLI进一步复杂化,使无监督方法的性能不理想。为了打破同构的限制,我们提出了一种用于远程BLI任务的半监督方法,即基于双向对抗模型的潜在空间半监督双语词典归纳方法。首先,两个自编码器分别学习源域和目标域的两个隐空间,以削弱嵌入空间中同构的约束;然后添加几对字典学习初始映射,连接潜空间。最后,在初始映射的基础上,在基于对抗模型的双向映射学习中,将循环一致性与距离约束相结合,保持两个嵌入空间的几何结构稳定。通过大量的实验,我们的方法在大多数语言对上得到了最先进的结果,特别是在远距离语言对上有了显著的改进。
{"title":"A Semi-supervised Bilingual Lexicon Induction Method for Distant Language Pairs Based on Bidirectional Adversarial Model","authors":"Wenwu Zhi, Yuhong Zhang","doi":"10.1109/ICKG52313.2021.00046","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00046","url":null,"abstract":"Bilingual lexicon induction (BLI) can transfer knowledgefrom well- to under- resourced language, and has been widelyapplied to various NLP tasks. Recent work on BLI is projection-based that learns a mapping to connect source and target embedding spaces, with the isomorphism assumption. Unfortunately, the isomorphism assumption doesn't hold gener-ally, especially in typologically distant language pairs. Moreover, without supervised signals guiding, the training will further com-plicates BLI, making the performance of unsupervised methods unsatisfactory. To broke the restrict of isomorphism, we propose a semi-supervised method for distant BLI tasks, named A Semi-supervised Bilingual Lexicon Induction method in Latent Space based on Bidirectional Adversarial Model. First, two latent spaces are learned by two autoencoders for source and target domain independently to weaken the constraint of isomorphism in the embedding spaces. Then we add a few pairs of dictionary to learn the initial mapping to connect the Latent Space. Last, based on initial mapping, Cycle-Consistency is combined with Distance constraint constraint to maintain the geometry structure of both embedding spaces stable in the learning of bi-direction mapping based on adversarial model. By conducting extensive experiments, our method gets state-of-the-art results on most language pairs, especially with significant improvements on distant language pairs.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"96 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125178566","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Fair Representation Learning in Knowledge-aware Recommendation 知识感知推荐中的公平代表学习
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00058
Bingke Xu, Yue Cui, Zipeng Sun, Liwei Deng, Kai Zheng
Knowledge-aware recommendation system has at-tracted considerable interest in academia and industry, which comes in handy to solve the cold-start problem and offer a reliable solution for the business to grow. It's particularly important to consider fairness issues when designing and using those systems. However, we find that though not explicitly introduced to a knowledge graph (KG), sensitive information can be implicitly learned by a recommender and thus leads to unfairness. Most existing debiasing methods require sophisticated model design or can only be applied to specific base models. In this paper, to address the above problems, we propose a method to ensure the fairness of any knowledge-aware recommendation models by introducing a sensitivity graph. Different from the majority of previous studies that only handle a single protected attribute, we also aim to make our method flexible to different combinations of fairness constraints during inference. Specifically, given a knowledge-based recommendation model, we first construct a sensitivity graph by taking protected attributes as nodes and dynamically learned relations between pairs of attributes as edges. Then we merge the sensitivity graph into the original knowledge graph and introduce an adversarial framework to enhance fairness criterion by extracting sensitive information of users from the original KG during the graph representation process, without changing the KG-based recommendation model. Extensive experimental results on two public real-world datasets show that the proposed framework can achieve state-of-the-art performance on improving the fairness of any KG-based recommendation model while only cause trivial overall accuracy declination.
知识感知推荐系统已经引起了学术界和工业界的极大兴趣,它可以很好地解决冷启动问题,为企业的发展提供可靠的解决方案。在设计和使用这些系统时,考虑公平性问题尤为重要。然而,我们发现,虽然没有明确地引入知识图(KG),但敏感信息可以被推荐器隐式学习,从而导致不公平。大多数现有的去偏方法需要复杂的模型设计,或者只能应用于特定的基础模型。针对上述问题,本文提出了一种通过引入敏感性图来保证任何知识感知推荐模型的公平性的方法。与以往大多数只处理单个保护属性的研究不同,我们的目标是使我们的方法在推理过程中能够灵活地适应不同的公平性约束组合。具体而言,给出基于知识的推荐模型,首先以受保护的属性为节点,动态学习属性对之间的关系为边,构造敏感性图;然后,在不改变基于KG的推荐模型的前提下,将灵敏度图合并到原始知识图中,引入对抗框架,在图表示过程中从原始KG中提取用户敏感信息,增强公平性准则。在两个公开的真实世界数据集上的大量实验结果表明,所提出的框架在提高任何基于kg的推荐模型的公平性方面都能达到最先进的性能,而只会导致轻微的总体准确性下降。
{"title":"Fair Representation Learning in Knowledge-aware Recommendation","authors":"Bingke Xu, Yue Cui, Zipeng Sun, Liwei Deng, Kai Zheng","doi":"10.1109/ICKG52313.2021.00058","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00058","url":null,"abstract":"Knowledge-aware recommendation system has at-tracted considerable interest in academia and industry, which comes in handy to solve the cold-start problem and offer a reliable solution for the business to grow. It's particularly important to consider fairness issues when designing and using those systems. However, we find that though not explicitly introduced to a knowledge graph (KG), sensitive information can be implicitly learned by a recommender and thus leads to unfairness. Most existing debiasing methods require sophisticated model design or can only be applied to specific base models. In this paper, to address the above problems, we propose a method to ensure the fairness of any knowledge-aware recommendation models by introducing a sensitivity graph. Different from the majority of previous studies that only handle a single protected attribute, we also aim to make our method flexible to different combinations of fairness constraints during inference. Specifically, given a knowledge-based recommendation model, we first construct a sensitivity graph by taking protected attributes as nodes and dynamically learned relations between pairs of attributes as edges. Then we merge the sensitivity graph into the original knowledge graph and introduce an adversarial framework to enhance fairness criterion by extracting sensitive information of users from the original KG during the graph representation process, without changing the KG-based recommendation model. Extensive experimental results on two public real-world datasets show that the proposed framework can achieve state-of-the-art performance on improving the fairness of any KG-based recommendation model while only cause trivial overall accuracy declination.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125520731","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Query-focused Abstractive Summarization via Question-answering Model 基于问答模型的以查询为中心的抽象摘要
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00065
Jiachen Du, Yang Gao
Text summarization is a task that creates a short version of a document while preserving the main content. In the age of information explosion, how to obtain the content that users care about from a large amount of information becomes par-ticularly significant. Under these circumstances, query-focused abstractive summarization (QFS) becomes more dominant since it is able to focus on user needs while generating fluent, con-cise, succinct paraphrased summaries. However, different from generic summarization that has achieved remarkable results driven by a large scale of parallel data, the QFS is suffering from lacking enough parallel corpus. To address the above issues, in this paper, we migrate the large-scale generic summarization datasets into query-focused datasets while preserving the informative summaries. Based on the synthetic queries and data, we proposed a new model, called SQAS, which is capable of extracting fine-grained factual information with respect to a specific question, and take into account the reasoning information by understanding the source document leveraged by the question-answering model. Receiving the extracted content, the summary generator can not only generate semantically relevant content but also assure fluent and readable sentences thanks to the language generation capability of a pre-trained language model. Experimental results on both generic datasets and query-focused summary datasets demonstrate the effectiveness of our proposed model in terms of automatic ROUGE metrics and investigating real cases.
文本摘要是在保留主要内容的同时创建文档的简短版本的任务。在信息爆炸的时代,如何从海量的信息中获取用户关心的内容就显得尤为重要。在这种情况下,以查询为中心的抽象摘要(QFS)变得更加重要,因为它能够在生成流畅、简洁、简明的释义摘要的同时关注用户需求。然而,与在大量并行数据的驱动下取得显著效果的通用摘要不同,QFS存在缺乏足够的并行语料库的问题。为了解决上述问题,本文在保留信息性摘要的同时,将大规模通用摘要数据集迁移到以查询为中心的数据集。在综合查询和数据的基础上,我们提出了一个新的SQAS模型,该模型能够提取关于特定问题的细粒度事实信息,并通过理解问答模型所利用的源文档来考虑推理信息。摘要生成器接收提取的内容后,利用预先训练好的语言模型的语言生成能力,不仅可以生成语义相关的内容,还可以保证句子的流畅和可读。在通用数据集和以查询为中心的摘要数据集上的实验结果表明,我们提出的模型在自动ROUGE度量和调查真实案例方面是有效的。
{"title":"Query-focused Abstractive Summarization via Question-answering Model","authors":"Jiachen Du, Yang Gao","doi":"10.1109/ICKG52313.2021.00065","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00065","url":null,"abstract":"Text summarization is a task that creates a short version of a document while preserving the main content. In the age of information explosion, how to obtain the content that users care about from a large amount of information becomes par-ticularly significant. Under these circumstances, query-focused abstractive summarization (QFS) becomes more dominant since it is able to focus on user needs while generating fluent, con-cise, succinct paraphrased summaries. However, different from generic summarization that has achieved remarkable results driven by a large scale of parallel data, the QFS is suffering from lacking enough parallel corpus. To address the above issues, in this paper, we migrate the large-scale generic summarization datasets into query-focused datasets while preserving the informative summaries. Based on the synthetic queries and data, we proposed a new model, called SQAS, which is capable of extracting fine-grained factual information with respect to a specific question, and take into account the reasoning information by understanding the source document leveraged by the question-answering model. Receiving the extracted content, the summary generator can not only generate semantically relevant content but also assure fluent and readable sentences thanks to the language generation capability of a pre-trained language model. Experimental results on both generic datasets and query-focused summary datasets demonstrate the effectiveness of our proposed model in terms of automatic ROUGE metrics and investigating real cases.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125631897","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Bridging the Language Gap: Knowledge Injected Multilingual Question Answering 弥合语言鸿沟:知识注入多语言问答
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00052
Zhichao Duan, Xiuxing Li, Zhengyan Zhang, Zhenyu Li, Ning Liu, Jianyong Wang
Question Answering (QA) is the task of automati-cally answering questions posed by humans in natural languages. There are different settings to answer a question, such as abstractive, extractive, boolean, and multiple-choice QA. As a popular topic in natural language processing tasks, extractive question answering task (extractive QA) has gained extensive attention in the past few years. With the continuous evolvement of the world, generalized cross-lingual transfer (G-XLT), where question and answer context are in different languages, poses some unique challenges over cross-lingual transfer (XLT), where question and answer context are in the same language. With the boost of corresponding development of related benchmarks, many works have been done to improve the performance of various language QA tasks. However, only a few works are dedicated to the G-XLT task. In this work, we propose a generalized cross-lingual transfer framework to enhance the model's ability to understand different languages. Specifically, we first assemble triples from different languages to form multilingual knowledge. Since the lack of knowledge between different languages greatly limits models' reasoning ability, we further design a knowledge injection strategy via leveraging link prediction techniques to enrich the model storage of multilingual knowledge. In this way, we can profoundly exploit rich semantic knowledge. Experiment results on real-world datasets MLQA demonstrate that the proposed method can improve the performance by a large margin, outperforming the baseline method by 13.18%/12.00% F1/EM on average.
问答(QA)是用自然语言自动回答人类提出的问题的任务。回答问题有不同的设置,例如抽象、抽取、布尔和多项选择QA。抽取问答任务作为自然语言处理任务中的一个热门话题,近年来得到了广泛的关注。随着世界的不断发展,问答语境为不同语言的广义跨语言迁移(G-XLT)对问答语境为同一语言的跨语言迁移(XLT)提出了一些独特的挑战。随着相关基准开发的推动,人们做了很多工作来提高各种语言QA任务的性能。然而,只有少数作品专门用于G-XLT任务。在这项工作中,我们提出了一个广义的跨语言迁移框架,以提高模型理解不同语言的能力。具体来说,我们首先将来自不同语言的三元组组合起来,形成多语言知识。由于不同语言之间缺乏知识,极大地限制了模型的推理能力,我们进一步设计了一种知识注入策略,利用链接预测技术来丰富多语言知识的模型存储。通过这种方式,我们可以深刻地挖掘丰富的语义知识。在真实数据集MLQA上的实验结果表明,该方法的性能有较大的提高,平均优于基准方法13.18%/12.00%的F1/EM。
{"title":"Bridging the Language Gap: Knowledge Injected Multilingual Question Answering","authors":"Zhichao Duan, Xiuxing Li, Zhengyan Zhang, Zhenyu Li, Ning Liu, Jianyong Wang","doi":"10.1109/ICKG52313.2021.00052","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00052","url":null,"abstract":"Question Answering (QA) is the task of automati-cally answering questions posed by humans in natural languages. There are different settings to answer a question, such as abstractive, extractive, boolean, and multiple-choice QA. As a popular topic in natural language processing tasks, extractive question answering task (extractive QA) has gained extensive attention in the past few years. With the continuous evolvement of the world, generalized cross-lingual transfer (G-XLT), where question and answer context are in different languages, poses some unique challenges over cross-lingual transfer (XLT), where question and answer context are in the same language. With the boost of corresponding development of related benchmarks, many works have been done to improve the performance of various language QA tasks. However, only a few works are dedicated to the G-XLT task. In this work, we propose a generalized cross-lingual transfer framework to enhance the model's ability to understand different languages. Specifically, we first assemble triples from different languages to form multilingual knowledge. Since the lack of knowledge between different languages greatly limits models' reasoning ability, we further design a knowledge injection strategy via leveraging link prediction techniques to enrich the model storage of multilingual knowledge. In this way, we can profoundly exploit rich semantic knowledge. Experiment results on real-world datasets MLQA demonstrate that the proposed method can improve the performance by a large margin, outperforming the baseline method by 13.18%/12.00% F1/EM on average.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127895673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
ICBK 2021 Keynote Abstracts ICBK 2021主题演讲摘要
Pub Date : 2021-12-01 DOI: 10.1109/ickg52313.2021.00009
Allen Bundy
{"title":"ICBK 2021 Keynote Abstracts","authors":"Allen Bundy","doi":"10.1109/ickg52313.2021.00009","DOIUrl":"https://doi.org/10.1109/ickg52313.2021.00009","url":null,"abstract":"","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123005867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-task Learning for Multi-turn Dialogue Generation with Topic Drift Modeling 基于主题漂移建模的多回合对话生成多任务学习
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00061
Hongwei Zeng, Zhenjie Hong, J. Liu, Bifan Wei
Multi-turn dialogue generation aims to generate natural and fluent responses that should be consistent with multiple consecutive utterances history. It is a more challenging task compared to its single-turn counterpart since it requires the model to capture the topic drift along with the multi-turn dialogue history. In this paper, we propose a multi-turn dialogue generation model which incorporates topic drift aware information into a hierarchical encoder-decoder framework to generate coherent responses. This model first utilizes a Convolutional Neural Network (CNN) based topic model to obtain the topic representation of each utterance. Then a topic drift model is employed to encode the sequential topics of multi-turn dialogue history to infer the topic of response. During the response generation, a specially designed topic drift aware generator is proposed to dynamically balance the impact of the inferred topic of response and local word structure. Fur-thermore, we employ multi-task learning to optimize the topic drift model and dialogue generation simultaneously. Extensive experimental results on two benchmark datasets (i.e. Cornell Movie Dialog Corpus and Ubuntu Dialogue Dataset) indicate that our proposed model can generate more coherent responses, and significantly outperform other dialogue generation models.
多回合对话生成的目的是生成自然流畅的响应,该响应应与多个连续的话语历史相一致。与单回合相比,这是一项更具挑战性的任务,因为它需要模型捕捉主题漂移以及多回合对话历史。在本文中,我们提出了一种多回合对话生成模型,该模型将主题漂移感知信息融入到分层编码器-解码器框架中,以产生连贯的响应。该模型首先利用基于卷积神经网络(CNN)的主题模型来获取每个话语的主题表示。然后利用话题漂移模型对多回合对话历史的顺序话题进行编码,从而推断出应答的话题。在响应生成过程中,提出了一个专门设计的主题漂移感知生成器,以动态平衡响应的推断主题和局部词结构的影响。此外,我们采用多任务学习同时优化话题漂移模型和对话生成。在两个基准数据集(即Cornell Movie对白语料库和Ubuntu对白数据集)上的大量实验结果表明,我们提出的模型可以生成更连贯的响应,并且显著优于其他对白生成模型。
{"title":"Multi-task Learning for Multi-turn Dialogue Generation with Topic Drift Modeling","authors":"Hongwei Zeng, Zhenjie Hong, J. Liu, Bifan Wei","doi":"10.1109/ICKG52313.2021.00061","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00061","url":null,"abstract":"Multi-turn dialogue generation aims to generate natural and fluent responses that should be consistent with multiple consecutive utterances history. It is a more challenging task compared to its single-turn counterpart since it requires the model to capture the topic drift along with the multi-turn dialogue history. In this paper, we propose a multi-turn dialogue generation model which incorporates topic drift aware information into a hierarchical encoder-decoder framework to generate coherent responses. This model first utilizes a Convolutional Neural Network (CNN) based topic model to obtain the topic representation of each utterance. Then a topic drift model is employed to encode the sequential topics of multi-turn dialogue history to infer the topic of response. During the response generation, a specially designed topic drift aware generator is proposed to dynamically balance the impact of the inferred topic of response and local word structure. Fur-thermore, we employ multi-task learning to optimize the topic drift model and dialogue generation simultaneously. Extensive experimental results on two benchmark datasets (i.e. Cornell Movie Dialog Corpus and Ubuntu Dialogue Dataset) indicate that our proposed model can generate more coherent responses, and significantly outperform other dialogue generation models.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125359986","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Label Distribution Learning by Exploiting Feature-Label Correlations Locally 利用局部特征-标签相关性的标签分布学习
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00021
Gui-Lin Li, Heng-Ru Zhang, Yuan-Yuan Xu, Yaoyao Lv, Fan Min
Label distribution learning (LDL) is a novel learning paradigm that predicts the degree of representation of multiple labels to an instance. Existing algorithms use all features to predict label distribution. However, each label is often related to part of the features, hence considering other irrelevant features may lead to deviation in both instance searching and model prediction. In this paper, we propose a new LDL algorithm by exploiting the local correlation between features and labels (LDL-LCFL). The main idea is to exploit the local correlations between features and labels, which will be used in the improved k NN algorithm for prediction. Experiments were conducted on eight well-known label distribution data sets with four distance measurements and two similarity measurements. Results show that compared with nine popular LDL methods, our algorithm's prediction ranking is superior.
标签分布学习(LDL)是一种新的学习范式,用于预测多个标签对一个实例的表示程度。现有算法使用所有特征来预测标签分布。然而,每个标签通常与部分特征相关,因此考虑其他不相关的特征可能会导致实例搜索和模型预测出现偏差。在本文中,我们提出了一种新的LDL算法,利用特征和标签之间的局部相关性(LDL- lcfl)。主要思想是利用特征和标签之间的局部相关性,这将用于改进的k神经网络算法进行预测。在8个已知标签分布数据集上进行了4次距离测量和2次相似度测量。结果表明,与常用的9种低密度脂蛋白方法相比,本文算法的预测排序更优。
{"title":"Label Distribution Learning by Exploiting Feature-Label Correlations Locally","authors":"Gui-Lin Li, Heng-Ru Zhang, Yuan-Yuan Xu, Yaoyao Lv, Fan Min","doi":"10.1109/ICKG52313.2021.00021","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00021","url":null,"abstract":"Label distribution learning (LDL) is a novel learning paradigm that predicts the degree of representation of multiple labels to an instance. Existing algorithms use all features to predict label distribution. However, each label is often related to part of the features, hence considering other irrelevant features may lead to deviation in both instance searching and model prediction. In this paper, we propose a new LDL algorithm by exploiting the local correlation between features and labels (LDL-LCFL). The main idea is to exploit the local correlations between features and labels, which will be used in the improved k NN algorithm for prediction. Experiments were conducted on eight well-known label distribution data sets with four distance measurements and two similarity measurements. Results show that compared with nine popular LDL methods, our algorithm's prediction ranking is superior.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122878525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Graph Neural Network for Ethereum Fraud Detection 用于以太坊欺诈检测的图神经网络
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00020
Runnan Tan, Qingfeng Tan, Peng Zhang, Zhao Li
Currently, the blockchain technology has been widely applied to various industries, and has attracted wide attention. However, because of its unique anonymity, digital currency has become a haven for all kinds of cyber crimes. It has been reported that Ethereum frauds provide huge profits, and pose a serious threat to the financial security of the Ethereum network. To create a desired financial environment, an effective method is urgently needed to automatically detect and identify Ethereum frauds in the governance of the Ethereum system. In view of this, this paper proposes a method for detecting Ethereum frauds by mining Ethereum-based transaction records. Specifically, web crawlers are used to capture labeled fraudulent addresses, and then a transaction network is reconstructed based on the public transaction book. Then, an amount-based network embedding algorithm is proposed to extract node features for identifying fraudulent transactions. At last, the graph convolutional network model is used to classify addresses into legal addresses and fraudulent addresses. The experimental results show that the system for detecting fraudulent transactions can achieve the accuracy of 95%, which reflects the excellent performance of the system for detecting Ethereum fraudulent transactions.
目前,区块链技术已广泛应用于各个行业,引起了广泛的关注。然而,由于其独特的匿名性,数字货币成为了各种网络犯罪的避风港。据报道,以太坊欺诈提供了巨额利润,并对以太坊网络的财务安全构成了严重威胁。为了创造理想的金融环境,迫切需要一种有效的方法来自动检测和识别以太坊系统治理中的以太坊欺诈行为。鉴于此,本文提出了一种通过挖掘基于以太坊的交易记录来检测以太坊欺诈的方法。具体来说,使用网络爬虫捕获标记的欺诈地址,然后基于公共交易账簿重构交易网络。然后,提出了一种基于数量的网络嵌入算法来提取节点特征,用于识别欺诈交易。最后,利用图卷积网络模型将地址分为合法地址和欺诈地址。实验结果表明,该系统检测欺诈性交易的准确率可达到95%,体现了该系统检测以太坊欺诈性交易的优异性能。
{"title":"Graph Neural Network for Ethereum Fraud Detection","authors":"Runnan Tan, Qingfeng Tan, Peng Zhang, Zhao Li","doi":"10.1109/ICKG52313.2021.00020","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00020","url":null,"abstract":"Currently, the blockchain technology has been widely applied to various industries, and has attracted wide attention. However, because of its unique anonymity, digital currency has become a haven for all kinds of cyber crimes. It has been reported that Ethereum frauds provide huge profits, and pose a serious threat to the financial security of the Ethereum network. To create a desired financial environment, an effective method is urgently needed to automatically detect and identify Ethereum frauds in the governance of the Ethereum system. In view of this, this paper proposes a method for detecting Ethereum frauds by mining Ethereum-based transaction records. Specifically, web crawlers are used to capture labeled fraudulent addresses, and then a transaction network is reconstructed based on the public transaction book. Then, an amount-based network embedding algorithm is proposed to extract node features for identifying fraudulent transactions. At last, the graph convolutional network model is used to classify addresses into legal addresses and fraudulent addresses. The experimental results show that the system for detecting fraudulent transactions can achieve the accuracy of 95%, which reflects the excellent performance of the system for detecting Ethereum fraudulent transactions.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126407078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Aggregation Enhanced Graph Convolutional Network for Graph Classification 基于聚合增强图卷积网络的图分类
Pub Date : 2021-12-01 DOI: 10.1109/ICKG52313.2021.00018
Guixian Zhang, Boyan Chen, Lijuan Wu, Kui Zhang, Shichao Zhang
Limited by the irregularity of graph topological structure and the sequence independence of nodes, existing graph neural networks usually generate graph-level representation by simple aggregation or sorting of node features for graph classification. These models are usually not deep enough to extract more abstract semantic information. Once the network deepens, it is easy to cause over-smoothing. To solve this problem, we propose a simple and efficient graph convolutional neural network based on DenseNet called AEGCN (Aggregation Enhanced Graph Convolutional Network) for graph clas-sification. We build a local extrema function named ELEConv (Enhanced Local Extrema Convolution) to reduce the noise in graphs, and then generate a large number of reusable feature maps through dense links. Extensive experiments on four real-world datasets validate that AEGCN not only alleviates the over-smoothing problem, but also has an advanced graph classification effect.
现有的图神经网络受图拓扑结构不规则性和节点序列独立性的限制,通常通过对节点特征进行简单的聚合或排序来生成图级表示,用于图分类。这些模型通常不够深入,无法提取更抽象的语义信息。一旦网络加深,很容易造成过度平滑。为了解决这一问题,我们提出了一种基于DenseNet的简单高效的图卷积神经网络,称为AEGCN (Aggregation Enhanced graph convolutional network),用于图分类。我们构建了一个局部极值函数ELEConv (Enhanced local extrema Convolution,增强局部极值卷积)来降低图中的噪声,然后通过密集链接生成大量可重用的特征图。在4个真实数据集上的大量实验验证了AEGCN不仅缓解了过度平滑问题,而且具有先进的图分类效果。
{"title":"Aggregation Enhanced Graph Convolutional Network for Graph Classification","authors":"Guixian Zhang, Boyan Chen, Lijuan Wu, Kui Zhang, Shichao Zhang","doi":"10.1109/ICKG52313.2021.00018","DOIUrl":"https://doi.org/10.1109/ICKG52313.2021.00018","url":null,"abstract":"Limited by the irregularity of graph topological structure and the sequence independence of nodes, existing graph neural networks usually generate graph-level representation by simple aggregation or sorting of node features for graph classification. These models are usually not deep enough to extract more abstract semantic information. Once the network deepens, it is easy to cause over-smoothing. To solve this problem, we propose a simple and efficient graph convolutional neural network based on DenseNet called AEGCN (Aggregation Enhanced Graph Convolutional Network) for graph clas-sification. We build a local extrema function named ELEConv (Enhanced Local Extrema Convolution) to reduce the noise in graphs, and then generate a large number of reusable feature maps through dense links. Extensive experiments on four real-world datasets validate that AEGCN not only alleviates the over-smoothing problem, but also has an advanced graph classification effect.","PeriodicalId":174126,"journal":{"name":"2021 IEEE International Conference on Big Knowledge (ICBK)","volume":"60 5","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132399277","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2021 IEEE International Conference on Big Knowledge (ICBK)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1