首页 > 最新文献

Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)最新文献

英文 中文
Train Your Own GNN Teacher: Graph-Aware Distillation on Textual Graphs 训练你自己的GNN老师:文本图的图形感知蒸馏
Costas Mavromatis, V. N. Ioannidis, Shen Wang, Da Zheng, Soji Adeshina, Jun Ma, Han Zhao, C. Faloutsos, G. Karypis
How can we learn effective node representations on textual graphs? Graph Neural Networks (GNNs) that use Language Models (LMs) to encode textual information of graphs achieve state-of-the-art performance in many node classification tasks. Yet, combining GNNs with LMs has not been widely explored for practical deployments due to its scalability issues. In this work, we tackle this challenge by developing a Graph-Aware Distillation framework (GRAD) to encode graph structures into an LM for graph-free, fast inference. Different from conventional knowledge distillation, GRAD jointly optimizes a GNN teacher and a graph-free student over the graph's nodes via a shared LM. This encourages the graph-free student to exploit graph information encoded by the GNN teacher while at the same time, enables the GNN teacher to better leverage textual information from unlabeled nodes. As a result, the teacher and the student models learn from each other to improve their overall performance. Experiments in eight node classification benchmarks in both transductive and inductive settings showcase GRAD's superiority over existing distillation approaches for textual graphs.
我们如何在文本图上学习有效的节点表示?使用语言模型(LMs)对图的文本信息进行编码的图神经网络(gnn)在许多节点分类任务中获得了最先进的性能。然而,由于其可扩展性问题,gnn与LMs的结合尚未被广泛探索用于实际部署。在这项工作中,我们通过开发一个图感知蒸馏框架(GRAD)来解决这一挑战,将图结构编码到LM中,以进行无图的快速推理。与传统的知识蒸馏不同,GRAD通过共享LM在图的节点上共同优化GNN教师和无图学生。这鼓励无图的学生利用GNN教师编码的图信息,同时使GNN教师能够更好地利用来自未标记节点的文本信息。因此,教师和学生模型相互学习,以提高他们的整体表现。在转导和归纳设置下的八个节点分类基准实验表明,GRAD优于现有的文本图蒸馏方法。
{"title":"Train Your Own GNN Teacher: Graph-Aware Distillation on Textual Graphs","authors":"Costas Mavromatis, V. N. Ioannidis, Shen Wang, Da Zheng, Soji Adeshina, Jun Ma, Han Zhao, C. Faloutsos, G. Karypis","doi":"10.48550/arXiv.2304.10668","DOIUrl":"https://doi.org/10.48550/arXiv.2304.10668","url":null,"abstract":"How can we learn effective node representations on textual graphs? Graph Neural Networks (GNNs) that use Language Models (LMs) to encode textual information of graphs achieve state-of-the-art performance in many node classification tasks. Yet, combining GNNs with LMs has not been widely explored for practical deployments due to its scalability issues. In this work, we tackle this challenge by developing a Graph-Aware Distillation framework (GRAD) to encode graph structures into an LM for graph-free, fast inference. Different from conventional knowledge distillation, GRAD jointly optimizes a GNN teacher and a graph-free student over the graph's nodes via a shared LM. This encourages the graph-free student to exploit graph information encoded by the GNN teacher while at the same time, enables the GNN teacher to better leverage textual information from unlabeled nodes. As a result, the teacher and the student models learn from each other to improve their overall performance. Experiments in eight node classification benchmarks in both transductive and inductive settings showcase GRAD's superiority over existing distillation approaches for textual graphs.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-04-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79203997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Two-stage Denoising Diffusion Model for Source Localization in Graph Inverse Problems 图反问题源定位的两阶段去噪扩散模型
Bosong Huang, Weihao Yu, Ruzhong Xie, Jing Xiao, Jin Huang
Source localization is the inverse problem of graph information dissemination and has broad practical applications. However, the inherent intricacy and uncertainty in information dissemination pose significant challenges, and the ill-posed nature of the source localization problem further exacerbates these challenges. Recently, deep generative models, particularly diffusion models inspired by classical non-equilibrium thermodynamics, have made significant progress. While diffusion models have proven to be powerful in solving inverse problems and producing high-quality reconstructions, applying them directly to the source localization is infeasible for two reasons. Firstly, it is impossible to calculate the posterior disseminated results on a large-scale network for iterative denoising sampling, which would incur enormous computational costs. Secondly, in the existing methods for this field, the training data itself are ill-posed (many-to-one); thus simply transferring the diffusion model would only lead to local optima. To address these challenges, we propose a two-stage optimization framework, the source localization denoising diffusion model (SL-Diff). In the coarse stage, we devise the source proximity degrees as the supervised signals to generate coarse-grained source predictions. This aims to efficiently initialize the next stage, significantly reducing its convergence time and calibrating the convergence process. Furthermore, the introduction of cascade temporal information in this training method transforms the many-to-one mapping relationship into a one-to-one relationship, perfectly addressing the ill-posed problem. In the fine stage, we design a diffusion model for the graph inverse problem that can quantify the uncertainty in the dissemination. The proposed SL-Diff yields excellent prediction results within a reasonable sampling time at extensive experiments.
源定位是图信息传播的逆问题,具有广泛的实际应用。然而,信息传播固有的复杂性和不确定性带来了重大挑战,而源定位问题的病态性进一步加剧了这些挑战。近年来,深度生成模型,特别是受经典非平衡热力学启发的扩散模型取得了重大进展。虽然扩散模型在求解逆问题和产生高质量重建方面已经被证明是强大的,但由于两个原因,将它们直接应用于源定位是不可行的。首先,不可能在大规模网络上计算后验传播结果进行迭代去噪采样,这将产生巨大的计算成本。其次,在该领域的现有方法中,训练数据本身是病态的(多对一);因此,简单地转移扩散模型只会导致局部最优。为了解决这些挑战,我们提出了一个两阶段优化框架,即源定位去噪扩散模型(SL-Diff)。在粗粒度阶段,我们将源接近度设计为监督信号,以生成粗粒度的源预测。这样做的目的是有效地初始化下一阶段,显著缩短其收敛时间并校准收敛过程。此外,在该训练方法中引入级联时间信息,将多对一映射关系转化为一对一映射关系,很好地解决了不适定问题。在精细阶段,我们针对图逆问题设计了一个扩散模型,可以量化传播中的不确定性。在大量的实验中,所提出的SL-Diff在合理的采样时间内产生了良好的预测结果。
{"title":"Two-stage Denoising Diffusion Model for Source Localization in Graph Inverse Problems","authors":"Bosong Huang, Weihao Yu, Ruzhong Xie, Jing Xiao, Jin Huang","doi":"10.48550/arXiv.2304.08841","DOIUrl":"https://doi.org/10.48550/arXiv.2304.08841","url":null,"abstract":"Source localization is the inverse problem of graph information dissemination and has broad practical applications. However, the inherent intricacy and uncertainty in information dissemination pose significant challenges, and the ill-posed nature of the source localization problem further exacerbates these challenges. Recently, deep generative models, particularly diffusion models inspired by classical non-equilibrium thermodynamics, have made significant progress. While diffusion models have proven to be powerful in solving inverse problems and producing high-quality reconstructions, applying them directly to the source localization is infeasible for two reasons. Firstly, it is impossible to calculate the posterior disseminated results on a large-scale network for iterative denoising sampling, which would incur enormous computational costs. Secondly, in the existing methods for this field, the training data itself are ill-posed (many-to-one); thus simply transferring the diffusion model would only lead to local optima. To address these challenges, we propose a two-stage optimization framework, the source localization denoising diffusion model (SL-Diff). In the coarse stage, we devise the source proximity degrees as the supervised signals to generate coarse-grained source predictions. This aims to efficiently initialize the next stage, significantly reducing its convergence time and calibrating the convergence process. Furthermore, the introduction of cascade temporal information in this training method transforms the many-to-one mapping relationship into a one-to-one relationship, perfectly addressing the ill-posed problem. In the fine stage, we design a diffusion model for the graph inverse problem that can quantify the uncertainty in the dissemination. The proposed SL-Diff yields excellent prediction results within a reasonable sampling time at extensive experiments.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83696207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving Autoregressive NLP Tasks via Modular Linearized Attention 模块化线性化注意力改进自回归NLP任务
Victor Agostinelli, Lizhong Chen
Various natural language processing (NLP) tasks necessitate models that are efficient and small based on their ultimate application at the edge or in other resource-constrained environments. While prior research has reduced the size of these models, increasing computational efficiency without considerable performance impacts remains difficult, especially for autoregressive tasks. This paper proposes modular linearized attention (MLA), which combines multiple efficient attention mechanisms, including cosFormer, to maximize inference quality while achieving notable speedups. We validate this approach on several autoregressive NLP tasks, including speech-to-text neural machine translation (S2T NMT), speech-to-text simultaneous translation (SimulST), and autoregressive text-to-spectrogram, noting efficiency gains on TTS and competitive performance for NMT and SimulST during training and inference.
各种自然语言处理(NLP)任务需要基于其在边缘或其他资源受限环境中的最终应用的高效和小型模型。虽然先前的研究已经减小了这些模型的大小,但在不显著影响性能的情况下提高计算效率仍然很困难,特别是对于自回归任务。本文提出了模块化线性化注意(MLA),它结合了多种有效的注意机制,包括cosFormer,以最大限度地提高推理质量,同时获得显著的速度。我们在几个自回归NLP任务上验证了这种方法,包括语音到文本的神经机器翻译(S2T NMT)、语音到文本的同声翻译(SimulST)和自回归文本到频谱图,注意到TTS的效率提高以及NMT和SimulST在训练和推理期间的竞争性能。
{"title":"Improving Autoregressive NLP Tasks via Modular Linearized Attention","authors":"Victor Agostinelli, Lizhong Chen","doi":"10.48550/arXiv.2304.08453","DOIUrl":"https://doi.org/10.48550/arXiv.2304.08453","url":null,"abstract":"Various natural language processing (NLP) tasks necessitate models that are efficient and small based on their ultimate application at the edge or in other resource-constrained environments. While prior research has reduced the size of these models, increasing computational efficiency without considerable performance impacts remains difficult, especially for autoregressive tasks. This paper proposes modular linearized attention (MLA), which combines multiple efficient attention mechanisms, including cosFormer, to maximize inference quality while achieving notable speedups. We validate this approach on several autoregressive NLP tasks, including speech-to-text neural machine translation (S2T NMT), speech-to-text simultaneous translation (SimulST), and autoregressive text-to-spectrogram, noting efficiency gains on TTS and competitive performance for NMT and SimulST during training and inference.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88071814","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Explainable Relational Reinforcement Learning: A Neuro-Symbolic Approach 深度可解释关系强化学习:一种神经符号方法
Rishi Hazra, L. D. Raedt
Despite numerous successes in Deep Reinforcement Learning (DRL), the learned policies are not interpretable. Moreover, since DRL does not exploit symbolic relational representations, it has difficulties in coping with structural changes in its environment (such as increasing the number of objects). Relational Reinforcement Learning, on the other hand, inherits the relational representations from symbolic planning to learn reusable policies. However, it has so far been unable to scale up and exploit the power of deep neural networks. We propose Deep Explainable Relational Reinforcement Learning (DERRL), a framework that exploits the best of both -- neural and symbolic worlds. By resorting to a neuro-symbolic approach, DERRL combines relational representations and constraints from symbolic planning with deep learning to extract interpretable policies. These policies are in the form of logical rules that explain how each decision (or action) is arrived at. Through several experiments, in setups like the Countdown Game, Blocks World, Gridworld, and Traffic, we show that the policies learned by DERRL can be applied to different configurations and contexts, hence generalizing to environmental modifications.
尽管深度强化学习(DRL)取得了许多成功,但学习到的策略是不可解释的。此外,由于DRL不利用符号关系表示,它在处理其环境中的结构变化(例如增加对象数量)方面存在困难。另一方面,关系强化学习从符号规划中继承关系表示来学习可重用策略。然而,到目前为止,它还无法扩大和利用深度神经网络的力量。我们提出深度可解释关系强化学习(DERRL),这是一个利用神经和符号世界最好的框架。通过采用神经符号方法,DERRL将符号规划中的关系表示和约束与深度学习相结合,以提取可解释的策略。这些策略以逻辑规则的形式出现,解释每个决策(或行动)是如何达成的。通过几个实验,比如倒计时游戏,方块世界,网格世界和交通,我们表明DERRL学习的策略可以应用于不同的配置和上下文,从而推广到环境修改。
{"title":"Deep Explainable Relational Reinforcement Learning: A Neuro-Symbolic Approach","authors":"Rishi Hazra, L. D. Raedt","doi":"10.48550/arXiv.2304.08349","DOIUrl":"https://doi.org/10.48550/arXiv.2304.08349","url":null,"abstract":"Despite numerous successes in Deep Reinforcement Learning (DRL), the learned policies are not interpretable. Moreover, since DRL does not exploit symbolic relational representations, it has difficulties in coping with structural changes in its environment (such as increasing the number of objects). Relational Reinforcement Learning, on the other hand, inherits the relational representations from symbolic planning to learn reusable policies. However, it has so far been unable to scale up and exploit the power of deep neural networks. We propose Deep Explainable Relational Reinforcement Learning (DERRL), a framework that exploits the best of both -- neural and symbolic worlds. By resorting to a neuro-symbolic approach, DERRL combines relational representations and constraints from symbolic planning with deep learning to extract interpretable policies. These policies are in the form of logical rules that explain how each decision (or action) is arrived at. Through several experiments, in setups like the Countdown Game, Blocks World, Gridworld, and Traffic, we show that the policies learned by DERRL can be applied to different configurations and contexts, hence generalizing to environmental modifications.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86261815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Click-aware Structure Transfer with Sample Weight Assignment for Post-Click Conversion Rate Estimation 点击感知结构转移与样本权重分配的点击后转化率估计
Kai Ouyang, Wenhao Zheng, Chen Tang, Xuanji Xiao, Haitao Zheng
Post-click Conversion Rate (CVR) prediction task plays an essential role in industrial applications, such as recommendation and advertising. Conventional CVR methods typically suffer from the data sparsity problem as they rely only on samples where the user has clicked. To address this problem, researchers have introduced the method of multi-task learning, which utilizes non-clicked samples and shares feature representations of the Click-Through Rate (CTR) task with the CVR task. However, it should be noted that the CVR and CTR tasks are fundamentally different and may even be contradictory. Therefore, introducing a large amount of CTR information without distinction may drown out valuable information related to CVR. This phenomenon is called the curse of knowledge problem in this paper. To tackle this issue, we argue that a trade-off should be achieved between the introduction of large amounts of auxiliary information and the protection of valuable information related to CVR. Hence, we propose a Click-aware Structure Transfer model with sample Weight Assignment, abbreviated as CSTWA. It pays more attention to the latent structure information, which can filter the input information that is related to CVR, instead of directly sharing feature representations. Meanwhile, to capture the representation conflict between CTR and CVR, we calibrate the representation layer and reweight the discriminant layer to excavate the click bias information from the CTR tower. Moreover, it incorporates a sample weight assignment algorithm biased towards CVR modeling, to make the knowledge from CTR would not mislead the CVR. Extensive experiments on industrial and public datasets have demonstrated that CSTWA significantly outperforms widely used and competitive models.
点击后转化率(CVR)预测任务在推荐和广告等行业应用中发挥着至关重要的作用。传统的CVR方法通常存在数据稀疏性问题,因为它们只依赖于用户点击的样本。为了解决这个问题,研究人员引入了多任务学习方法,该方法利用非点击样本,并与CVR任务共享点击率(CTR)任务的特征表示。然而,需要注意的是,CVR和CTR的任务是根本不同的,甚至可能是矛盾的。因此,不加区分地引入大量的CTR信息可能会淹没与CVR相关的有价值的信息。本文将这种现象称为知识诅咒问题。为了解决这一问题,我们认为应该在引入大量辅助信息和保护与CVR相关的有价值信息之间实现权衡。因此,我们提出了一个带有样本权重分配的点击感知结构转移模型,简称CSTWA。它更注重潜在的结构信息,可以过滤与CVR相关的输入信息,而不是直接共享特征表示。同时,为了捕获CTR和CVR之间的表示冲突,我们对表示层进行了校准,并对判别层进行了重新加权,从而从CTR塔中挖掘出点击偏差信息。此外,该算法还引入了偏向于CVR建模的样本权重分配算法,以使来自CTR的知识不会误导CVR。在工业和公共数据集上进行的大量实验表明,CSTWA显著优于广泛使用的竞争模型。
{"title":"Click-aware Structure Transfer with Sample Weight Assignment for Post-Click Conversion Rate Estimation","authors":"Kai Ouyang, Wenhao Zheng, Chen Tang, Xuanji Xiao, Haitao Zheng","doi":"10.48550/arXiv.2304.01169","DOIUrl":"https://doi.org/10.48550/arXiv.2304.01169","url":null,"abstract":"Post-click Conversion Rate (CVR) prediction task plays an essential role in industrial applications, such as recommendation and advertising. Conventional CVR methods typically suffer from the data sparsity problem as they rely only on samples where the user has clicked. To address this problem, researchers have introduced the method of multi-task learning, which utilizes non-clicked samples and shares feature representations of the Click-Through Rate (CTR) task with the CVR task. However, it should be noted that the CVR and CTR tasks are fundamentally different and may even be contradictory. Therefore, introducing a large amount of CTR information without distinction may drown out valuable information related to CVR. This phenomenon is called the curse of knowledge problem in this paper. To tackle this issue, we argue that a trade-off should be achieved between the introduction of large amounts of auxiliary information and the protection of valuable information related to CVR. Hence, we propose a Click-aware Structure Transfer model with sample Weight Assignment, abbreviated as CSTWA. It pays more attention to the latent structure information, which can filter the input information that is related to CVR, instead of directly sharing feature representations. Meanwhile, to capture the representation conflict between CTR and CVR, we calibrate the representation layer and reweight the discriminant layer to excavate the click bias information from the CTR tower. Moreover, it incorporates a sample weight assignment algorithm biased towards CVR modeling, to make the knowledge from CTR would not mislead the CVR. Extensive experiments on industrial and public datasets have demonstrated that CSTWA significantly outperforms widely used and competitive models.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-04-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76983112","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Improving Few-Shot Inductive Learning on Temporal Knowledge Graphs using Confidence-Augmented Reinforcement Learning 基于置信度增强强化学习改进时态知识图的少量归纳学习
Zifeng Ding, Jingpei Wu, Zong-Xun Li, Yunpu Ma, Volker Tresp
Temporal knowledge graph completion (TKGC) aims to predict the missing links among the entities in a temporal knwoledge graph (TKG). Most previous TKGC methods only consider predicting the missing links among the entities seen in the training set, while they are unable to achieve great performance in link prediction concerning newly-emerged unseen entities. Recently, a new task, i.e., TKG few-shot out-of-graph (OOG) link prediction, is proposed, where TKGC models are required to achieve great link prediction performance concerning newly-emerged entities that only have few-shot observed examples. In this work, we propose a TKGC method FITCARL that combines few-shot learning with reinforcement learning to solve this task. In FITCARL, an agent traverses through the whole TKG to search for the prediction answer. A policy network is designed to guide the search process based on the traversed path. To better address the data scarcity problem in the few-shot setting, we introduce a module that computes the confidence of each candidate action and integrate it into the policy for action selection. We also exploit the entity concept information with a novel concept regularizer to boost model performance. Experimental results show that FITCARL achieves stat-of-the-art performance on TKG few-shot OOG link prediction.
时间知识图补全(TKGC)的目的是预测时间知识图中实体之间缺失的链接。大多数以前的TKGC方法只考虑预测训练集中看到的实体之间的缺失链接,而对于新出现的未见实体的链接预测无法取得很好的性能。最近,提出了一种新的任务,即TKG少镜头图外(OOG)链接预测,该任务要求TKGC模型对只有很少镜头观察样例的新出现实体具有很高的链接预测性能。在这项工作中,我们提出了一种结合了few-shot学习和强化学习的TKGC方法FITCARL来解决这个问题。在FITCARL中,agent遍历整个TKG来寻找预测答案。策略网络的设计是基于所遍历的路径来引导搜索过程。为了更好地解决少镜头设置中的数据稀缺性问题,我们引入了一个模块来计算每个候选动作的置信度,并将其集成到动作选择的策略中。我们还利用实体概念信息和一个新的概念正则化器来提高模型的性能。实验结果表明,FITCARL在TKG少射OOG链路预测上达到了最先进的性能。
{"title":"Improving Few-Shot Inductive Learning on Temporal Knowledge Graphs using Confidence-Augmented Reinforcement Learning","authors":"Zifeng Ding, Jingpei Wu, Zong-Xun Li, Yunpu Ma, Volker Tresp","doi":"10.48550/arXiv.2304.00613","DOIUrl":"https://doi.org/10.48550/arXiv.2304.00613","url":null,"abstract":"Temporal knowledge graph completion (TKGC) aims to predict the missing links among the entities in a temporal knwoledge graph (TKG). Most previous TKGC methods only consider predicting the missing links among the entities seen in the training set, while they are unable to achieve great performance in link prediction concerning newly-emerged unseen entities. Recently, a new task, i.e., TKG few-shot out-of-graph (OOG) link prediction, is proposed, where TKGC models are required to achieve great link prediction performance concerning newly-emerged entities that only have few-shot observed examples. In this work, we propose a TKGC method FITCARL that combines few-shot learning with reinforcement learning to solve this task. In FITCARL, an agent traverses through the whole TKG to search for the prediction answer. A policy network is designed to guide the search process based on the traversed path. To better address the data scarcity problem in the few-shot setting, we introduce a module that computes the confidence of each candidate action and integrate it into the policy for action selection. We also exploit the entity concept information with a novel concept regularizer to boost model performance. Experimental results show that FITCARL achieves stat-of-the-art performance on TKG few-shot OOG link prediction.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73925491","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
marl-jax: Multi-agent Reinforcement Leaning framework for Social Generalization 社会泛化的多智能体强化学习框架
K. Mehta, Anuj Mahajan, Priyesh Kumar
Recent advances in Reinforcement Learning (RL) have led to many exciting applications. These advancements have been driven by improvements in both algorithms and engineering, which have resulted in faster training of RL agents. We present marl-jax, a multi-agent reinforcement learning software package for training and evaluating social generalization of the agents. The package is designed for training a population of agents in multi-agent environments and evaluating their ability to generalize to diverse background agents. It is built on top of DeepMind's JAX ecosystem~cite{deepmind2020jax} and leverages the RL ecosystem developed by DeepMind. Our framework marl-jax is capable of working in cooperative and competitive, simultaneous-acting environments with multiple agents. The package offers an intuitive and user-friendly command-line interface for training a population and evaluating its generalization capabilities. In conclusion, marl-jax provides a valuable resource for researchers interested in exploring social generalization in the context of MARL. The open-source code for marl-jax is available at: href{https://github.com/kinalmehta/marl-jax}{https://github.com/kinalmehta/marl-jax}
强化学习(RL)的最新进展导致了许多令人兴奋的应用。这些进步是由算法和工程的改进推动的,这导致了强化学习代理的更快训练。我们提出了marl-jax,一个多智能体强化学习软件包,用于训练和评估智能体的社会泛化。该包旨在训练多智能体环境中的智能体群体,并评估它们泛化到不同背景智能体的能力。它建立在DeepMind的JAX生态系统cite{deepmind2020jax}之上,并利用了DeepMind开发的强化学习生态系统。我们的框架marjax能够在具有多个代理的合作、竞争、同时作用的环境中工作。该包提供了一个直观和用户友好的命令行界面,用于训练种群和评估其泛化能力。总之,MARL -jax为有兴趣在MARL背景下探索社会概括的研究人员提供了宝贵的资源。marl-jax的开源代码可在: href{https://github.com/kinalmehta/marl-jax}{https://github.com/kinalmehta/marl-jax}
{"title":"marl-jax: Multi-agent Reinforcement Leaning framework for Social Generalization","authors":"K. Mehta, Anuj Mahajan, Priyesh Kumar","doi":"10.48550/arXiv.2303.13808","DOIUrl":"https://doi.org/10.48550/arXiv.2303.13808","url":null,"abstract":"Recent advances in Reinforcement Learning (RL) have led to many exciting applications. These advancements have been driven by improvements in both algorithms and engineering, which have resulted in faster training of RL agents. We present marl-jax, a multi-agent reinforcement learning software package for training and evaluating social generalization of the agents. The package is designed for training a population of agents in multi-agent environments and evaluating their ability to generalize to diverse background agents. It is built on top of DeepMind's JAX ecosystem~cite{deepmind2020jax} and leverages the RL ecosystem developed by DeepMind. Our framework marl-jax is capable of working in cooperative and competitive, simultaneous-acting environments with multiple agents. The package offers an intuitive and user-friendly command-line interface for training a population and evaluating its generalization capabilities. In conclusion, marl-jax provides a valuable resource for researchers interested in exploring social generalization in the context of MARL. The open-source code for marl-jax is available at: href{https://github.com/kinalmehta/marl-jax}{https://github.com/kinalmehta/marl-jax}","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90815754","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Improved Regret Bounds for Online Kernel Selection under Bandit Feedback 改进了强盗反馈下在线内核选择的遗憾边界
Junfan Li, Shizhong Liao
In this paper, we improve the regret bound for online kernel selection under bandit feedback. Previous algorithm enjoys a $O((Vert fVert^2_{mathcal{H}_i}+1)K^{frac{1}{3}}T^{frac{2}{3}})$ expected bound for Lipschitz loss functions. We prove two types of regret bounds improving the previous bound. For smooth loss functions, we propose an algorithm with a $O(U^{frac{2}{3}}K^{-frac{1}{3}}(sum^K_{i=1}L_T(f^ast_i))^{frac{2}{3}})$ expected bound where $L_T(f^ast_i)$ is the cumulative losses of optimal hypothesis in $mathbb{H}_{i}={finmathcal{H}_i:Vert fVert_{mathcal{H}_i}leq U}$. The data-dependent bound keeps the previous worst-case bound and is smaller if most of candidate kernels match well with the data. For Lipschitz loss functions, we propose an algorithm with a $O(Usqrt{KT}ln^{frac{2}{3}}{T})$ expected bound asymptotically improving the previous bound. We apply the two algorithms to online kernel selection with time constraint and prove new regret bounds matching or improving the previous $O(sqrt{Tln{K}} +Vert fVert^2_{mathcal{H}_i}max{sqrt{T},frac{T}{sqrt{mathcal{R}}}})$ expected bound where $mathcal{R}$ is the time budget. Finally, we empirically verify our algorithms on online regression and classification tasks.
本文改进了强盗反馈下在线核选择的遗憾界。先前的算法对Lipschitz损失函数具有$O((Vert fVert^2_{mathcal{H}_i}+1)K^{frac{1}{3}}T^{frac{2}{3}})$期望界。我们证明了两种改进前一界的后悔界。对于平滑损失函数,我们提出了一个具有$O(U^{frac{2}{3}}K^{-frac{1}{3}}(sum^K_{i=1}L_T(f^ast_i))^{frac{2}{3}})$期望界的算法,其中$L_T(f^ast_i)$为$mathbb{H}_{i}={finmathcal{H}_i:Vert fVert_{mathcal{H}_i}leq U}$中最优假设的累积损失。与数据相关的边界保持之前的最坏情况边界,如果大多数候选核与数据匹配得很好,则边界更小。对于Lipschitz损失函数,我们提出了一种具有$O(Usqrt{KT}ln^{frac{2}{3}}{T})$期望界的算法,该算法渐近地改进了之前的期望界。我们将这两种算法应用于有时间约束的在线核选择,并证明了新的遗憾边界匹配或改进了先前的$O(sqrt{Tln{K}} +Vert fVert^2_{mathcal{H}_i}max{sqrt{T},frac{T}{sqrt{mathcal{R}}}})$期望边界,其中$mathcal{R}$为时间预算。最后,我们在在线回归和分类任务上验证了我们的算法。
{"title":"Improved Regret Bounds for Online Kernel Selection under Bandit Feedback","authors":"Junfan Li, Shizhong Liao","doi":"10.48550/arXiv.2303.05018","DOIUrl":"https://doi.org/10.48550/arXiv.2303.05018","url":null,"abstract":"In this paper, we improve the regret bound for online kernel selection under bandit feedback. Previous algorithm enjoys a $O((Vert fVert^2_{mathcal{H}_i}+1)K^{frac{1}{3}}T^{frac{2}{3}})$ expected bound for Lipschitz loss functions. We prove two types of regret bounds improving the previous bound. For smooth loss functions, we propose an algorithm with a $O(U^{frac{2}{3}}K^{-frac{1}{3}}(sum^K_{i=1}L_T(f^ast_i))^{frac{2}{3}})$ expected bound where $L_T(f^ast_i)$ is the cumulative losses of optimal hypothesis in $mathbb{H}_{i}={finmathcal{H}_i:Vert fVert_{mathcal{H}_i}leq U}$. The data-dependent bound keeps the previous worst-case bound and is smaller if most of candidate kernels match well with the data. For Lipschitz loss functions, we propose an algorithm with a $O(Usqrt{KT}ln^{frac{2}{3}}{T})$ expected bound asymptotically improving the previous bound. We apply the two algorithms to online kernel selection with time constraint and prove new regret bounds matching or improving the previous $O(sqrt{Tln{K}} +Vert fVert^2_{mathcal{H}_i}max{sqrt{T},frac{T}{sqrt{mathcal{R}}}})$ expected bound where $mathcal{R}$ is the time budget. Finally, we empirically verify our algorithms on online regression and classification tasks.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91255541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Imbalanced Time-series Forecasting via Local Discrepancy Density 基于局部差异密度的深度不平衡时间序列预测
Junwoo Park, Jungsoo Lee, Youngin Cho, W. Shin, Dongmin Kim, J. Choo, E. Choi
Time-series forecasting models often encounter abrupt changes in a given period of time which generally occur due to unexpected or unknown events. Despite their scarce occurrences in the training set, abrupt changes incur loss that significantly contributes to the total loss. Therefore, they act as noisy training samples and prevent the model from learning generalizable patterns, namely the normal states. Based on our findings, we propose a reweighting framework that down-weights the losses incurred by abrupt changes and up-weights those by normal states. For the reweighting framework, we first define a measurement termed Local Discrepancy (LD) which measures the degree of abruptness of a change in a given period of time. Since a training set is mostly composed of normal states, we then consider how frequently the temporal changes appear in the training set based on LD. Our reweighting framework is applicable to existing time-series forecasting models regardless of the architectures. Through extensive experiments on 12 time-series forecasting models over eight datasets with various in-output sequence lengths, we demonstrate that applying our reweighting framework reduces MSE by 10.1% on average and by up to 18.6% in the state-of-the-art model.
时间序列预测模型经常遇到在给定时间段内的突然变化,这种变化通常是由于意外或未知事件引起的。尽管它们在训练集中很少出现,但突变会导致损失,这对总损失有很大的贡献。因此,它们作为有噪声的训练样本,并阻止模型学习可推广的模式,即正常状态。基于我们的研究结果,我们提出了一个重加权框架,该框架降低了突变引起的损失的权重,提高了正常状态引起的损失的权重。对于重新加权框架,我们首先定义了一个称为局部差异(LD)的度量,它测量给定时间段内变化的突然性程度。由于训练集主要由正常状态组成,我们然后考虑基于LD的训练集中时间变化出现的频率。我们的重加权框架适用于现有的时间序列预测模型,而不管结构如何。通过对8个不同输出序列长度的数据集上的12个时间序列预测模型进行广泛的实验,我们证明,应用我们的加权框架可以将MSE平均降低10.1%,在最先进的模型中可降低高达18.6%。
{"title":"Deep Imbalanced Time-series Forecasting via Local Discrepancy Density","authors":"Junwoo Park, Jungsoo Lee, Youngin Cho, W. Shin, Dongmin Kim, J. Choo, E. Choi","doi":"10.48550/arXiv.2302.13563","DOIUrl":"https://doi.org/10.48550/arXiv.2302.13563","url":null,"abstract":"Time-series forecasting models often encounter abrupt changes in a given period of time which generally occur due to unexpected or unknown events. Despite their scarce occurrences in the training set, abrupt changes incur loss that significantly contributes to the total loss. Therefore, they act as noisy training samples and prevent the model from learning generalizable patterns, namely the normal states. Based on our findings, we propose a reweighting framework that down-weights the losses incurred by abrupt changes and up-weights those by normal states. For the reweighting framework, we first define a measurement termed Local Discrepancy (LD) which measures the degree of abruptness of a change in a given period of time. Since a training set is mostly composed of normal states, we then consider how frequently the temporal changes appear in the training set based on LD. Our reweighting framework is applicable to existing time-series forecasting models regardless of the architectures. Through extensive experiments on 12 time-series forecasting models over eight datasets with various in-output sequence lengths, we demonstrate that applying our reweighting framework reduces MSE by 10.1% on average and by up to 18.6% in the state-of-the-art model.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78848676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
MiDi: Mixed Graph and 3D Denoising Diffusion for Molecule Generation MiDi:用于分子生成的混合图和三维去噪扩散
Clément Vignac, Nagham Osman, L. Toni, P. Frossard
This work introduces MiDi, a novel diffusion model for jointly generating molecular graphs and their corresponding 3D arrangement of atoms. Unlike existing methods that rely on predefined rules to determine molecular bonds based on the 3D conformation, MiDi offers an end-to-end differentiable approach that streamlines the molecule generation process. Our experimental results demonstrate the effectiveness of this approach. On the challenging GEOM-DRUGS dataset, MiDi generates 92% of stable molecules, against 6% for the previous EDM model that uses interatomic distances for bond prediction, and 40% using EDM followed by an algorithm that directly optimize bond orders for validity. Our code is available at github.com/cvignac/MiDi.
本文介绍了一种新的扩散模型MiDi,用于共同生成分子图及其相应的原子三维排列。现有的方法依赖于基于3D构象的预定义规则来确定分子键,MiDi提供了一种端到端可微分的方法,简化了分子生成过程。实验结果证明了该方法的有效性。在具有挑战性的geomo - drugs数据集上,MiDi生成了92%的稳定分子,而之前使用原子间距离进行键预测的EDM模型生成了6%的稳定分子,使用EDM后直接优化键顺序的算法生成了40%的稳定分子。我们的代码可在github.com/cvignac/MiDi上获得。
{"title":"MiDi: Mixed Graph and 3D Denoising Diffusion for Molecule Generation","authors":"Clément Vignac, Nagham Osman, L. Toni, P. Frossard","doi":"10.48550/arXiv.2302.09048","DOIUrl":"https://doi.org/10.48550/arXiv.2302.09048","url":null,"abstract":"This work introduces MiDi, a novel diffusion model for jointly generating molecular graphs and their corresponding 3D arrangement of atoms. Unlike existing methods that rely on predefined rules to determine molecular bonds based on the 3D conformation, MiDi offers an end-to-end differentiable approach that streamlines the molecule generation process. Our experimental results demonstrate the effectiveness of this approach. On the challenging GEOM-DRUGS dataset, MiDi generates 92% of stable molecules, against 6% for the previous EDM model that uses interatomic distances for bond prediction, and 40% using EDM followed by an algorithm that directly optimize bond orders for validity. Our code is available at github.com/cvignac/MiDi.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-02-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74653158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1