首页 > 最新文献

Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)最新文献

英文 中文
Powershap: A Power-full Shapley Feature Selection Method Powershap: Power-full Shapley特征选择方法
Jarne Verhaeghe, Jeroen Van Der Donckt, F. Ongenae, S. Hoecke
Feature selection is a crucial step in developing robust and powerful machine learning models. Feature selection techniques can be divided into two categories: filter and wrapper methods. While wrapper methods commonly result in strong predictive performances, they suffer from a large computational complexity and therefore take a significant amount of time to complete, especially when dealing with high-dimensional feature sets. Alternatively, filter methods are considerably faster, but suffer from several other disadvantages, such as (i) requiring a threshold value, (ii) not taking into account intercorrelation between features, and (iii) ignoring feature interactions with the model. To this end, we present powershap, a novel wrapper feature selection method, which leverages statistical hypothesis testing and power calculations in combination with Shapley values for quick and intuitive feature selection. Powershap is built on the core assumption that an informative feature will have a larger impact on the prediction compared to a known random feature. Benchmarks and simulations show that powershap outperforms other filter methods with predictive performances on par with wrapper methods while being significantly faster, often even reaching half or a third of the execution time. As such, powershap provides a competitive and quick algorithm that can be used by various models in different domains. Furthermore, powershap is implemented as a plug-and-play and open-source sklearn component, enabling easy integration in conventional data science pipelines. User experience is even further enhanced by also providing an automatic mode that automatically tunes the hyper-parameters of the powershap algorithm, allowing to use the algorithm without any configuration needed.
特征选择是开发健壮和强大的机器学习模型的关键步骤。特征选择技术可以分为两类:过滤方法和包装方法。虽然包装器方法通常会产生很强的预测性能,但它们的计算复杂度很高,因此需要花费大量的时间来完成,特别是在处理高维特征集时。另外,过滤方法要快得多,但也有其他缺点,比如(i)需要一个阈值,(ii)不考虑特征之间的相互关系,(iii)忽略特征与模型的相互作用。为此,我们提出了一种新的包装特征选择方法powershap,该方法利用统计假设检验和功率计算结合Shapley值进行快速直观的特征选择。Powershap建立在一个核心假设之上,即与已知的随机特征相比,信息特征对预测的影响更大。基准测试和模拟表明,powershap优于其他过滤器方法,其预测性能与包装器方法相当,同时速度快得多,通常甚至可以达到执行时间的一半或三分之一。因此,powershap提供了一种具有竞争力的快速算法,可用于不同领域的各种模型。此外,powershap是作为即插即用和开源的sklearn组件实现的,可以轻松集成到传统的数据科学管道中。通过提供自动模式,可以自动调整powershap算法的超参数,从而进一步增强用户体验,从而允许在不需要任何配置的情况下使用该算法。
{"title":"Powershap: A Power-full Shapley Feature Selection Method","authors":"Jarne Verhaeghe, Jeroen Van Der Donckt, F. Ongenae, S. Hoecke","doi":"10.48550/arXiv.2206.08394","DOIUrl":"https://doi.org/10.48550/arXiv.2206.08394","url":null,"abstract":"Feature selection is a crucial step in developing robust and powerful machine learning models. Feature selection techniques can be divided into two categories: filter and wrapper methods. While wrapper methods commonly result in strong predictive performances, they suffer from a large computational complexity and therefore take a significant amount of time to complete, especially when dealing with high-dimensional feature sets. Alternatively, filter methods are considerably faster, but suffer from several other disadvantages, such as (i) requiring a threshold value, (ii) not taking into account intercorrelation between features, and (iii) ignoring feature interactions with the model. To this end, we present powershap, a novel wrapper feature selection method, which leverages statistical hypothesis testing and power calculations in combination with Shapley values for quick and intuitive feature selection. Powershap is built on the core assumption that an informative feature will have a larger impact on the prediction compared to a known random feature. Benchmarks and simulations show that powershap outperforms other filter methods with predictive performances on par with wrapper methods while being significantly faster, often even reaching half or a third of the execution time. As such, powershap provides a competitive and quick algorithm that can be used by various models in different domains. Furthermore, powershap is implemented as a plug-and-play and open-source sklearn component, enabling easy integration in conventional data science pipelines. User experience is even further enhanced by also providing an automatic mode that automatically tunes the hyper-parameters of the powershap algorithm, allowing to use the algorithm without any configuration needed.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77563693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
ARES: Locally Adaptive Reconstruction-based Anomaly Scoring ARES:基于局部自适应重构的异常评分
Adam Goodge, Bryan Hooi, See-Kiong Ng, W. Ng
How can we detect anomalies: that is, samples that significantly differ from a given set of high-dimensional data, such as images or sensor data? This is a practical problem with numerous applications and is also relevant to the goal of making learning algorithms more robust to unexpected inputs. Autoencoders are a popular approach, partly due to their simplicity and their ability to perform dimension reduction. However, the anomaly scoring function is not adaptive to the natural variation in reconstruction error across the range of normal samples, which hinders their ability to detect real anomalies. In this paper, we empirically demonstrate the importance of local adaptivity for anomaly scoring in experiments with real data. We then propose our novel Adaptive Reconstruction Error-based Scoring approach, which adapts its scoring based on the local behaviour of reconstruction error over the latent space. We show that this improves anomaly detection performance over relevant baselines in a wide variety of benchmark datasets.
我们如何检测异常:即与一组给定的高维数据(如图像或传感器数据)显著不同的样本?这是许多应用中的一个实际问题,也与使学习算法对意外输入更健壮的目标有关。自编码器是一种流行的方法,部分原因是它们的简单性和执行降维的能力。然而,异常评分函数不能适应重构误差在正常样本范围内的自然变化,这阻碍了它们检测真实异常的能力。在本文中,我们用实际数据的实验证明了局部自适应对异常评分的重要性。然后,我们提出了一种新的基于自适应重构误差的评分方法,该方法基于潜在空间上重构误差的局部行为来调整其评分。我们表明,这提高了在各种基准数据集的相关基线上的异常检测性能。
{"title":"ARES: Locally Adaptive Reconstruction-based Anomaly Scoring","authors":"Adam Goodge, Bryan Hooi, See-Kiong Ng, W. Ng","doi":"10.48550/arXiv.2206.07604","DOIUrl":"https://doi.org/10.48550/arXiv.2206.07604","url":null,"abstract":"How can we detect anomalies: that is, samples that significantly differ from a given set of high-dimensional data, such as images or sensor data? This is a practical problem with numerous applications and is also relevant to the goal of making learning algorithms more robust to unexpected inputs. Autoencoders are a popular approach, partly due to their simplicity and their ability to perform dimension reduction. However, the anomaly scoring function is not adaptive to the natural variation in reconstruction error across the range of normal samples, which hinders their ability to detect real anomalies. In this paper, we empirically demonstrate the importance of local adaptivity for anomaly scoring in experiments with real data. We then propose our novel Adaptive Reconstruction Error-based Scoring approach, which adapts its scoring based on the local behaviour of reconstruction error over the latent space. We show that this improves anomaly detection performance over relevant baselines in a wide variety of benchmark datasets.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89213332","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Summarizing Labeled Multi-Graphs 标注多图总结
Dimitris Berberidis, P. Liang, L. Akoglu
Real-world graphs can be difficult to interpret and visualize beyond a certain size. To address this issue, graph summarization aims to simplify and shrink a graph, while maintaining its high-level structure and characteristics. Most summarization methods are designed for homogeneous, undirected, simple graphs; however, many real-world graphs are ornate; with characteristics including node labels, directed edges, edge multiplicities, and self-loops. In this paper we propose LM-Gsum, a versatile yet rigorous graph summarization model that (to the best of our knowledge, for the first time) can handle graphs with all the aforementioned characteristics (and any combination thereof). Moreover, our proposed model captures basic sub-structures that are prevalent in real-world graphs, such as cliques, stars, etc. LM-Gsum compactly quantifies the information content of a complex graph using a novel encoding scheme, where it seeks to minimize the total number of bits required to encode (i) the summary graph, as well as (ii) the corrections required for reconstructing the input graph losslessly. To accelerate the summary construction, it creates super-nodes efficiently by merging nodes in groups. Experiments demonstrate that LM-Gsum facilitates the visualization of real-world complex graphs, revealing interpretable structures and high- level relationships. Furthermore, LM-Gsum achieves better trade-off between compression rate and running time, relative to existing methods (only) on comparable settings.
现实世界的图形可能很难解释和可视化超过一定的大小。为了解决这个问题,图形摘要旨在简化和缩小图形,同时保持其高级结构和特征。大多数总结方法是为齐次的、无向的、简单的图设计的;然而,许多现实世界的图表都是华丽的;具有节点标签、有向边、边多重性和自环等特征。在本文中,我们提出了LM-Gsum,这是一个通用但严格的图摘要模型,(据我们所知,这是第一次)可以处理具有上述所有特征(以及它们的任何组合)的图。此外,我们提出的模型捕获了在现实世界图中普遍存在的基本子结构,如派系、星形等。LM-Gsum使用一种新颖的编码方案紧凑地量化了复杂图的信息内容,其中它寻求最小化编码(i)汇总图所需的总比特数,以及(ii)无损重建输入图所需的校正。该算法通过分组合并节点,高效地创建超级节点,加快了摘要的构建速度。实验表明,LM-Gsum有助于现实世界复杂图形的可视化,揭示可解释的结构和高层关系。此外,LM-Gsum在压缩率和运行时间之间实现了更好的权衡,相对于现有的方法(仅)在可比较的设置。
{"title":"Summarizing Labeled Multi-Graphs","authors":"Dimitris Berberidis, P. Liang, L. Akoglu","doi":"10.48550/arXiv.2206.07674","DOIUrl":"https://doi.org/10.48550/arXiv.2206.07674","url":null,"abstract":"Real-world graphs can be difficult to interpret and visualize beyond a certain size. To address this issue, graph summarization aims to simplify and shrink a graph, while maintaining its high-level structure and characteristics. Most summarization methods are designed for homogeneous, undirected, simple graphs; however, many real-world graphs are ornate; with characteristics including node labels, directed edges, edge multiplicities, and self-loops. In this paper we propose LM-Gsum, a versatile yet rigorous graph summarization model that (to the best of our knowledge, for the first time) can handle graphs with all the aforementioned characteristics (and any combination thereof). Moreover, our proposed model captures basic sub-structures that are prevalent in real-world graphs, such as cliques, stars, etc. LM-Gsum compactly quantifies the information content of a complex graph using a novel encoding scheme, where it seeks to minimize the total number of bits required to encode (i) the summary graph, as well as (ii) the corrections required for reconstructing the input graph losslessly. To accelerate the summary construction, it creates super-nodes efficiently by merging nodes in groups. Experiments demonstrate that LM-Gsum facilitates the visualization of real-world complex graphs, revealing interpretable structures and high- level relationships. Furthermore, LM-Gsum achieves better trade-off between compression rate and running time, relative to existing methods (only) on comparable settings.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87918419","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Defending Observation Attacks in Deep Reinforcement Learning via Detection and Denoising 利用检测和去噪防御深度强化学习中的观察攻击
Zikang Xiong, Joe Eappen, He Zhu, S. Jagannathan
Neural network policies trained using Deep Reinforcement Learning (DRL) are well-known to be susceptible to adversarial attacks. In this paper, we consider attacks manifesting as perturbations in the observation space managed by the external environment. These attacks have been shown to downgrade policy performance significantly. We focus our attention on well-trained deterministic and stochastic neural network policies in the context of continuous control benchmarks subject to four well-studied observation space adversarial attacks. To defend against these attacks, we propose a novel defense strategy using a detect-and-denoise schema. Unlike previous adversarial training approaches that sample data in adversarial scenarios, our solution does not require sampling data in an environment under attack, thereby greatly reducing risk during training. Detailed experimental results show that our technique is comparable with state-of-the-art adversarial training approaches.
众所周知,使用深度强化学习(DRL)训练的神经网络策略容易受到对抗性攻击。在本文中,我们考虑攻击表现为由外部环境管理的观测空间中的扰动。这些攻击已被证明会显著降低策略性能。我们将注意力集中在连续控制基准中训练有素的确定性和随机神经网络策略上,这些基准受到四种经过充分研究的观察空间对抗性攻击的影响。为了防御这些攻击,我们提出了一种使用检测-降噪模式的新防御策略。与之前在对抗场景中采样数据的对抗训练方法不同,我们的解决方案不需要在受到攻击的环境中采样数据,从而大大降低了训练过程中的风险。详细的实验结果表明,我们的技术可以与最先进的对抗性训练方法相媲美。
{"title":"Defending Observation Attacks in Deep Reinforcement Learning via Detection and Denoising","authors":"Zikang Xiong, Joe Eappen, He Zhu, S. Jagannathan","doi":"10.48550/arXiv.2206.07188","DOIUrl":"https://doi.org/10.48550/arXiv.2206.07188","url":null,"abstract":"Neural network policies trained using Deep Reinforcement Learning (DRL) are well-known to be susceptible to adversarial attacks. In this paper, we consider attacks manifesting as perturbations in the observation space managed by the external environment. These attacks have been shown to downgrade policy performance significantly. We focus our attention on well-trained deterministic and stochastic neural network policies in the context of continuous control benchmarks subject to four well-studied observation space adversarial attacks. To defend against these attacks, we propose a novel defense strategy using a detect-and-denoise schema. Unlike previous adversarial training approaches that sample data in adversarial scenarios, our solution does not require sampling data in an environment under attack, thereby greatly reducing risk during training. Detailed experimental results show that our technique is comparable with state-of-the-art adversarial training approaches.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78883023","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
On the Generalization of Neural Combinatorial Optimization Heuristics 神经组合优化启发式的推广
S. Manchanda, Sofia Michel, Darko Drakulic, J. Andreoli
Neural Combinatorial Optimization approaches have recently leveraged the expressiveness and flexibility of deep neural networks to learn efficient heuristics for hard Combinatorial Optimization (CO) problems. However, most of the current methods lack generalization: for a given CO problem, heuristics which are trained on instances with certain characteristics underperform when tested on instances with different characteristics. While some previous works have focused on varying the training instances properties, we postulate that a one-size-fit-all model is out of reach. Instead, we formalize solving a CO problem over a given instance distribution as a separate learning task and investigate meta-learning techniques to learn a model on a variety of tasks, in order to optimize its capacity to adapt to new tasks. Through extensive experiments, on two CO problems, using both synthetic and realistic instances, we show that our proposed meta-learning approach significantly improves the generalization of two state-of-the-art models.
神经组合优化方法最近利用深度神经网络的表达能力和灵活性来学习难组合优化(CO)问题的有效启发式。然而,目前的大多数方法缺乏泛化:对于给定的CO问题,在具有某些特征的实例上训练的启发式方法在具有不同特征的实例上测试时表现不佳。虽然以前的一些工作专注于改变训练实例的属性,但我们假设一个放之四海而皆准的模型是遥不可及的。相反,我们将解决给定实例分布上的CO问题形式化为一个单独的学习任务,并研究元学习技术来学习各种任务上的模型,以优化其适应新任务的能力。通过对两个CO问题的广泛实验,使用合成和现实实例,我们表明我们提出的元学习方法显着提高了两个最先进模型的泛化性。
{"title":"On the Generalization of Neural Combinatorial Optimization Heuristics","authors":"S. Manchanda, Sofia Michel, Darko Drakulic, J. Andreoli","doi":"10.48550/arXiv.2206.00787","DOIUrl":"https://doi.org/10.48550/arXiv.2206.00787","url":null,"abstract":"Neural Combinatorial Optimization approaches have recently leveraged the expressiveness and flexibility of deep neural networks to learn efficient heuristics for hard Combinatorial Optimization (CO) problems. However, most of the current methods lack generalization: for a given CO problem, heuristics which are trained on instances with certain characteristics underperform when tested on instances with different characteristics. While some previous works have focused on varying the training instances properties, we postulate that a one-size-fit-all model is out of reach. Instead, we formalize solving a CO problem over a given instance distribution as a separate learning task and investigate meta-learning techniques to learn a model on a variety of tasks, in order to optimize its capacity to adapt to new tasks. Through extensive experiments, on two CO problems, using both synthetic and realistic instances, we show that our proposed meta-learning approach significantly improves the generalization of two state-of-the-art models.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76555711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Factorized Structured Regression for Large-Scale Varying Coefficient Models 大型变系数模型的因式结构回归
David Rugamer, Andreas Bender, Simon Wiegrebe, Daniel Racek, Bernd Bischl, Christian L. Muller, Clemens Stachl
Recommender Systems (RS) pervade many aspects of our everyday digital life. Proposed to work at scale, state-of-the-art RS allow the modeling of thousands of interactions and facilitate highly individualized recommendations. Conceptually, many RS can be viewed as instances of statistical regression models that incorporate complex feature effects and potentially non-Gaussian outcomes. Such structured regression models, including time-aware varying coefficients models, are, however, limited in their applicability to categorical effects and inclusion of a large number of interactions. Here, we propose Factorized Structured Regression (FaStR) for scalable varying coefficient models. FaStR overcomes limitations of general regression models for large-scale data by combining structured additive regression and factorization approaches in a neural network-based model implementation. This fusion provides a scalable framework for the estimation of statistical models in previously infeasible data settings. Empirical results confirm that the estimation of varying coefficients of our approach is on par with state-of-the-art regression techniques, while scaling notably better and also being competitive with other time-aware RS in terms of prediction performance. We illustrate FaStR's performance and interpretability on a large-scale behavioral study with smartphone user data.
推荐系统(RS)遍及我们日常数字生活的许多方面。建议在规模上工作,最先进的RS允许对数千个交互进行建模,并促进高度个性化的推荐。从概念上讲,许多RS可以被视为包含复杂特征效应和潜在非高斯结果的统计回归模型的实例。然而,这种结构化回归模型,包括时间感知的变系数模型,在适用于分类效应和包含大量相互作用方面受到限制。在这里,我们提出了可扩展变系数模型的分解结构回归(FaStR)。FaStR通过在基于神经网络的模型实现中结合结构化加性回归和因子分解方法,克服了一般回归模型对大规模数据的局限性。这种融合为以前不可行的数据设置中统计模型的估计提供了一个可扩展的框架。实证结果证实,我们的方法对变化系数的估计与最先进的回归技术相当,同时缩放明显更好,并且在预测性能方面与其他时间感知RS竞争。我们在智能手机用户数据的大规模行为研究中说明了FaStR的性能和可解释性。
{"title":"Factorized Structured Regression for Large-Scale Varying Coefficient Models","authors":"David Rugamer, Andreas Bender, Simon Wiegrebe, Daniel Racek, Bernd Bischl, Christian L. Muller, Clemens Stachl","doi":"10.48550/arXiv.2205.13080","DOIUrl":"https://doi.org/10.48550/arXiv.2205.13080","url":null,"abstract":"Recommender Systems (RS) pervade many aspects of our everyday digital life. Proposed to work at scale, state-of-the-art RS allow the modeling of thousands of interactions and facilitate highly individualized recommendations. Conceptually, many RS can be viewed as instances of statistical regression models that incorporate complex feature effects and potentially non-Gaussian outcomes. Such structured regression models, including time-aware varying coefficients models, are, however, limited in their applicability to categorical effects and inclusion of a large number of interactions. Here, we propose Factorized Structured Regression (FaStR) for scalable varying coefficient models. FaStR overcomes limitations of general regression models for large-scale data by combining structured additive regression and factorization approaches in a neural network-based model implementation. This fusion provides a scalable framework for the estimation of statistical models in previously infeasible data settings. Empirical results confirm that the estimation of varying coefficients of our approach is on par with state-of-the-art regression techniques, while scaling notably better and also being competitive with other time-aware RS in terms of prediction performance. We illustrate FaStR's performance and interpretability on a large-scale behavioral study with smartphone user data.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82680514","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
MAVIPER: Learning Decision Tree Policies for Interpretable Multi-Agent Reinforcement Learning MAVIPER:可解释多智能体强化学习的学习决策树策略
Stephanie Milani, Zhicheng Zhang, Nicholay Topin, Z. Shi, C. Kamhoua, E. Papalexakis, Fei Fang
Many recent breakthroughs in multi-agent reinforcement learning (MARL) require the use of deep neural networks, which are challenging for human experts to interpret and understand. On the other hand, existing work on interpretable reinforcement learning (RL) has shown promise in extracting more interpretable decision tree-based policies from neural networks, but only in the single-agent setting. To fill this gap, we propose the first set of algorithms that extract interpretable decision-tree policies from neural networks trained with MARL. The first algorithm, IVIPER, extends VIPER, a recent method for single-agent interpretable RL, to the multi-agent setting. We demonstrate that IVIPER learns high-quality decision-tree policies for each agent. To better capture coordination between agents, we propose a novel centralized decision-tree training algorithm, MAVIPER. MAVIPER jointly grows the trees of each agent by predicting the behavior of the other agents using their anticipated trees, and uses resampling to focus on states that are critical for its interactions with other agents. We show that both algorithms generally outperform the baselines and that MAVIPER-trained agents achieve better-coordinated performance than IVIPER-trained agents on three different multi-agent particle-world environments.
最近在多智能体强化学习(MARL)方面的许多突破都需要使用深度神经网络,这对人类专家来说是一个挑战。另一方面,可解释强化学习(RL)的现有工作已经显示出从神经网络中提取更多可解释的基于决策树的策略的希望,但仅限于单智能体设置。为了填补这一空白,我们提出了第一组算法,从用MARL训练的神经网络中提取可解释的决策树策略。第一个算法IVIPER将VIPER(一种用于单代理可解释强化学习的最新方法)扩展到多代理设置。我们证明了IVIPER为每个代理学习高质量的决策树策略。为了更好地捕捉智能体之间的协调,我们提出了一种新的集中式决策树训练算法MAVIPER。mavper通过使用其他代理的预期树预测其他代理的行为来共同生长每个代理的树,并使用重新采样来关注对其与其他代理的交互至关重要的状态。我们表明,这两种算法通常都优于基线,并且在三种不同的多智能体粒子世界环境中,maviper训练的智能体比iviper训练的智能体实现了更好的协调性能。
{"title":"MAVIPER: Learning Decision Tree Policies for Interpretable Multi-Agent Reinforcement Learning","authors":"Stephanie Milani, Zhicheng Zhang, Nicholay Topin, Z. Shi, C. Kamhoua, E. Papalexakis, Fei Fang","doi":"10.48550/arXiv.2205.12449","DOIUrl":"https://doi.org/10.48550/arXiv.2205.12449","url":null,"abstract":"Many recent breakthroughs in multi-agent reinforcement learning (MARL) require the use of deep neural networks, which are challenging for human experts to interpret and understand. On the other hand, existing work on interpretable reinforcement learning (RL) has shown promise in extracting more interpretable decision tree-based policies from neural networks, but only in the single-agent setting. To fill this gap, we propose the first set of algorithms that extract interpretable decision-tree policies from neural networks trained with MARL. The first algorithm, IVIPER, extends VIPER, a recent method for single-agent interpretable RL, to the multi-agent setting. We demonstrate that IVIPER learns high-quality decision-tree policies for each agent. To better capture coordination between agents, we propose a novel centralized decision-tree training algorithm, MAVIPER. MAVIPER jointly grows the trees of each agent by predicting the behavior of the other agents using their anticipated trees, and uses resampling to focus on states that are critical for its interactions with other agents. We show that both algorithms generally outperform the baselines and that MAVIPER-trained agents achieve better-coordinated performance than IVIPER-trained agents on three different multi-agent particle-world environments.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80667173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
On the Prediction Instability of Graph Neural Networks 论图神经网络的预测不稳定性
Max Klabunde, F. Lemmerich
Instability of trained models, i.e., the dependence of individual node predictions on random factors, can affect reproducibility, reliability, and trust in machine learning systems. In this paper, we systematically assess the prediction instability of node classification with state-of-the-art Graph Neural Networks (GNNs). With our experiments, we establish that multiple instantiations of popular GNN models trained on the same data with the same model hyperparameters result in almost identical aggregated performance but display substantial disagreement in the predictions for individual nodes. We find that up to one third of the incorrectly classified nodes differ across algorithm runs. We identify correlations between hyperparameters, node properties, and the size of the training set with the stability of predictions. In general, maximizing model performance implicitly also reduces model instability.
训练模型的不稳定性,即单个节点预测对随机因素的依赖,会影响机器学习系统的可重复性、可靠性和信任度。在本文中,我们系统地评估了最先进的图神经网络(gnn)节点分类的预测不稳定性。通过我们的实验,我们建立了在相同数据上使用相同模型超参数训练的流行GNN模型的多个实例产生几乎相同的聚合性能,但在单个节点的预测中显示出实质性的分歧。我们发现,在不同的算法运行中,多达三分之一的错误分类节点是不同的。我们确定了超参数、节点属性和训练集大小与预测稳定性之间的相关性。一般来说,最大化模型性能也隐含地减少了模型的不稳定性。
{"title":"On the Prediction Instability of Graph Neural Networks","authors":"Max Klabunde, F. Lemmerich","doi":"10.48550/arXiv.2205.10070","DOIUrl":"https://doi.org/10.48550/arXiv.2205.10070","url":null,"abstract":"Instability of trained models, i.e., the dependence of individual node predictions on random factors, can affect reproducibility, reliability, and trust in machine learning systems. In this paper, we systematically assess the prediction instability of node classification with state-of-the-art Graph Neural Networks (GNNs). With our experiments, we establish that multiple instantiations of popular GNN models trained on the same data with the same model hyperparameters result in almost identical aggregated performance but display substantial disagreement in the predictions for individual nodes. We find that up to one third of the incorrectly classified nodes differ across algorithm runs. We identify correlations between hyperparameters, node properties, and the size of the training set with the stability of predictions. In general, maximizing model performance implicitly also reduces model instability.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73947934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Wasserstein t-SNE
Fynn Bachmann, Philipp Hennig, Dmitry Kobak
Scientific datasets often have hierarchical structure: for example, in surveys, individual participants (samples) might be grouped at a higher level (units) such as their geographical region. In these settings, the interest is often in exploring the structure on the unit level rather than on the sample level. Units can be compared based on the distance between their means, however this ignores the within-unit distribution of samples. Here we develop an approach for exploratory analysis of hierarchical datasets using the Wasserstein distance metric that takes into account the shapes of within-unit distributions. We use t-SNE to construct 2D embeddings of the units, based on the matrix of pairwise Wasserstein distances between them. The distance matrix can be efficiently computed by approximating each unit with a Gaussian distribution, but we also provide a scalable method to compute exact Wasserstein distances. We use synthetic data to demonstrate the effectiveness of our Wasserstein t-SNE, and apply it to data from the 2017 German parliamentary election, considering polling stations as samples and voting districts as units. The resulting embedding uncovers meaningful structure in the data.
科学数据集通常具有层次结构:例如,在调查中,个体参与者(样本)可能被分组在更高的层次(单位),例如他们的地理区域。在这些设置中,兴趣通常是在单位水平上探索结构,而不是在样本水平上。单位可以根据其均值之间的距离进行比较,但这忽略了样本的单位内分布。在这里,我们开发了一种使用Wasserstein距离度量对分层数据集进行探索性分析的方法,该度量考虑了单位内分布的形状。我们使用t-SNE来构建单元的二维嵌入,基于它们之间的成对Wasserstein距离矩阵。距离矩阵可以通过用高斯分布近似每个单元来有效地计算,但我们也提供了一种可扩展的方法来计算精确的Wasserstein距离。我们使用合成数据来证明我们的Wasserstein t-SNE的有效性,并将其应用于2017年德国议会选举的数据,以投票站为样本,以投票区为单位。由此产生的嵌入揭示了数据中有意义的结构。
{"title":"Wasserstein t-SNE","authors":"Fynn Bachmann, Philipp Hennig, Dmitry Kobak","doi":"10.48550/arXiv.2205.07531","DOIUrl":"https://doi.org/10.48550/arXiv.2205.07531","url":null,"abstract":"Scientific datasets often have hierarchical structure: for example, in surveys, individual participants (samples) might be grouped at a higher level (units) such as their geographical region. In these settings, the interest is often in exploring the structure on the unit level rather than on the sample level. Units can be compared based on the distance between their means, however this ignores the within-unit distribution of samples. Here we develop an approach for exploratory analysis of hierarchical datasets using the Wasserstein distance metric that takes into account the shapes of within-unit distributions. We use t-SNE to construct 2D embeddings of the units, based on the matrix of pairwise Wasserstein distances between them. The distance matrix can be efficiently computed by approximating each unit with a Gaussian distribution, but we also provide a scalable method to compute exact Wasserstein distances. We use synthetic data to demonstrate the effectiveness of our Wasserstein t-SNE, and apply it to data from the 2017 German parliamentary election, considering polling stations as samples and voting districts as units. The resulting embedding uncovers meaningful structure in the data.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86326688","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Near out-of-distribution detection for low-resolution radar micro-Doppler signatures 低分辨率雷达微多普勒特征的近离分布检测
Martin Bauw, S. Velasco-Forero, J. Angulo, C. Adnet, O. Airiau
Near out-of-distribution detection (OODD) aims at discriminating semantically similar data points without the supervision required for classification. This paper puts forward an OODD use case for radar targets detection extensible to other kinds of sensors and detection scenarios. We emphasize the relevance of OODD and its specific supervision requirements for the detection of a multimodal, diverse targets class among other similar radar targets and clutter in real-life critical systems. We propose a comparison of deep and non-deep OODD methods on simulated low-resolution pulse radar micro-Doppler signatures, considering both a spectral and a covariance matrix input representation. The covariance representation aims at estimating whether dedicated second-order processing is appropriate to discriminate signatures. The potential contributions of labeled anomalies in training, self-supervised learning, contrastive learning insights and innovative training losses are discussed, and the impact of training set contamination caused by mislabelling is investigated.
近离分布检测(OODD)的目的是在没有分类所需的监督的情况下区分语义上相似的数据点。提出了一种面向对象的雷达目标检测用例,可扩展到其他类型的传感器和检测场景。我们强调OODD的相关性及其在实际关键系统中对其他类似雷达目标和杂波中的多模态、不同目标类别的检测的具体监督要求。我们提出了模拟低分辨率脉冲雷达微多普勒特征的深度和非深度OODD方法的比较,同时考虑了谱和协方差矩阵输入表示。协方差表示的目的是估计专用二阶处理是否适合区分签名。讨论了标记异常在训练、自监督学习、对比学习洞察和创新训练损失中的潜在贡献,并研究了错误标记导致的训练集污染的影响。
{"title":"Near out-of-distribution detection for low-resolution radar micro-Doppler signatures","authors":"Martin Bauw, S. Velasco-Forero, J. Angulo, C. Adnet, O. Airiau","doi":"10.48550/arXiv.2205.07869","DOIUrl":"https://doi.org/10.48550/arXiv.2205.07869","url":null,"abstract":"Near out-of-distribution detection (OODD) aims at discriminating semantically similar data points without the supervision required for classification. This paper puts forward an OODD use case for radar targets detection extensible to other kinds of sensors and detection scenarios. We emphasize the relevance of OODD and its specific supervision requirements for the detection of a multimodal, diverse targets class among other similar radar targets and clutter in real-life critical systems. We propose a comparison of deep and non-deep OODD methods on simulated low-resolution pulse radar micro-Doppler signatures, considering both a spectral and a covariance matrix input representation. The covariance representation aims at estimating whether dedicated second-order processing is appropriate to discriminate signatures. The potential contributions of labeled anomalies in training, self-supervised learning, contrastive learning insights and innovative training losses are discussed, and the impact of training set contamination caused by mislabelling is investigated.","PeriodicalId":74091,"journal":{"name":"Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83909963","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
Machine learning and knowledge discovery in databases : European Conference, ECML PKDD ... : proceedings. ECML PKDD (Conference)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1