首页 > 最新文献

Artificial Intelligence最新文献

英文 中文
An abstract and structured account of dialectical argument strength 抽象而有条理地阐述辩证论证的力量
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-30 DOI: 10.1016/j.artint.2024.104193

This paper presents a formal model of dialectical argument strength in terms of the number of ways in which an argument can be successfully attacked in expansions of an abstract argumentation framework. First a model is proposed that is abstract but designed to avoid overly limiting assumptions on instantiations or dialogue contexts. It is then shown that most principles for argument strength proposed in the literature fail to hold for the proposed notions of dialectical strength, which clarifies the rational foundations of these principles and highlights the importance of distinguishing between kinds of argument strength, in particular logical, dialectical and rhetorical argument strength. The abstract model is then instantiated with ASPIC+ to test the claim that it does not make overly limiting assumptions on the structure of arguments and the nature of their relations.

本文提出了一个辩证论证强度的正式模型,即在抽象论证框架的扩展中,论证可以成功攻击的方式数量。首先,本文提出了一个抽象模型,但旨在避免对实例或对话语境的过度限制性假设。然后证明了文献中提出的大多数论证强度原则对于所提出的辩证强度概念都是不成立的,这澄清了这些原则的合理性基础,并强调了区分论证强度类型的重要性,尤其是逻辑、辩证和修辞论证强度。然后,我们将抽象模型实例化,以检验该模型是否对论证结构及其关系的性质做出了过度限制性的假设。
{"title":"An abstract and structured account of dialectical argument strength","authors":"","doi":"10.1016/j.artint.2024.104193","DOIUrl":"10.1016/j.artint.2024.104193","url":null,"abstract":"<div><p>This paper presents a formal model of dialectical argument strength in terms of the number of ways in which an argument can be successfully attacked in expansions of an abstract argumentation framework. First a model is proposed that is abstract but designed to avoid overly limiting assumptions on instantiations or dialogue contexts. It is then shown that most principles for argument strength proposed in the literature fail to hold for the proposed notions of dialectical strength, which clarifies the rational foundations of these principles and highlights the importance of distinguishing between kinds of argument strength, in particular logical, dialectical and rhetorical argument strength. The abstract model is then instantiated with <em>ASPIC</em><sup>+</sup> to test the claim that it does not make overly limiting assumptions on the structure of arguments and the nature of their relations.</p></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0004370224001292/pdfft?md5=ccf1642809aafb2e2582cad1cfca01a9&pid=1-s2.0-S0004370224001292-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141891926","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-objective meta-learning 多目标元学习
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-25 DOI: 10.1016/j.artint.2024.104184

Meta-learning has arisen as a powerful tool for many machine learning problems. With multiple factors to be considered when designing learning models for real-world applications, meta-learning with multiple objectives has attracted much attention recently. However, existing works either linearly combine multiple objectives into one objective or adopt evolutionary algorithms to handle it, where the former approach needs to pay high computational cost to tune the combination coefficients while the latter approach is computationally heavy and incapable to be integrated into gradient-based optimization. To alleviate those limitations, in this paper, we aim to propose a generic gradient-based Multi-Objective Meta-Learning (MOML) framework with applications in many machine learning problems. Specifically, the MOML framework formulates the objective function of meta-learning with multiple objectives as a Multi-Objective Bi-Level optimization Problem (MOBLP) where the upper-level subproblem is to solve several possibly conflicting objectives for the meta-learner. Different from those existing works, in this paper, we propose a gradient-based algorithm to solve the MOBLP. Specifically, we devise the first gradient-based optimization algorithm by alternately solving the lower-level and upper-level subproblems via the gradient descent method and the gradient-based multi-objective optimization method, respectively. Theoretically, we prove the convergence property and provide a non-asymptotic analysis of the proposed gradient-based optimization algorithm. Empirically, extensive experiments justify our theoretical results and demonstrate the superiority of the proposed MOML framework for different learning problems, including few-shot learning, domain adaptation, multi-task learning, neural architecture search, and reinforcement learning. The source code of MOML is available at https://github.com/Baijiong-Lin/MOML.

元学习已成为解决许多机器学习问题的有力工具。在为实际应用设计学习模型时,需要考虑多种因素,因此具有多个目标的元学习近年来备受关注。然而,现有的研究要么将多个目标线性组合为一个目标,要么采用进化算法来处理,前者需要付出高昂的计算成本来调整组合系数,而后者计算量大,无法集成到基于梯度的优化中。为了缓解这些局限性,我们在本文中提出了一种通用的基于梯度的多目标元学习(MOML)框架,可应用于许多机器学习问题。具体来说,MOML 框架将多目标元学习的目标函数表述为多目标双层优化问题(MOBLP),其中上层子问题是解决元学习器的几个可能相互冲突的目标。与现有研究不同,本文提出了一种基于梯度的算法来解决 MOBLP。具体来说,我们设计了第一种基于梯度的优化算法,分别通过梯度下降法和基于梯度的多目标优化法交替求解低级子问题和高级子问题。我们从理论上证明了收敛性,并对所提出的基于梯度的优化算法进行了非渐近分析。从经验上讲,大量实验证明了我们的理论结果,并证明了所提出的 MOML 框架在不同学习问题上的优越性,包括少量学习、领域适应、多任务学习、神经架构搜索和强化学习。MOML的源代码可在https://github.com/Baijiong-Lin/MOML。
{"title":"Multi-objective meta-learning","authors":"","doi":"10.1016/j.artint.2024.104184","DOIUrl":"10.1016/j.artint.2024.104184","url":null,"abstract":"<div><p>Meta-learning has arisen as a powerful tool for many machine learning problems. With multiple factors to be considered when designing learning models for real-world applications, meta-learning with multiple objectives has attracted much attention recently. However, existing works either linearly combine multiple objectives into one objective or adopt evolutionary algorithms to handle it, where the former approach needs to pay high computational cost to tune the combination coefficients while the latter approach is computationally heavy and incapable to be integrated into gradient-based optimization. To alleviate those limitations, in this paper, we aim to propose a generic gradient-based Multi-Objective Meta-Learning (MOML) framework with applications in many machine learning problems. Specifically, the MOML framework formulates the objective function of meta-learning with multiple objectives as a Multi-Objective Bi-Level optimization Problem (MOBLP) where the upper-level subproblem is to solve several possibly conflicting objectives for the meta-learner. Different from those existing works, in this paper, we propose a gradient-based algorithm to solve the MOBLP. Specifically, we devise the first gradient-based optimization algorithm by alternately solving the lower-level and upper-level subproblems via the gradient descent method and the gradient-based multi-objective optimization method, respectively. Theoretically, we prove the convergence property and provide a non-asymptotic analysis of the proposed gradient-based optimization algorithm. Empirically, extensive experiments justify our theoretical results and demonstrate the superiority of the proposed MOML framework for different learning problems, including few-shot learning, domain adaptation, multi-task learning, neural architecture search, and reinforcement learning. The source code of MOML is available at <span><span>https://github.com/Baijiong-Lin/MOML</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141851427","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A crossword solving system based on Monte Carlo tree search 基于蒙特卡洛树搜索的填字游戏解题系统
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-25 DOI: 10.1016/j.artint.2024.104192

Although the development of AI in games is remarkable, intelligent machines still lag behind humans in games that require the ability of language understanding. In this paper, we focus on the crossword puzzle resolution task. Solving crossword puzzles is a challenging task since it requires the ability to answer natural language questions with knowledge and the ability to execute a search over possible answers to find an optimal set of solutions for the grid. Previous solutions are devoted to exploiting heuristic strategies in search to find solutions while having limited ability to explore the search space. We build a comprehensive system for crossword puzzle resolution based on Monte Carlo Tree Search (MCTS). As far as we know, we are the first to model the crossword puzzle resolution problem as a Markov Decision Process and apply the MCTS to solve it. We construct a dataset for crossword puzzle resolution based on daily puzzles from The New York Times with detailed specifications of both the puzzle and clue database selection. Our method achieves state-of-the-art performance on the dataset. The code of the system and experiments in this paper is publicly available: https://www.github.com/lhlclhl/CP.

尽管人工智能在游戏领域的发展令人瞩目,但在需要语言理解能力的游戏中,智能机器仍然落后于人类。在本文中,我们将重点讨论填字游戏的解题任务。解决填字游戏是一项极具挑战性的任务,因为它需要用知识回答自然语言问题的能力,以及对可能的答案进行搜索以找到网格的最优解集的能力。以往的解决方案都是利用启发式搜索策略来寻找解决方案,但探索搜索空间的能力有限。我们基于蒙特卡洛树搜索(Monte Carlo Tree Search,MCTS)建立了一个全面的填字游戏解题系统。据我们所知,我们是第一个将填字谜题解析问题建模为马尔可夫决策过程并应用 MCTS 解决该问题的人。我们基于《纽约时报》的每日谜题构建了一个字谜解析数据集,并对谜题和线索数据库的选择进行了详细说明。我们的方法在该数据集上取得了最先进的性能。本文中的系统和实验代码已公开:https://www.github.com/lhlclhl/CP。
{"title":"A crossword solving system based on Monte Carlo tree search","authors":"","doi":"10.1016/j.artint.2024.104192","DOIUrl":"10.1016/j.artint.2024.104192","url":null,"abstract":"<div><p>Although the development of AI in games is remarkable, intelligent machines still lag behind humans in games that require the ability of language understanding. In this paper, we focus on the crossword puzzle resolution task. Solving crossword puzzles is a challenging task since it requires the ability to answer natural language questions with knowledge and the ability to execute a search over possible answers to find an optimal set of solutions for the grid. Previous solutions are devoted to exploiting heuristic strategies in search to find solutions while having limited ability to explore the search space. We build a comprehensive system for crossword puzzle resolution based on Monte Carlo Tree Search (MCTS). As far as we know, we are the first to model the crossword puzzle resolution problem as a Markov Decision Process and apply the MCTS to solve it. We construct a dataset for crossword puzzle resolution based on daily puzzles from The New York Times with detailed specifications of both the puzzle and clue database selection. Our method achieves state-of-the-art performance on the dataset. The code of the system and experiments in this paper is publicly available: <span><span>https://www.github.com/lhlclhl/CP</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141842906","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ASQ-IT: Interactive explanations for reinforcement-learning agents ASQ-IT:强化学习代理的交互式解释
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-22 DOI: 10.1016/j.artint.2024.104182

As reinforcement learning methods increasingly amass accomplishments, the need for comprehending their solutions becomes more crucial. Most explainable reinforcement learning (XRL) methods generate a static explanation depicting their developers' intuition of what should be explained and how. In contrast, literature from the social sciences proposes that meaningful explanations are structured as a dialog between the explainer and the explainee, suggesting a more active role for the user and her communication with the agent. In this paper, we present ASQ-IT – an interactive explanation system that presents video clips of the agent acting in its environment based on queries given by the user that describe temporal properties of behaviors of interest. Our approach is based on formal methods: queries in ASQ-IT's user interface map to a fragment of Linear Temporal Logic over finite traces (LTLf), which we developed, and our algorithm for query processing is based on automata theory. User studies show that end-users can understand and formulate queries in ASQ-IT and that using ASQ-IT assists users in identifying faulty agent behaviors.

随着强化学习方法取得越来越多的成就,理解这些方法的解决方案变得越来越重要。大多数可解释强化学习(XRL)方法都会生成静态的解释,描述其开发者对应该解释什么以及如何解释的直觉。与此相反,社会科学文献提出,有意义的解释是解释者与被解释者之间的对话结构,这表明用户在与代理的交流中扮演着更积极的角色。在本文中,我们介绍了 ASQ-IT--一个交互式解释系统,它可以根据用户提出的描述相关行为时间属性的询问,展示代理在其环境中行动的视频片段。我们的方法基于形式化方法:ASQ-IT 用户界面中的查询映射到我们开发的有限踪迹线性时态逻辑(LTLf)片段,我们的查询处理算法基于自动机理论。用户研究表明,终端用户能够理解并在 ASQ-IT 中提出查询,而且 ASQ-IT 还能帮助用户识别故障代理行为。
{"title":"ASQ-IT: Interactive explanations for reinforcement-learning agents","authors":"","doi":"10.1016/j.artint.2024.104182","DOIUrl":"10.1016/j.artint.2024.104182","url":null,"abstract":"<div><p>As reinforcement learning methods increasingly amass accomplishments, the need for comprehending their solutions becomes more crucial. Most explainable reinforcement learning (XRL) methods generate a static explanation depicting their developers' intuition of what should be explained and how. In contrast, literature from the social sciences proposes that meaningful explanations are structured as a dialog between the explainer and the explainee, suggesting a more active role for the user and her communication with the agent. In this paper, we present ASQ-IT – an interactive explanation system that presents video clips of the agent acting in its environment based on queries given by the user that describe temporal properties of behaviors of interest. Our approach is based on formal methods: queries in ASQ-IT's user interface map to a fragment of Linear Temporal Logic over finite traces (LTLf), which we developed, and our algorithm for query processing is based on automata theory. User studies show that end-users can understand and formulate queries in ASQ-IT and that using ASQ-IT assists users in identifying faulty agent behaviors.</p></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141952324","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Planning with mental models – Balancing explanations and explicability 用心智模式进行规划 - 平衡解释与可解释性之间的关系--用心智模式进行规划
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-18 DOI: 10.1016/j.artint.2024.104181

Human-aware planning involves generating plans that are explicable, i.e. conform to user expectations, as well as providing explanations when such plans cannot be found. In this paper, we bring these two concepts together and show how an agent can achieve a trade-off between these two competing characteristics of a plan. To achieve this, we conceive a first-of-its-kind planner MEGA that can reason about the possibility of explaining a plan in the plan generation process itself. We will also explore how solutions to such problems can be expressed as “self-explaining plans” – and show how this representation allows us to leverage classical planning compilations of epistemic planning to reason about this trade-off at plan generation time without having to incur the computational burden of having to search in the space of differences between the agent model and the mental model of the human in the loop in order to come up with the optimal trade-off. We will illustrate these concepts in two well-known planning domains, as well as with a robot in a typical search and reconnaissance task. Human factor studies in the latter highlight the usefulness of the proposed approach.

人类感知规划涉及生成可解释的计划,即符合用户期望的计划,以及在无法找到此类计划时提供解释。在本文中,我们将这两个概念结合在一起,并展示了代理如何在计划的这两个相互竞争的特性之间实现权衡。为了实现这一目标,我们构想了一种首创的计划器 MEGA,它可以在计划生成过程中推理出解释计划的可能性。我们还将探讨如何将此类问题的解决方案表述为 "自我解释计划",并展示这种表述方式如何让我们在计划生成时利用经典的认识论计划编译来推理这种权衡,而无需为了得出最优权衡结果而在代理模型和环路中人的心智模型之间的差异空间中进行搜索,从而造成计算负担。我们将在两个著名的规划领域以及典型的搜索和侦察任务中使用机器人来说明这些概念。在后者中进行的人为因素研究突出了所建议方法的实用性。
{"title":"Planning with mental models – Balancing explanations and explicability","authors":"","doi":"10.1016/j.artint.2024.104181","DOIUrl":"10.1016/j.artint.2024.104181","url":null,"abstract":"<div><p>Human-aware planning involves generating plans that are explicable, i.e. conform to user expectations, as well as providing explanations when such plans cannot be found. In this paper, we bring these two concepts together and show how an agent can achieve a trade-off between these two competing characteristics of a plan. To achieve this, we conceive a first-of-its-kind planner <span>MEGA</span> that can reason about the possibility of explaining a plan <em>in the plan generation process itself</em>. We will also explore how solutions to such problems can be expressed as “self-explaining plans” – and show how this representation allows us to leverage classical planning compilations of epistemic planning to reason about this trade-off at plan generation time without having to incur the computational burden of having to search in the space of differences between the agent model and the mental model of the human in the loop in order to come up with the optimal trade-off. We will illustrate these concepts in two well-known planning domains, as well as with a robot in a typical search and reconnaissance task. Human factor studies in the latter highlight the usefulness of the proposed approach.</p></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141842294","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A note on incorrect inferences in non-binary qualitative probabilistic networks 关于非二元定性概率网络中不正确推论的说明
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-14 DOI: 10.1016/j.artint.2024.104180

Qualitative probabilistic networks (QPNs) combine the conditional independence assumptions of Bayesian networks with the qualitative properties of positive and negative dependence. They formalise various intuitive properties of positive dependence to allow inferences over a large network of variables. However, we will demonstrate in this paper that, due to an incorrect symmetry property, many inferences obtained in non-binary QPNs are not mathematically true. We will provide examples of such incorrect inferences and briefly discuss possible resolutions.

定性概率网络(QPN)结合了贝叶斯网络的条件独立性假设和正负依赖性的定性特性。定性概率网络将正相关性的各种直观特性形式化,允许对大型变量网络进行推断。然而,我们将在本文中证明,由于不正确的对称属性,在非二元 QPN 中得到的许多推论在数学上并不正确。我们将举例说明这种不正确的推论,并简要讨论可能的解决方法。
{"title":"A note on incorrect inferences in non-binary qualitative probabilistic networks","authors":"","doi":"10.1016/j.artint.2024.104180","DOIUrl":"10.1016/j.artint.2024.104180","url":null,"abstract":"<div><p>Qualitative probabilistic networks (QPNs) combine the conditional independence assumptions of Bayesian networks with the qualitative properties of positive and negative dependence. They formalise various intuitive properties of positive dependence to allow inferences over a large network of variables. However, we will demonstrate in this paper that, due to an incorrect symmetry property, many inferences obtained in non-binary QPNs are not mathematically true. We will provide examples of such incorrect inferences and briefly discuss possible resolutions.</p></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141729395","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Assessing fidelity in XAI post-hoc techniques: A comparative study with ground truth explanations datasets 评估 XAI 事后技术的保真度:与地面实况解释数据集的比较研究
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-11 DOI: 10.1016/j.artint.2024.104179

The evaluation of the fidelity of eXplainable Artificial Intelligence (XAI) methods to their underlying models is a challenging task, primarily due to the absence of a ground truth for explanations. However, assessing fidelity is a necessary step for ensuring a correct XAI methodology. In this study, we conduct a fair and objective comparison of the current state-of-the-art XAI methods by introducing three novel image datasets with reliable ground truth for explanations. The primary objective of this comparison is to identify methods with low fidelity and eliminate them from further research, thereby promoting the development of more trustworthy and effective XAI techniques. Our results demonstrate that XAI methods based on the direct gradient calculation and the backpropagation of output information to input yield higher accuracy and reliability compared to methods relying on perturbation based or Class Activation Maps (CAM). However, these methods tend to generate more noisy saliency maps. These findings have significant implications for the advancement of XAI methods, enabling the elimination of erroneous explanations and fostering the development of more robust and reliable XAI.

评估可解释人工智能(XAI)方法与其基础模型的保真度是一项具有挑战性的任务,这主要是由于缺乏解释的基本真相。然而,评估保真度是确保 XAI 方法正确性的必要步骤。在本研究中,我们通过引入三个具有可靠基本真相解释的新型图像数据集,对当前最先进的 XAI 方法进行了公平客观的比较。比较的主要目的是找出低保真度的方法,并将其从进一步的研究中剔除,从而促进开发更可信、更有效的 XAI 技术。我们的研究结果表明,与基于扰动或类激活图(CAM)的方法相比,基于直接梯度计算和将输出信息反向传播到输入的 XAI 方法具有更高的准确性和可靠性。不过,这些方法往往会生成噪声更大的显著性地图。这些发现对 XAI 方法的发展具有重要意义,可以消除错误的解释,促进更稳健、更可靠的 XAI 的发展。
{"title":"Assessing fidelity in XAI post-hoc techniques: A comparative study with ground truth explanations datasets","authors":"","doi":"10.1016/j.artint.2024.104179","DOIUrl":"10.1016/j.artint.2024.104179","url":null,"abstract":"<div><p>The evaluation of the fidelity of eXplainable Artificial Intelligence (XAI) methods to their underlying models is a challenging task, primarily due to the absence of a ground truth for explanations. However, assessing fidelity is a necessary step for ensuring a correct XAI methodology. In this study, we conduct a fair and objective comparison of the current state-of-the-art XAI methods by introducing three novel image datasets with reliable ground truth for explanations. The primary objective of this comparison is to identify methods with low fidelity and eliminate them from further research, thereby promoting the development of more trustworthy and effective XAI techniques. Our results demonstrate that XAI methods based on the direct gradient calculation and the backpropagation of output information to input yield higher accuracy and reliability compared to methods relying on perturbation based or Class Activation Maps (CAM). However, these methods tend to generate more noisy saliency maps. These findings have significant implications for the advancement of XAI methods, enabling the elimination of erroneous explanations and fostering the development of more robust and reliable XAI.</p></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0004370224001152/pdfft?md5=aa7feedd4216bc7f0fc0893dee291096&pid=1-s2.0-S0004370224001152-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141704429","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Class fairness in online matching 在线匹配中的班级公平性
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-09 DOI: 10.1016/j.artint.2024.104177

We initiate the study of fairness among classes of agents in online bipartite matching where there is a given set of offline vertices (aka agents) and another set of vertices (aka items) that arrive online and must be matched irrevocably upon arrival. In this setting, agents are partitioned into classes and the matching is required to be fair with respect to the classes. We adapt popular fairness notions (e.g. envy-freeness, proportionality, and maximin share) and their relaxations to this setting and study deterministic algorithms for matching indivisible items (leading to integral matchings) and for matching divisible items (leading to fractional matchings). For matching indivisible items, we propose an adaptive-priority-based algorithm, Match-and-Shift, prove that it achieves 12-approximation of both class envy-freeness up to one item and class maximin share fairness, and show that each guarantee is tight. For matching divisible items, we design a water-filling-based algorithm, Equal-Filling, that achieves (11e)-approximation of class envy-freeness and class proportionality; we prove 11e to be tight for class proportionality and establish a 34 upper bound on class envy-freeness. Finally, we discuss several challenges in designing randomized algorithms that achieve reasonable fairness approximation ratios. Nonetheless, we build upon Equal-Filling to design a randomized algorithm for matching indivisible items, Equal-Filling-OCS, which achieves 0.593-approximation of class proportionality.

我们开始研究在线双向匹配中不同类别代理之间的公平性,其中有一组给定的离线顶点(又称代理)和另一组在线到达的顶点(又称项目),到达后必须进行不可撤销的匹配。在这种情况下,代理被划分为不同的类别,而匹配要求在不同类别之间是公平的。我们将流行的公平性概念(如无妒忌、比例性和最大份额)及其松弛概念应用于这一环境,并研究不可分割项目匹配(导致积分匹配)和可分割项目匹配(导致分数匹配)的确定性算法。对于不可分割项的匹配,我们提出了一种基于自适应优先级的算法--"匹配与转移"(Match-and-Shift),并证明该算法可以实现 12 近似值的类嫉妒无忧(最多一个项)和类最大份额公平性,还证明了每种保证都很严密。对于可分割项的匹配,我们设计了一种基于注水的算法 Equal-Filling,它能实现 (1-1e)- 近似的类嫉妒无绿化度和类比例度;我们证明了 1-1e 对于类比例度是紧密的,并建立了类嫉妒无绿化度的 34 上界。最后,我们讨论了设计能达到合理公平近似率的随机算法所面临的几个挑战。尽管如此,我们还是在 Equal-Filling 的基础上设计了一种用于匹配不可分割项的随机算法 Equal-Filling-OCS,它达到了 0.593 的类比例近似值。
{"title":"Class fairness in online matching","authors":"","doi":"10.1016/j.artint.2024.104177","DOIUrl":"10.1016/j.artint.2024.104177","url":null,"abstract":"<div><p>We initiate the study of fairness among <em>classes</em> of agents in online bipartite matching where there is a given set of offline vertices (aka agents) and another set of vertices (aka items) that arrive online and must be matched irrevocably upon arrival. In this setting, agents are partitioned into classes and the matching is required to be fair with respect to the classes. We adapt popular fairness notions (e.g. envy-freeness, proportionality, and maximin share) and their relaxations to this setting and study deterministic algorithms for matching indivisible items (leading to integral matchings) and for matching divisible items (leading to fractional matchings). For matching indivisible items, we propose an adaptive-priority-based algorithm, <span>Match-and-Shift</span>, prove that it achieves <span><math><mfrac><mrow><mn>1</mn></mrow><mrow><mn>2</mn></mrow></mfrac></math></span>-approximation of both class envy-freeness up to one item and class maximin share fairness, and show that each guarantee is tight. For matching divisible items, we design a water-filling-based algorithm, <span>Equal-Filling</span>, that achieves <span><math><mo>(</mo><mn>1</mn><mo>−</mo><mfrac><mrow><mn>1</mn></mrow><mrow><mi>e</mi></mrow></mfrac><mo>)</mo></math></span>-approximation of class envy-freeness and class proportionality; we prove <span><math><mn>1</mn><mo>−</mo><mfrac><mrow><mn>1</mn></mrow><mrow><mi>e</mi></mrow></mfrac></math></span> to be tight for class proportionality and establish a <span><math><mfrac><mrow><mn>3</mn></mrow><mrow><mn>4</mn></mrow></mfrac></math></span> upper bound on class envy-freeness. Finally, we discuss several challenges in designing randomized algorithms that achieve reasonable fairness approximation ratios. Nonetheless, we build upon <span>Equal-Filling</span> to design a randomized algorithm for matching indivisible items, <span>Equal-Filling-OCS</span>, which achieves 0.593-approximation of class proportionality.</p></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141638644","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Incremental measurement of structural entropy for dynamic graphs 动态图形结构熵的增量测量
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-02 DOI: 10.1016/j.artint.2024.104175
Runze Yang , Hao Peng , Chunyang Liu , Angsheng Li

Structural entropy is a metric that measures the amount of information embedded in graph structure data under a strategy of hierarchical abstracting. To measure the structural entropy of a dynamic graph, we need to decode the optimal encoding tree corresponding to the best community partitioning for each snapshot. However, the current methods do not support dynamic encoding tree updating and incremental structural entropy computation. To address this issue, we propose Incre-2dSE, a novel incremental measurement framework that dynamically adjusts the community partitioning and efficiently computes the updated structural entropy for each updated graph. Specifically, Incre-2dSE includes incremental algorithms based on two dynamic adjustment strategies for two-dimensional encoding trees, i.e., the naive adjustment strategy and the node-shifting adjustment strategy, which support theoretical analysis of updated structural entropy and incrementally optimize community partitioning towards a lower structural entropy. We conduct extensive experiments on 3 artificial datasets generated by Hawkes Process and 3 real-world datasets. Experimental results confirm that our incremental algorithms effectively capture the dynamic evolution of the communities, reduce time consumption, and provide great interpretability.

结构熵是一种度量指标,用于衡量在分层抽象策略下图结构数据所蕴含的信息量。要测量动态图的结构熵,我们需要解码与每个快照的最佳群落划分相对应的最优编码树。然而,目前的方法不支持动态编码树更新和增量结构熵计算。为了解决这个问题,我们提出了 Incre-2dSE,这是一个新颖的增量测量框架,可以动态调整社区划分,并为每个更新的图有效计算更新的结构熵。具体来说,Incre-2dSE 包括基于二维编码树的两种动态调整策略(即天真调整策略和节点移动调整策略)的增量算法,它们支持对更新结构熵的理论分析,并朝着更低的结构熵增量优化群落划分。我们在霍克斯过程生成的 3 个人工数据集和 3 个真实世界数据集上进行了大量实验。实验结果证实,我们的增量算法能有效捕捉社群的动态演化,减少时间消耗,并提供很好的可解释性。
{"title":"Incremental measurement of structural entropy for dynamic graphs","authors":"Runze Yang ,&nbsp;Hao Peng ,&nbsp;Chunyang Liu ,&nbsp;Angsheng Li","doi":"10.1016/j.artint.2024.104175","DOIUrl":"https://doi.org/10.1016/j.artint.2024.104175","url":null,"abstract":"<div><p>Structural entropy is a metric that measures the amount of information embedded in graph structure data under a strategy of hierarchical abstracting. To measure the structural entropy of a dynamic graph, we need to decode the optimal encoding tree corresponding to the best community partitioning for each snapshot. However, the current methods do not support dynamic encoding tree updating and incremental structural entropy computation. To address this issue, we propose <em>Incre-2dSE</em>, a novel incremental measurement framework that dynamically adjusts the community partitioning and efficiently computes the updated structural entropy for each updated graph. Specifically, <em>Incre-2dSE</em> includes incremental algorithms based on two dynamic adjustment strategies for two-dimensional encoding trees, i.e., <em>the naive adjustment strategy</em> and <em>the node-shifting adjustment strategy</em>, which support theoretical analysis of updated structural entropy and incrementally optimize community partitioning towards a lower structural entropy. We conduct extensive experiments on 3 artificial datasets generated by <em>Hawkes Process</em> and 3 real-world datasets. Experimental results confirm that our incremental algorithms effectively capture the dynamic evolution of the communities, reduce time consumption, and provide great interpretability.</p></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141542832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Controlled query evaluation in description logics through consistent query answering 通过一致的查询回答在描述符逻辑中进行受控查询评估
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-02 DOI: 10.1016/j.artint.2024.104176
Gianluca Cima , Domenico Lembo , Riccardo Rosati , Domenico Fabio Savo

Controlled Query Evaluation (CQE) is a framework for the protection of confidential data, where a policy given in terms of logic formulae indicates which information must be kept private. Functions called censors filter query answering so that no answers are returned that may lead a user to infer data protected by the policy. The preferred censors, called optimal censors, are the ones that conceal only what is necessary, thus maximizing the returned answers. Typically, given a policy over a data or knowledge base, several optimal censors exist.

Our research on CQE is based on the following intuition: confidential data are those that violate the logical assertions specifying the policy, and thus censoring them in query answering is similar to processing queries in the presence of inconsistent data as studied in Consistent Query Answering (CQA). In this paper, we investigate the relationship between CQE and CQA in the context of Description Logic ontologies. We borrow the idea from CQA that query answering is a form of skeptical reasoning that takes into account all possible optimal censors. This approach leads to a revised notion of CQE, which allows us to avoid making an arbitrary choice on the censor to be selected, as done by previous research on the topic.

We then study the data complexity of query answering in our CQE framework, for conjunctive queries issued over ontologies specified in the popular Description Logics DL-LiteR and EL. In our analysis, we consider some variants of the censor language, which is the language used by the censor to enforce the policy. Whereas the problem is in general intractable for simple censor languages, we show that for DL-LiteR ontologies it is first-order rewritable, and thus in AC0 in data complexity, for the most expressive censor language we propose.

受控查询评估(CQE)是一种用于保护机密数据的框架,它通过逻辑公式给出的政策来指明哪些信息必须保密。称为审查器的函数对查询回答进行过滤,以避免返回的答案可能导致用户推断出受政策保护的数据。被称为最优审查器的首选审查器只隐藏必要的信息,从而使返回的答案最大化。我们对 CQE 的研究基于以下直觉:机密数据是那些违反指定策略的逻辑断言的数据,因此在查询回答中审查这些数据类似于在一致性查询回答 (CQA) 中研究的在存在不一致数据的情况下处理查询。在本文中,我们将在描述逻辑本体的背景下研究 CQE 和 CQA 之间的关系。我们借鉴了 CQA 的思想,即查询回答是一种怀疑推理,它考虑到了所有可能的最优审查者。然后,我们在 CQE 框架中研究了查询回答的数据复杂性,适用于在流行的描述逻辑 DL-LiteR 和 EL⊥ 中指定的本体上发出的连接查询。在分析中,我们考虑了审查员语言的一些变体,即审查员用于执行策略的语言。对于简单的审查员语言来说,这个问题一般是难以解决的,而对于我们提出的最具表现力的审查员语言来说,我们证明了对于 DL-LiteR 本体来说,这个问题是一阶可重写的,因此数据复杂度为 AC0。
{"title":"Controlled query evaluation in description logics through consistent query answering","authors":"Gianluca Cima ,&nbsp;Domenico Lembo ,&nbsp;Riccardo Rosati ,&nbsp;Domenico Fabio Savo","doi":"10.1016/j.artint.2024.104176","DOIUrl":"https://doi.org/10.1016/j.artint.2024.104176","url":null,"abstract":"<div><p>Controlled Query Evaluation (CQE) is a framework for the protection of confidential data, where a <em>policy</em> given in terms of logic formulae indicates which information must be kept private. Functions called <em>censors</em> filter query answering so that no answers are returned that may lead a user to infer data protected by the policy. The preferred censors, called <em>optimal</em> censors, are the ones that conceal only what is necessary, thus maximizing the returned answers. Typically, given a policy over a data or knowledge base, several optimal censors exist.</p><p>Our research on CQE is based on the following intuition: confidential data are those that violate the logical assertions specifying the policy, and thus censoring them in query answering is similar to processing queries in the presence of inconsistent data as studied in Consistent Query Answering (CQA). In this paper, we investigate the relationship between CQE and CQA in the context of Description Logic ontologies. We borrow the idea from CQA that query answering is a form of skeptical reasoning that takes into account all possible optimal censors. This approach leads to a revised notion of CQE, which allows us to avoid making an arbitrary choice on the censor to be selected, as done by previous research on the topic.</p><p>We then study the data complexity of query answering in our CQE framework, for conjunctive queries issued over ontologies specified in the popular Description Logics <span><math><msub><mrow><mtext>DL-Lite</mtext></mrow><mrow><mi>R</mi></mrow></msub></math></span> and <span><math><msub><mrow><mi>EL</mi></mrow><mrow><mo>⊥</mo></mrow></msub></math></span>. In our analysis, we consider some variants of the censor language, which is the language used by the censor to enforce the policy. Whereas the problem is in general intractable for simple censor languages, we show that for <span><math><msub><mrow><mtext>DL-Lite</mtext></mrow><mrow><mi>R</mi></mrow></msub></math></span> ontologies it is first-order rewritable, and thus in AC<sup>0</sup> in data complexity, for the most expressive censor language we propose.</p></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0004370224001127/pdfft?md5=ee177d55b636c08d6ce8c57b16674343&pid=1-s2.0-S0004370224001127-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141594823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Artificial Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1