首页 > 最新文献

Artificial Intelligence最新文献

英文 中文
A semantics for probabilistic hybrid knowledge bases with function symbols 带有函数符号的概率混合知识库的语义
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-20 DOI: 10.1016/j.artint.2025.104361
Marco Alberti , Evelina Lamma , Fabrizio Riguzzi , Riccardo Zese
Hybrid Knowledge Bases (HKBs) successfully integrate Logic Programming (LP) and Description Logics (DL) under the Minimal Knowledge with Negation as Failure semantics. Both world closure assumptions (open and closed) can be used in the same HKB, a feature required in many domains, such as the legal and health-care ones. In previous work, we proposed (function-free) Probabilistic HKBs, whose semantics applied Sato's distribution semantics approach to the well-founded HKB semantics proposed by Knorr et al. and Lyu and You. This semantics relied on the fact that the grounding of a function-free Probabilistic HKB (PHKB) is finite. In this article, we extend the PHKB language to allow function symbols, obtaining PHKBFS. Because the grounding of a PHKBFS can be infinite, we propose a novel semantics which does not require the PHKBFS's grounding to be finite. We show that the proposed semantics extends the previously proposed semantics and that, for a large class of PHKBFS, every query can be assigned a probability.
混合知识库以否定为失效语义,成功地集成了最小知识下的逻辑规划(LP)和描述逻辑(DL)。两个世界关闭假设(开放和封闭)都可以在同一个HKB中使用,这是许多领域(如法律和保健领域)所需的功能。在之前的工作中,我们提出了(无函数的)概率HKB,其语义将Sato的分布语义方法应用于Knorr等人以及Lyu和You提出的有充分根据的HKB语义。这种语义依赖于这样一个事实:无函数概率HKB (PHKB)的基础是有限的。在本文中,我们扩展PHKB语言以允许函数符号,从而获得PHKBFS。由于PHKBFS的基础可以是无限的,我们提出了一种新的语义,它不要求PHKBFS的基础是有限的。我们展示了建议的语义扩展了之前提出的语义,并且对于一个大的PHKBFS类,每个查询都可以分配一个概率。
{"title":"A semantics for probabilistic hybrid knowledge bases with function symbols","authors":"Marco Alberti ,&nbsp;Evelina Lamma ,&nbsp;Fabrizio Riguzzi ,&nbsp;Riccardo Zese","doi":"10.1016/j.artint.2025.104361","DOIUrl":"10.1016/j.artint.2025.104361","url":null,"abstract":"<div><div>Hybrid Knowledge Bases (HKBs) successfully integrate Logic Programming (LP) and Description Logics (DL) under the Minimal Knowledge with Negation as Failure semantics. Both world closure assumptions (open and closed) can be used in the same HKB, a feature required in many domains, such as the legal and health-care ones. In previous work, we proposed (function-free) Probabilistic HKBs, whose semantics applied Sato's distribution semantics approach to the well-founded HKB semantics proposed by Knorr et al. and Lyu and You. This semantics relied on the fact that the grounding of a function-free Probabilistic HKB (PHKB) is finite. In this article, we extend the PHKB language to allow function symbols, obtaining PHKB<sup><em>FS</em></sup>. Because the grounding of a PHKB<sup><em>FS</em></sup> can be infinite, we propose a novel semantics which does not require the PHKB<sup><em>FS</em></sup>'s grounding to be finite. We show that the proposed semantics extends the previously proposed semantics and that, for a large class of PHKB<sup><em>FS</em></sup>, every query can be assigned a probability.</div></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":"346 ","pages":"Article 104361"},"PeriodicalIF":5.1,"publicationDate":"2025-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144098911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Active legibility in multiagent reinforcement learning 多智能体强化学习中的主动易读性
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-19 DOI: 10.1016/j.artint.2025.104357
Yanyu Liu , Yinghui Pan , Yifeng Zeng , Biyang Ma , Prashant Doshi
A multiagent sequential decision problem has been seen in many critical applications including urban transportation, autonomous driving cars, military operations, etc. Its widely known solution, namely multiagent reinforcement learning, has evolved tremendously in recent years. Among them, the solution paradigm of modeling other agents attracts our interest, which is different from traditional value decomposition or communication mechanisms. It enables agents to understand and anticipate others' behaviors and facilitates their collaboration. Inspired by recent research on the legibility that allows agents to reveal their intentions through their behavior, we propose a multiagent active legibility framework to improve their performance. The legibility-oriented framework drives agents to conduct legible actions so as to help others optimize their behaviors. In addition, we design a series of problem domains that emulate a common legibility-needed scenario and effectively characterize the legibility in multiagent reinforcement learning. The experimental results demonstrate that the new framework is more efficient and requires less training time compared to several multiagent reinforcement learning algorithms.
多智能体顺序决策问题在城市交通、自动驾驶汽车、军事行动等许多关键应用中都有应用。它广为人知的解决方案,即多智能体强化学习,在最近几年有了巨大的发展。其中,与传统的价值分解或沟通机制不同,对其他agent建模的解决范式引起了我们的兴趣。它使代理能够理解和预测他人的行为,并促进他们的合作。受最近对易读性研究的启发,我们提出了一个多智能体主动易读性框架来提高它们的性能。以易读性为导向的框架驱动agent进行易读的行为,从而帮助他人优化自己的行为。此外,我们设计了一系列问题域,模拟常见的易读性需求场景,并有效地表征了多智能体强化学习中的易读性。实验结果表明,与几种多智能体强化学习算法相比,新框架具有更高的效率和更少的训练时间。
{"title":"Active legibility in multiagent reinforcement learning","authors":"Yanyu Liu ,&nbsp;Yinghui Pan ,&nbsp;Yifeng Zeng ,&nbsp;Biyang Ma ,&nbsp;Prashant Doshi","doi":"10.1016/j.artint.2025.104357","DOIUrl":"10.1016/j.artint.2025.104357","url":null,"abstract":"<div><div>A multiagent sequential decision problem has been seen in many critical applications including urban transportation, autonomous driving cars, military operations, etc. Its widely known solution, namely multiagent reinforcement learning, has evolved tremendously in recent years. Among them, the solution paradigm of modeling other agents attracts our interest, which is different from traditional value decomposition or communication mechanisms. It enables agents to understand and anticipate others' behaviors and facilitates their collaboration. Inspired by recent research on the legibility that allows agents to reveal their intentions through their behavior, we propose a <em>multiagent active legibility framework</em> to improve their performance. The legibility-oriented framework drives agents to conduct legible actions so as to help others optimize their behaviors. In addition, we design a series of problem domains that emulate a common legibility-needed scenario and effectively characterize the legibility in multiagent reinforcement learning. The experimental results demonstrate that the new framework is more efficient and requires less training time compared to several multiagent reinforcement learning algorithms.</div></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":"346 ","pages":"Article 104357"},"PeriodicalIF":5.1,"publicationDate":"2025-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144098910","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A theory of synaptic neural balance: From local to global order 突触神经平衡理论:从局部秩序到全局秩序
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-16 DOI: 10.1016/j.artint.2025.104360
Pierre Baldi, Antonios Alexos, Ian Domingo, Alireza Rahmansetayesh
We develop a general theory of synaptic neural balance and how it can emerge or be enforced in neural networks. For a given additive cost function R (regularizer), a neuron is said to be in balance if the total cost of its input weights is equal to the total cost of its output weights. The basic example is provided by feedforward networks of ReLU units trained with L2 regularizers, which exhibit balance after proper training. The theory explains this phenomenon and extends it in several directions. The first direction is the extension to bilinear and other activation functions. The second direction is the extension to more general regularizers, including all Lp (p>0) regularizers. The third direction is the extension to non-layered architectures, recurrent architectures, convolutional architectures, as well as architectures with mixed activation functions and to different balancing algorithms. Gradient descent on the error function alone does not converge in general to a balanced state, where every neuron is in balance, even when starting from a balanced state. However, gradient descent on the regularized error function ought to converge to a balanced state, and thus network balance can be used to assess learning progress. The theory is based on two local neuronal operations: scaling which is commutative, and balancing which is not commutative. Finally, and most importantly, given any set of weights, when local balancing operations are applied to each neuron in a stochastic manner, global order always emerges through the convergence of the stochastic balancing algorithm to the same unique set of balanced weights. The reason for this convergence is the existence of an underlying strictly convex optimization problem where the relevant variables are constrained to a linear, only architecture-dependent, manifold. Simulations show that balancing neurons prior to learning, or during learning in alternation with gradient descent steps, can improve learning speed and performance thereby expanding the arsenal of available training tools. Scaling and balancing operations are entirely local and thus physically plausible in biological and neuromorphic neural networks.
我们发展了突触神经平衡的一般理论,以及它如何在神经网络中出现或被强制执行。对于给定的加性代价函数R(正则化器),如果一个神经元的输入权值的总代价等于输出权值的总代价,则该神经元处于平衡状态。用L2正则化器训练的ReLU单元前馈网络提供了一个基本的例子,经过适当的训练后,ReLU单元表现出平衡。该理论解释了这一现象,并在几个方向上进行了扩展。第一个方向是对双线性和其他激活函数的扩展。第二个方向是扩展到更一般的正则子,包括所有Lp (p>0)正则子。第三个方向是扩展到非分层架构、循环架构、卷积架构、混合激活函数架构和不同的平衡算法。单独的误差函数的梯度下降通常不会收敛到平衡状态,在平衡状态下,每个神经元都处于平衡状态,即使从平衡状态开始。然而,正则化误差函数上的梯度下降应该收敛到平衡状态,因此网络平衡可以用来评估学习进度。该理论基于两种局部神经元操作:伸缩是可交换的,平衡是不可交换的。最后,也是最重要的是,给定任何一组权值,当以随机方式对每个神经元应用局部平衡操作时,随机平衡算法总是通过收敛到相同的唯一的平衡权值集而出现全局顺序。这种收敛的原因是存在一个潜在的严格凸优化问题,其中相关变量被约束为线性的,仅与体系结构相关的流形。仿真表明,在学习前或学习过程中,通过梯度下降步骤交替平衡神经元,可以提高学习速度和性能,从而扩展可用训练工具的武器库。缩放和平衡操作完全是局部的,因此在生物和神经形态神经网络中物理上是可行的。
{"title":"A theory of synaptic neural balance: From local to global order","authors":"Pierre Baldi,&nbsp;Antonios Alexos,&nbsp;Ian Domingo,&nbsp;Alireza Rahmansetayesh","doi":"10.1016/j.artint.2025.104360","DOIUrl":"10.1016/j.artint.2025.104360","url":null,"abstract":"<div><div>We develop a general theory of synaptic neural balance and how it can emerge or be enforced in neural networks. For a given additive cost function <em>R</em> (regularizer), a neuron is said to be in balance if the total cost of its input weights is equal to the total cost of its output weights. The basic example is provided by feedforward networks of ReLU units trained with <span><math><msub><mrow><mi>L</mi></mrow><mrow><mn>2</mn></mrow></msub></math></span> regularizers, which exhibit balance after proper training. The theory explains this phenomenon and extends it in several directions. The first direction is the extension to bilinear and other activation functions. The second direction is the extension to more general regularizers, including all <span><math><msub><mrow><mi>L</mi></mrow><mrow><mi>p</mi></mrow></msub></math></span> (<span><math><mi>p</mi><mo>&gt;</mo><mn>0</mn></math></span>) regularizers. The third direction is the extension to non-layered architectures, recurrent architectures, convolutional architectures, as well as architectures with mixed activation functions and to different balancing algorithms. Gradient descent on the error function alone does not converge in general to a balanced state, where every neuron is in balance, even when starting from a balanced state. However, gradient descent on the regularized error function ought to converge to a balanced state, and thus network balance can be used to assess learning progress. The theory is based on two local neuronal operations: scaling which is commutative, and balancing which is not commutative. Finally, and most importantly, given any set of weights, when local balancing operations are applied to each neuron in a stochastic manner, global order always emerges through the convergence of the stochastic balancing algorithm to the same unique set of balanced weights. The reason for this convergence is the existence of an underlying strictly convex optimization problem where the relevant variables are constrained to a linear, only architecture-dependent, manifold. Simulations show that balancing neurons prior to learning, or during learning in alternation with gradient descent steps, can improve learning speed and performance thereby expanding the arsenal of available training tools. Scaling and balancing operations are entirely local and thus physically plausible in biological and neuromorphic neural networks.</div></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":"346 ","pages":"Article 104360"},"PeriodicalIF":5.1,"publicationDate":"2025-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144106474","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RelBERT: Embedding relations with language models 用语言模型嵌入关系
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-15 DOI: 10.1016/j.artint.2025.104359
Asahi Ushio, Jose Camacho-Collados, Steven Schockaert
Many applications need access to background knowledge about how different concepts and entities are related. Although Large Language Models (LLM) can address this need to some extent, LLMs are inefficient and difficult to control. As an alternative, we propose to extract relation embeddings from relatively small language models. In particular, we show that masked language models such as RoBERTa can be straightforwardly fine-tuned for this purpose, using only a small amount of training data. The resulting model, which we call RelBERT, captures relational similarity in a surprisingly fine-grained way, allowing us to set a new state-of-the-art in analogy benchmarks. Crucially, RelBERT is capable of modelling relations that go well beyond what the model has seen during training. For instance, we obtained strong results on relations between named entities with a model that was only trained on lexical relations between concepts, and we observed that RelBERT can recognise morphological analogies despite not being trained on such examples. Overall, we find that RelBERT significantly outperforms strategies based on prompting language models that are several orders of magnitude larger, including recent GPT-based models and open source models.1
许多应用程序需要访问有关不同概念和实体如何相关的背景知识。尽管大型语言模型(LLM)可以在一定程度上解决这一需求,但LLM效率低下且难以控制。作为替代方案,我们建议从相对较小的语言模型中提取关系嵌入。特别地,我们展示了像RoBERTa这样的屏蔽语言模型可以直接为此目的进行微调,只使用少量的训练数据。由此产生的模型,我们称之为RelBERT,以一种令人惊讶的细粒度方式捕获关系相似性,使我们能够在类比基准中设置新的最先进的技术。至关重要的是,RelBERT能够对远远超出模型在训练期间所看到的关系进行建模。例如,我们使用一个只在概念之间的词汇关系上训练的模型,在命名实体之间的关系上获得了强有力的结果,我们观察到RelBERT可以识别形态类比,尽管没有在这样的例子上训练。总的来说,我们发现RelBERT显著优于基于提示语言模型的策略,这些模型要大几个数量级,包括最近的基于gpt的模型和开源模型
{"title":"RelBERT: Embedding relations with language models","authors":"Asahi Ushio,&nbsp;Jose Camacho-Collados,&nbsp;Steven Schockaert","doi":"10.1016/j.artint.2025.104359","DOIUrl":"10.1016/j.artint.2025.104359","url":null,"abstract":"<div><div>Many applications need access to background knowledge about how different concepts and entities are related. Although Large Language Models (LLM) can address this need to some extent, LLMs are inefficient and difficult to control. As an alternative, we propose to extract relation embeddings from relatively small language models. In particular, we show that masked language models such as RoBERTa can be straightforwardly fine-tuned for this purpose, using only a small amount of training data. The resulting model, which we call RelBERT, captures relational similarity in a surprisingly fine-grained way, allowing us to set a new state-of-the-art in analogy benchmarks. Crucially, RelBERT is capable of modelling relations that go well beyond what the model has seen during training. For instance, we obtained strong results on relations between named entities with a model that was only trained on lexical relations between concepts, and we observed that RelBERT can recognise morphological analogies despite not being trained on such examples. Overall, we find that RelBERT significantly outperforms strategies based on prompting language models that are several orders of magnitude larger, including recent GPT-based models and open source models.<span><span><sup>1</sup></span></span></div></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":"347 ","pages":"Article 104359"},"PeriodicalIF":5.1,"publicationDate":"2025-05-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144190068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CBS-Budget (CBSB): A complete and bounded suboptimal search for multi-agent path finding CBS-Budget (CBSB):用于多智能体路径查找的完整和有界次优搜索
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-08 DOI: 10.1016/j.artint.2025.104349
Jaein Lim , Panagiotis Tsiotras
Multi-Agent Path Finding (MAPF) is the problem of finding a collection of conflict-free paths for a team of multiple agents while minimizing some global cost, such as the sum of the travel time of all agents, or the travel time of the last agent. Conflict Based Search (CBS) is a leading complete and optimal MAPF algorithm that lazily explores the joint agent state space, using an admissible heuristic joint plan. Such an admissible heuristic joint plan is computed by combining individual shortest paths computed without considering inter-agent conflicts, and becoming gradually more informed as constraints are added to the individual agents' path-planning problems to avoid discovered conflicts. In this paper, we seek to speed up CBS by finding a more informed heuristic joint plan that is bounded. We first propose the budgeted Class-Ordered A* (bCOA*), a novel algorithm that finds the least-cost path with the minimal number of conflicts that is upper bounded in terms of path length. Then, we propose a novel bounded-cost variant of CBS, called CBS-Budget (CBSB) by using bCOA* search at the low-level search of the CBS and by using a modified focal search at the high-level search of the CBS. We prove that CBSB is complete and bounded-suboptimal. In our numerical experiments, CBSB finds a near-optimal solution for hundreds of agents within a fraction of a second. CBSB shows state-of-the-art performance, comparable to Explicit Estimation CBS (EECBS), an enhanced recent version of CBS. On the other hand, CBSB is much easier to implement than EECBS, since only one priority queue at the low-level search is needed, as in CBS, and only two priority queues at the high-level search are needed, as in Enhanced CBS (ECBS).
多代理寻路(Multi-Agent Path Finding, MAPF)的问题是为多个代理组成的团队找到一组无冲突的路径,同时最小化一些全局成本,比如所有代理的旅行时间之和,或者最后一个代理的旅行时间。基于冲突的搜索(CBS)是一种领先的完备最优MAPF算法,它使用可接受的启发式联合计划惰性地探索联合代理状态空间。这种可接受的启发式联合规划是在不考虑智能体间冲突的情况下,将计算得到的单个最短路径组合在一起计算得到的,并随着对单个智能体路径规划问题的约束的增加而逐渐变得更加知情,以避免发现冲突。在本文中,我们寻求通过寻找一个更知情的有界启发式联合计划来加速CBS。我们首先提出了预算类有序A* (bCOA*)算法,这是一种新的算法,可以找到具有最小冲突数的最小代价路径,该路径在路径长度上是上界的。在此基础上,提出了一种新的有界成本CBS算法,即CBS- budget (CBSB)算法,该算法在CBS的低阶搜索中使用bCOA*搜索,在CBS的高阶搜索中使用改进的焦点搜索。证明了CBSB是完全的、有界次优的。在我们的数值实验中,CBSB在几分之一秒内为数百个代理找到了近乎最佳的解决方案。CBSB显示了最先进的性能,可与显式估计CBS (EECBS)相媲美,后者是CBS的最新增强版本。另一方面,CBSB比EECBS更容易实现,因为在低级搜索中只需要一个优先级队列(如CBS),在高级搜索中只需要两个优先级队列(如增强型CBS (ECBS))。
{"title":"CBS-Budget (CBSB): A complete and bounded suboptimal search for multi-agent path finding","authors":"Jaein Lim ,&nbsp;Panagiotis Tsiotras","doi":"10.1016/j.artint.2025.104349","DOIUrl":"10.1016/j.artint.2025.104349","url":null,"abstract":"<div><div>Multi-Agent Path Finding (MAPF) is the problem of finding a collection of conflict-free paths for a team of multiple agents while minimizing some global cost, such as the sum of the travel time of all agents, or the travel time of the last agent. Conflict Based Search (CBS) is a leading complete and optimal MAPF algorithm that lazily explores the joint agent state space, using an admissible heuristic joint plan. Such an admissible heuristic joint plan is computed by combining individual shortest paths computed without considering inter-agent conflicts, and becoming gradually more informed as constraints are added to the individual agents' path-planning problems to avoid discovered conflicts. In this paper, we seek to speed up CBS by finding a more informed heuristic joint plan that is bounded. We first propose the budgeted Class-Ordered A* (bCOA*), a novel algorithm that finds the least-cost path with the minimal number of conflicts that is upper bounded in terms of path length. Then, we propose a novel bounded-cost variant of CBS, called CBS-Budget (CBSB) by using bCOA* search at the low-level search of the CBS and by using a modified focal search at the high-level search of the CBS. We prove that CBSB is complete and bounded-suboptimal. In our numerical experiments, CBSB finds a near-optimal solution for hundreds of agents within a fraction of a second. CBSB shows state-of-the-art performance, comparable to Explicit Estimation CBS (EECBS), an enhanced recent version of CBS. On the other hand, CBSB is much easier to implement than EECBS, since only one priority queue at the low-level search is needed, as in CBS, and only two priority queues at the high-level search are needed, as in Enhanced CBS (ECBS).</div></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":"346 ","pages":"Article 104349"},"PeriodicalIF":5.1,"publicationDate":"2025-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144067939","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient and effective budget-feasible mechanisms for submodular valuations 分模块估值的高效率和有效的预算可行机制
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-07 DOI: 10.1016/j.artint.2025.104348
Kai Han , Haotian Zhang , Shuang Cui
We revisit the classical problem of designing Budget-Feasible Mechanisms (BFMs) for submodular valuation functions, which has been extensively studied since the seminal paper of Singer [FOCS'10] due to their wide applications in crowdsourcing and social marketing. We propose TripleEagle, a novel algorithmic framework for designing BFMs, based on which we present several simple yet effective BFMs that achieve better approximation ratios than the state-of-the-art work. Moreover, our BFMs are the first in the literature to achieve linear query complexity under the value oracle model while ensuring obvious strategyproofness, making them more practical than the previous BFMs. We conduct extensive experiments to evaluate the empirical performance of our BFMs, and the experimental results demonstrate the superiorities of our approach in terms of efficiency and effectiveness compared to the state-of-the-art BFMs.
我们重新审视为子模块估值函数设计预算可行机制(BFMs)的经典问题,自Singer的开创性论文[FOCS'10]以来,由于其在众包和社会营销中的广泛应用,该问题得到了广泛研究。我们提出了TripleEagle,一个设计bfm的新算法框架,在此基础上,我们提出了几个简单而有效的bfm,它们比最先进的工作实现了更好的近似比。此外,我们的bfm是文献中第一个在值oracle模型下实现线性查询复杂度的,同时保证了明显的策略证明性,使它们比以前的bfm更实用。我们进行了大量的实验来评估我们的bfm的经验性能,实验结果表明,与最先进的bfm相比,我们的方法在效率和有效性方面具有优势。
{"title":"Efficient and effective budget-feasible mechanisms for submodular valuations","authors":"Kai Han ,&nbsp;Haotian Zhang ,&nbsp;Shuang Cui","doi":"10.1016/j.artint.2025.104348","DOIUrl":"10.1016/j.artint.2025.104348","url":null,"abstract":"<div><div>We revisit the classical problem of designing Budget-Feasible Mechanisms (BFMs) for submodular valuation functions, which has been extensively studied since the seminal paper of Singer [FOCS'10] due to their wide applications in crowdsourcing and social marketing. We propose <span><math><mi>TripleEagle</mi></math></span>, a novel algorithmic framework for designing BFMs, based on which we present several simple yet effective BFMs that achieve better approximation ratios than the state-of-the-art work. Moreover, our BFMs are the first in the literature to achieve linear query complexity under the value oracle model while ensuring obvious strategyproofness, making them more practical than the previous BFMs. We conduct extensive experiments to evaluate the empirical performance of our BFMs, and the experimental results demonstrate the superiorities of our approach in terms of efficiency and effectiveness compared to the state-of-the-art BFMs.</div></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":"345 ","pages":"Article 104348"},"PeriodicalIF":5.1,"publicationDate":"2025-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143921715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep optimal transport for domain adaptation on SPD manifolds SPD流形域自适应的深度最优输运
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-02 DOI: 10.1016/j.artint.2025.104347
Ce Ju , Cuntai Guan
Recent progress in geometric deep learning has drawn increasing attention from the machine learning community toward domain adaptation on symmetric positive definite (SPD) manifolds—especially for neuroimaging data that often suffer from distribution shifts across sessions. These data, typically represented as covariance matrices of brain signals, inherently lie on SPD manifolds due to their symmetry and positive definiteness. However, conventional domain adaptation methods often overlook this geometric structure when applied directly to covariance matrices, which can result in suboptimal performance. To address this issue, we introduce a new geometric deep learning framework that combines optimal transport theory with the geometry of SPD manifolds. Our approach aligns data distributions while respecting the manifold structure, effectively reducing both marginal and conditional discrepancies. We validate our method on three cross-session brain-computer interface datasets—KU, BNCI2014001, and BNCI2015001—where it consistently outperforms baseline approaches while maintaining the intrinsic geometry of the data. We also provide quantitative results and visualizations to better illustrate the behavior of the learned embeddings.
几何深度学习的最新进展引起了机器学习社区对对称正定流形(SPD)的域适应的越来越多的关注,特别是对于经常遭受分布变化的神经成像数据。这些数据通常表示为大脑信号的协方差矩阵,由于其对称性和正确定性,它们固有地位于SPD流形上。然而,传统的域自适应方法在直接应用于协方差矩阵时往往忽略了这种几何结构,从而导致性能不理想。为了解决这个问题,我们引入了一个新的几何深度学习框架,该框架结合了最优输运理论和SPD流形的几何特征。我们的方法在尊重流形结构的同时对齐数据分布,有效地减少了边际和条件差异。我们在三个跨会话脑机接口数据集(ku、BNCI2014001和bnci2015001)上验证了我们的方法,在保持数据固有几何形状的同时,它始终优于基线方法。我们还提供了定量结果和可视化,以更好地说明学习嵌入的行为。
{"title":"Deep optimal transport for domain adaptation on SPD manifolds","authors":"Ce Ju ,&nbsp;Cuntai Guan","doi":"10.1016/j.artint.2025.104347","DOIUrl":"10.1016/j.artint.2025.104347","url":null,"abstract":"<div><div>Recent progress in geometric deep learning has drawn increasing attention from the machine learning community toward domain adaptation on symmetric positive definite (SPD) manifolds—especially for neuroimaging data that often suffer from distribution shifts across sessions. These data, typically represented as covariance matrices of brain signals, inherently lie on SPD manifolds due to their symmetry and positive definiteness. However, conventional domain adaptation methods often overlook this geometric structure when applied directly to covariance matrices, which can result in suboptimal performance. To address this issue, we introduce a new geometric deep learning framework that combines optimal transport theory with the geometry of SPD manifolds. Our approach aligns data distributions while respecting the manifold structure, effectively reducing both marginal and conditional discrepancies. We validate our method on three cross-session brain-computer interface datasets—KU, BNCI2014001, and BNCI2015001—where it consistently outperforms baseline approaches while maintaining the intrinsic geometry of the data. We also provide quantitative results and visualizations to better illustrate the behavior of the learned embeddings.</div></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":"345 ","pages":"Article 104347"},"PeriodicalIF":5.1,"publicationDate":"2025-05-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143906599","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Disjoint projected enumeration for SAT and SMT without blocking clauses 不带阻塞子句的SAT和SMT的不相交投影枚举
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-29 DOI: 10.1016/j.artint.2025.104346
Giuseppe Spallitta , Roberto Sebastiani , Armin Biere
All-Solution Satisfiability (AllSAT) and its extension, All-Solution Satisfiability Modulo Theories (AllSMT), have become more relevant in recent years, mainly in formal verification and artificial intelligence applications. The goal of these problems is the enumeration of all satisfying assignments of a formula (for SAT and SMT problems, respectively), making them useful for test generation, model checking, and probabilistic inference. Nevertheless, traditional AllSAT algorithms face significant computational challenges due to the exponential growth of the search space and inefficiencies caused by blocking clauses, which cause memory blowups and degrade unit propagation performance in the long term. This paper presents two novel solvers: TabularAllSAT, a projected AllSAT solver, and TabularAllSMT, a projected AllSMT solver. Both solvers combine Conflict-Driven Clause Learning (CDCL) with chronological backtracking to improve efficiency while ensuring disjoint enumeration. To retrieve compact partial assignments we propose a novel aggressive implicant shrinking algorithm, compatible with chronological backtracking, to minimize the number of partial assignments, reducing overall search complexity. Furthermore, we extend the solver framework to handle projected enumeration and SMT formulas effectively and efficiently, adapting the baseline framework to integrate theory reasoning and the distinction between important and non-important variables. An extensive experimental evaluation demonstrates the superiority of our approach compared to state-of-the-art solvers, particularly in scenarios requiring projection and SMT-based reasoning.
近年来,全解可满足性(AllSAT)及其扩展全解可满足模理论(AllSMT)在形式验证和人工智能应用中变得越来越重要。这些问题的目标是枚举一个公式的所有令人满意的赋值(分别针对SAT和SMT问题),使它们对测试生成、模型检查和概率推断有用。然而,由于搜索空间的指数增长和阻塞子句导致的低效率,传统的AllSAT算法面临着巨大的计算挑战,阻塞子句会导致内存爆炸并降低长期的单元传播性能。本文提出了两个新的求解器:TabularAllSAT和TabularAllSMT。这两种解决方案都将冲突驱动子句学习(CDCL)与时间回溯相结合,以提高效率,同时确保不连接枚举。为了检索紧凑的部分分配,我们提出了一种新的兼容时间回溯的主动隐含收缩算法,以最小化部分分配的数量,降低整体搜索复杂度。此外,我们扩展了求解器框架,以有效地处理投影枚举和SMT公式,并采用基线框架来整合理论推理以及重要变量和非重要变量的区分。广泛的实验评估表明,与最先进的求解器相比,我们的方法具有优势,特别是在需要投影和基于smt的推理的场景中。
{"title":"Disjoint projected enumeration for SAT and SMT without blocking clauses","authors":"Giuseppe Spallitta ,&nbsp;Roberto Sebastiani ,&nbsp;Armin Biere","doi":"10.1016/j.artint.2025.104346","DOIUrl":"10.1016/j.artint.2025.104346","url":null,"abstract":"<div><div>All-Solution Satisfiability (AllSAT) and its extension, All-Solution Satisfiability Modulo Theories (AllSMT), have become more relevant in recent years, mainly in formal verification and artificial intelligence applications. The goal of these problems is the enumeration of all satisfying assignments of a formula (for SAT and SMT problems, respectively), making them useful for test generation, model checking, and probabilistic inference. Nevertheless, traditional AllSAT algorithms face significant computational challenges due to the exponential growth of the search space and inefficiencies caused by blocking clauses, which cause memory blowups and degrade unit propagation performance in the long term. This paper presents two novel solvers: <span>TabularAllSAT</span>, a projected AllSAT solver, and <span>TabularAllSMT</span>, a projected AllSMT solver. Both solvers combine Conflict-Driven Clause Learning (CDCL) with chronological backtracking to improve efficiency while ensuring disjoint enumeration. To retrieve compact partial assignments we propose a novel aggressive implicant shrinking algorithm, compatible with chronological backtracking, to minimize the number of partial assignments, reducing overall search complexity. Furthermore, we extend the solver framework to handle projected enumeration and SMT formulas effectively and efficiently, adapting the baseline framework to integrate theory reasoning and the distinction between important and non-important variables. An extensive experimental evaluation demonstrates the superiority of our approach compared to state-of-the-art solvers, particularly in scenarios requiring projection and SMT-based reasoning.</div></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":"345 ","pages":"Article 104346"},"PeriodicalIF":5.1,"publicationDate":"2025-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143911498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FedHM: Efficient federated learning for heterogeneous models via low-rank factorization FedHM:通过低秩分解对异构模型进行有效的联邦学习
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-25 DOI: 10.1016/j.artint.2025.104333
Dezhong Yao , Wanning Pan , Yuexin Shi , Michael J. O'Neill , Yutong Dai , Yao Wan , Peilin Zhao , Hai Jin , Lichao Sun
One underlying assumption of recent Federated Learning (FL) paradigms is that all local models share an identical network architecture. However, this assumption is inefficient for heterogeneous systems where devices possess varying computation and communication capabilities. The presence of such heterogeneity among devices negatively impacts the scalability of FL and slows down the training process due to the existence of stragglers. To this end, this paper proposes a novel federated compression framework for heterogeneous models, named FedHM, distributing the heterogeneous low-rank models to clients and then aggregating them into a full-rank global model. Furthermore, FedHM significantly reduces communication costs by utilizing low-rank models. Compared with state-of-the-art heterogeneous FL methods under various FL settings, FedHM is superior in the performance and robustness of models with different sizes. Additionally, the convergence guarantee of FL for heterogeneous devices is first theoretically analyzed.
最近联邦学习(FL)范例的一个基本假设是,所有本地模型共享相同的网络体系结构。然而,这种假设对于异构系统来说是低效的,在异构系统中,设备具有不同的计算和通信能力。设备之间的这种异质性的存在对FL的可扩展性产生了负面影响,并且由于离散体的存在而减慢了训练过程。为此,本文提出了一种新的异构模型联邦压缩框架FedHM,将异构低秩模型分发到客户端,再聚合成全秩全局模型。此外,FedHM通过使用低秩模型显著降低了通信成本。与目前最先进的异构FL方法相比,在不同FL设置下,FedHM在不同尺寸模型的性能和鲁棒性方面都具有优势。此外,本文还从理论上分析了异质器件下FL的收敛性保证。
{"title":"FedHM: Efficient federated learning for heterogeneous models via low-rank factorization","authors":"Dezhong Yao ,&nbsp;Wanning Pan ,&nbsp;Yuexin Shi ,&nbsp;Michael J. O'Neill ,&nbsp;Yutong Dai ,&nbsp;Yao Wan ,&nbsp;Peilin Zhao ,&nbsp;Hai Jin ,&nbsp;Lichao Sun","doi":"10.1016/j.artint.2025.104333","DOIUrl":"10.1016/j.artint.2025.104333","url":null,"abstract":"<div><div>One underlying assumption of recent <em>Federated Learning</em> (FL) paradigms is that all local models share an identical network architecture. However, this assumption is inefficient for heterogeneous systems where devices possess varying computation and communication capabilities. The presence of such heterogeneity among devices negatively impacts the scalability of FL and slows down the training process due to the existence of stragglers. To this end, this paper proposes a novel <em>federated compression framework for heterogeneous models</em>, named FedHM, distributing the heterogeneous low-rank models to clients and then aggregating them into a full-rank global model. Furthermore, FedHM significantly reduces communication costs by utilizing low-rank models. Compared with state-of-the-art heterogeneous FL methods under various FL settings, FedHM is superior in the performance and robustness of models with different sizes. Additionally, the convergence guarantee of FL for heterogeneous devices is first theoretically analyzed.</div></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":"344 ","pages":"Article 104333"},"PeriodicalIF":5.1,"publicationDate":"2025-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143881427","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Effective and fast module extraction for nonempty ABoxes 有效和快速的模块提取非空框
IF 5.1 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-24 DOI: 10.1016/j.artint.2025.104345
Piero Andrea Bonatti , Francesco Magliocca , Iliana Mineva Petrova , Luigi Sauro
A deductive module of a knowledge base KB is a subset of KB that preserves a specified class of consequences. Module extraction is applied in ontology design, debugging, and reasoning. The locality-based module extractors of the OWL API are less effective when the knowledge base contains facts such as ABox assertions. The competing module extractor PrisM computes smaller modules at the cost of higher computation time. In this paper, we introduce and study a novel module extraction technique, called conditional module extraction, that can be applied to satisfiable SRIQ(D) knowledge bases. Experimental analysis shows that conditional module extraction constitutes an appealing alternative to PrisM and to the locality-based extractors of the OWL API, when the ABox is nonempty.
知识库KB的演绎模块是知识库的一个子集,它保留了指定的结果类。将模块提取应用于本体设计、调试和推理。当知识库包含诸如ABox断言之类的事实时,OWL API的基于位置的模块提取器的效果较差。与之竞争的模块提取器PrisM以更高的计算时间为代价计算更小的模块。本文介绍并研究了一种新的模块提取技术,即条件模块提取,它可以应用于可满足的SRIQ(D)知识库。实验分析表明,当ABox为非空时,条件模块提取可以替代PrisM和OWL API的基于位置的提取器。
{"title":"Effective and fast module extraction for nonempty ABoxes","authors":"Piero Andrea Bonatti ,&nbsp;Francesco Magliocca ,&nbsp;Iliana Mineva Petrova ,&nbsp;Luigi Sauro","doi":"10.1016/j.artint.2025.104345","DOIUrl":"10.1016/j.artint.2025.104345","url":null,"abstract":"<div><div>A deductive module of a knowledge base <span><math><mi>KB</mi></math></span> is a subset of <span><math><mi>KB</mi></math></span> that preserves a specified class of consequences. Module extraction is applied in ontology design, debugging, and reasoning. The locality-based module extractors of the OWL API are less effective when the knowledge base contains facts such as ABox assertions. The competing module extractor PrisM computes smaller modules at the cost of higher computation time. In this paper, we introduce and study a novel module extraction technique, called <em>conditional module extraction</em>, that can be applied to satisfiable <span><math><mrow><mi>SRIQ</mi></mrow><mo>(</mo><mi>D</mi><mo>)</mo></math></span> knowledge bases. Experimental analysis shows that conditional module extraction constitutes an appealing alternative to PrisM and to the locality-based extractors of the OWL API, when the ABox is nonempty.</div></div>","PeriodicalId":8434,"journal":{"name":"Artificial Intelligence","volume":"344 ","pages":"Article 104345"},"PeriodicalIF":5.1,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143881426","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Artificial Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1