Pub Date : 2024-07-27DOI: 10.1007/s10472-024-09953-7
Noam Simon, Priel Levy, David Sarne
Contests are commonly used as a mechanism for eliciting effort and participation in multi-agent settings. Naturally, and much like with various other mechanisms, the information provided to the agents prior to and throughout the contest fundamentally influences its outcomes. In this paper we study the problem of information providing whenever the contest organizer does not initially hold the information and obtaining it is potentially costly. As the underlying contest mechanism for our model we use the binary contest, where contestants’ strategy is captured by their decision whether or not to participate in the contest in the first place. Here, it is often the case that the contest organizer can proactively obtain and provide contestants information related to their expected performance in the contest. We provide a comprehensive equilibrium analysis of the model, showing that even when such information is costless, it is not necessarily the case that the contest organizer will prefer to obtain and provide it to all agents, let alone when the information is costly.
{"title":"Costly information providing in binary contests","authors":"Noam Simon, Priel Levy, David Sarne","doi":"10.1007/s10472-024-09953-7","DOIUrl":"10.1007/s10472-024-09953-7","url":null,"abstract":"<div><p>Contests are commonly used as a mechanism for eliciting effort and participation in multi-agent settings. Naturally, and much like with various other mechanisms, the information provided to the agents prior to and throughout the contest fundamentally influences its outcomes. In this paper we study the problem of information providing whenever the contest organizer does not initially hold the information and obtaining it is potentially costly. As the underlying contest mechanism for our model we use the binary contest, where contestants’ strategy is captured by their decision whether or not to participate in the contest in the first place. Here, it is often the case that the contest organizer can proactively obtain and provide contestants information related to their expected performance in the contest. We provide a comprehensive equilibrium analysis of the model, showing that even when such information is costless, it is not necessarily the case that the contest organizer will prefer to obtain and provide it to all agents, let alone when the information is costly.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 5","pages":"1353 - 1375"},"PeriodicalIF":1.2,"publicationDate":"2024-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10472-024-09953-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141779355","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-19DOI: 10.1007/s10472-024-09952-8
Théo Guilbert, Olivier Caelen, Andrei Chirita, Marco Saerens
The calibration problem in machine learning classification tasks arises when a model’s output score does not align with the ground truth observed probability of the target class. There exist several parametric and non-parametric post-processing methods that can help to calibrate an existing classifier. In this work, we focus on binary classification cases where the dataset is imbalanced, meaning that the negative target class significantly outnumbers the positive one. We propose new parametric calibration methods designed to this specific case and a new calibration measure focusing on the primary objective in imbalanced problems: detecting infrequent positive cases. Experiments on several datasets show that, for imbalanced problems, our approaches outperform state-of-the-art methods in many cases.
{"title":"Calibration methods in imbalanced binary classification","authors":"Théo Guilbert, Olivier Caelen, Andrei Chirita, Marco Saerens","doi":"10.1007/s10472-024-09952-8","DOIUrl":"10.1007/s10472-024-09952-8","url":null,"abstract":"<div><p>The calibration problem in machine learning classification tasks arises when a model’s output score does not align with the ground truth observed probability of the target class. There exist several parametric and non-parametric post-processing methods that can help to calibrate an existing classifier. In this work, we focus on binary classification cases where the dataset is imbalanced, meaning that the negative target class significantly outnumbers the positive one. We propose new parametric calibration methods designed to this specific case and a new calibration measure focusing on the primary objective in imbalanced problems: detecting infrequent positive cases. Experiments on several datasets show that, for imbalanced problems, our approaches outperform state-of-the-art methods in many cases.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 5","pages":"1319 - 1352"},"PeriodicalIF":1.2,"publicationDate":"2024-07-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141745855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-10DOI: 10.1007/s10472-024-09946-6
Amit K. Chopra, Jürgen Dix, Rym Zalila-Wenkstern
{"title":"Introduction to the special issue: selected papers from EMAS 2022","authors":"Amit K. Chopra, Jürgen Dix, Rym Zalila-Wenkstern","doi":"10.1007/s10472-024-09946-6","DOIUrl":"10.1007/s10472-024-09946-6","url":null,"abstract":"","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 4","pages":"773 - 774"},"PeriodicalIF":1.2,"publicationDate":"2024-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142411128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-03DOI: 10.1007/s10472-024-09935-9
Hélène Fargier, Stefan Mengel, Jérôme Mengin
Conditional preference statements have been used to compactly represent preferences over combinatorial domains. They are at the core of CP-nets and their generalizations, and lexicographic preference trees. Several works have addressed the complexity of some queries (optimization, dominance in particular). We extend in this paper some of these results, and study other queries which have not been addressed so far, like equivalence, and transformations, like conditioning and variable elimination, thereby contributing to a knowledge compilation map for languages based on conditional preference statements. We also study the expressiveness and complexity of queries and transformations for generalized additive utilities.
{"title":"An extended knowledge compilation map for conditional preference statements-based and generalized additive utilities-based languages","authors":"Hélène Fargier, Stefan Mengel, Jérôme Mengin","doi":"10.1007/s10472-024-09935-9","DOIUrl":"10.1007/s10472-024-09935-9","url":null,"abstract":"<div><p>Conditional preference statements have been used to compactly represent preferences over combinatorial domains. They are at the core of CP-nets and their generalizations, and lexicographic preference trees. Several works have addressed the complexity of some queries (optimization, dominance in particular). We extend in this paper some of these results, and study other queries which have not been addressed so far, like equivalence, and transformations, like conditioning and variable elimination, thereby contributing to a knowledge compilation map for languages based on conditional preference statements. We also study the expressiveness and complexity of queries and transformations for generalized additive utilities.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 5","pages":"1161 - 1196"},"PeriodicalIF":1.2,"publicationDate":"2024-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141257456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-17DOI: 10.1007/s10472-024-09938-6
Jérémy Lemée, Danai Vachtsevanou, Simon Mayer, Andrei Ciortea
The ecological psychologist James J. Gibson defined the notion of affordances to refer to what action possibilities environments offer to animals. In this paper, we show how (artificial) agents can discover and exploit affordances in a Multi-Agent System (MAS) environment to achieve their goals. To indicate to agents what affordances are present in their environment and whether it is likely that these may help the agents to achieve their objectives, the environment may expose signifiers while taking into account the current situation of the environment and of the agent. On this basis, we define a Signifier Exposure Mechanism that is used by the environment to compute which signifiers should be exposed to agents in order to permit agents to only perceive information about affordances that are likely to be relevant to them, and thereby increase their interaction efficiency. If this is successful, agents can interact with partially observable environments more efficiently because the signifiers indicate the affordances they can exploit towards given purposes. Signifiers thereby facilitate the exploration and the exploitation of MAS environments. Implementations of signifiers and of the Signifier Exposure Mechanism are presented within the context of a Hypermedia Multi-Agent System, and the utility of this approach is presented through the development of a scenario.
生态心理学家詹姆斯-吉布森(James J. Gibson)定义了 "可负担性"(affordances)这一概念,指的是环境为动物提供的行动可能性。在本文中,我们将展示(人工)代理如何在多代理系统(MAS)环境中发现并利用可负担性来实现其目标。为了向代理指明其所处环境中存在哪些可负担性,以及这些可负担性是否有可能帮助代理实现其目标,环境可以在考虑到环境和代理当前情况的情况下暴露出标志物。在此基础上,我们定义了一种标识符暴露机制,由环境来计算哪些标识符应暴露给代理,以便让代理只感知可能与其相关的负担能力信息,从而提高其交互效率。如果这样做成功的话,代理就能更有效地与部分可观测环境进行交互,因为标识符指明了他们可以利用的能力,以达到特定目的。因此,标识符有助于探索和利用 MAS 环境。在超媒体多代理系统的背景下,介绍了标识符和标识符暴露机制的实施,并通过一个场景的开发介绍了这种方法的实用性。
{"title":"Signifiers for conveying and exploiting affordances: from human-computer interaction to multi-agent systems","authors":"Jérémy Lemée, Danai Vachtsevanou, Simon Mayer, Andrei Ciortea","doi":"10.1007/s10472-024-09938-6","DOIUrl":"10.1007/s10472-024-09938-6","url":null,"abstract":"<div><p>The ecological psychologist James J. Gibson defined the notion of affordances to refer to what action possibilities environments offer to animals. In this paper, we show how (artificial) agents can discover and exploit affordances in a Multi-Agent System (MAS) environment to achieve their goals. To indicate to agents what affordances are present in their environment and whether it is likely that these may help the agents to achieve their objectives, the environment may expose signifiers while taking into account the current situation of the environment and of the agent. On this basis, we define a Signifier Exposure Mechanism that is used by the environment to compute which signifiers should be exposed to agents in order to permit agents to only perceive information about affordances that are likely to be relevant to them, and thereby increase their interaction efficiency. If this is successful, agents can interact with partially observable environments more efficiently because the signifiers indicate the affordances they can exploit towards given purposes. Signifiers thereby facilitate the exploration and the exploitation of MAS environments. Implementations of signifiers and of the Signifier Exposure Mechanism are presented within the context of a Hypermedia Multi-Agent System, and the utility of this approach is presented through the development of a scenario.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 4","pages":"815 - 835"},"PeriodicalIF":1.2,"publicationDate":"2024-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10472-024-09938-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140616802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This paper describes experiments showing that some tasks in natural language processing (NLP) can already be performed using quantum computers, though so far only with small datasets. We demonstrate various approaches to topic classification. The first uses an explicit word-based approach, in which word-topic weights are implemented as fractional rotations of individual qubits, and a phrase is classified based on the accumulation of these weights onto a scoring qubit, using entangling quantum gates. This is compared with more scalable quantum encodings of word embedding vectors, which are used to compute kernel values in a quantum support vector machine: this approach achieved an average of 62% accuracy on classification tasks involving over 10000 words, which is the largest such quantum computing experiment to date. We describe a quantum probability approach to bigram modeling that can be applied to understand sequences of words and formal concepts, investigate a generative approximation to these distributions using a quantum circuit Born machine, and introduce an approach to ambiguity resolution in verb-noun composition using single-qubit rotations for simple nouns and 2-qubit entangling gates for simple verbs. The smaller systems presented have been run successfully on physical quantum computers, and the larger ones have been simulated. We show that statistically meaningful results can be obtained, but the quality of individual results varies much more using real datasets than using artificial language examples from previous quantum NLP research. Related NLP research is compared, partly with respect to contemporary challenges including informal language, fluency, and truthfulness.
{"title":"Near-term advances in quantum natural language processing","authors":"Dominic Widdows, Aaranya Alexander, Daiwei Zhu, Chase Zimmerman, Arunava Majumder","doi":"10.1007/s10472-024-09940-y","DOIUrl":"10.1007/s10472-024-09940-y","url":null,"abstract":"<div><p>This paper describes experiments showing that some tasks in natural language processing (NLP) can already be performed using quantum computers, though so far only with small datasets. We demonstrate various approaches to topic classification. The first uses an explicit word-based approach, in which word-topic weights are implemented as fractional rotations of individual qubits, and a phrase is classified based on the accumulation of these weights onto a scoring qubit, using entangling quantum gates. This is compared with more scalable quantum encodings of word embedding vectors, which are used to compute kernel values in a quantum support vector machine: this approach achieved an average of 62% accuracy on classification tasks involving over 10000 words, which is the largest such quantum computing experiment to date. We describe a quantum probability approach to bigram modeling that can be applied to understand sequences of words and formal concepts, investigate a generative approximation to these distributions using a quantum circuit Born machine, and introduce an approach to ambiguity resolution in verb-noun composition using single-qubit rotations for simple nouns and 2-qubit entangling gates for simple verbs. The smaller systems presented have been run successfully on physical quantum computers, and the larger ones have been simulated. We show that statistically meaningful results can be obtained, but the quality of individual results varies much more using real datasets than using artificial language examples from previous quantum NLP research. Related NLP research is compared, partly with respect to contemporary challenges including informal language, fluency, and truthfulness.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 5","pages":"1249 - 1272"},"PeriodicalIF":1.2,"publicationDate":"2024-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140582189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-21DOI: 10.1007/s10472-024-09939-5
Vikram Voleti, Chris Finlay, Adam Oberman, Christopher Pal
Recent work has shown that Neural Ordinary Differential Equations (ODEs) can serve as generative models of images using the perspective of Continuous Normalizing Flows (CNFs). Such models offer exact likelihood calculation, and invertible generation/density estimation. In this work we introduce a Multi-Resolution variant of such models (MRCNF), by characterizing the conditional distribution over the additional information required to generate a fine image that is consistent with the coarse image. We introduce a transformation between resolutions that allows for no change in the log likelihood. We show that this approach yields comparable likelihood values for various image datasets, with improved performance at higher resolutions, with fewer parameters, using only one GPU. Further, we examine the out-of-distribution properties of MRCNFs, and find that they are similar to those of other likelihood-based generative models.
{"title":"Multi-resolution continuous normalizing flows","authors":"Vikram Voleti, Chris Finlay, Adam Oberman, Christopher Pal","doi":"10.1007/s10472-024-09939-5","DOIUrl":"10.1007/s10472-024-09939-5","url":null,"abstract":"<div><p>Recent work has shown that Neural Ordinary Differential Equations (ODEs) can serve as generative models of images using the perspective of Continuous Normalizing Flows (CNFs). Such models offer exact likelihood calculation, and invertible generation/density estimation. In this work we introduce a Multi-Resolution variant of such models (MRCNF), by characterizing the conditional distribution over the additional information required to generate a fine image that is consistent with the coarse image. We introduce a transformation between resolutions that allows for no change in the log likelihood. We show that this approach yields comparable likelihood values for various image datasets, with improved performance at higher resolutions, with fewer parameters, using only one GPU. Further, we examine the out-of-distribution properties of MRCNFs, and find that they are similar to those of other likelihood-based generative models.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 5","pages":"1295 - 1317"},"PeriodicalIF":1.2,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140203649","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-19DOI: 10.1007/s10472-024-09929-7
L. Thorne McCarty
This paper develops a theory of clustering and coding that combines a geometric model with a probabilistic model in a principled way. The geometric model is a Riemannian manifold with a Riemannian metric, ({g}_{ij}(textbf{x})), which we interpret as a measure of dissimilarity. The probabilistic model consists of a stochastic process with an invariant probability measure that matches the density of the sample input data. The link between the two models is a potential function, (U(textbf{x})), and its gradient, (nabla U(textbf{x})). We use the gradient to define the dissimilarity metric, which guarantees that our measure of dissimilarity will depend on the probability measure. Finally, we use the dissimilarity metric to define a coordinate system on the embedded Riemannian manifold, which gives us a low-dimensional encoding of our original data.
{"title":"Clustering, coding, and the concept of similarity","authors":"L. Thorne McCarty","doi":"10.1007/s10472-024-09929-7","DOIUrl":"10.1007/s10472-024-09929-7","url":null,"abstract":"<div><p>This paper develops a theory of <i>clustering</i> and <i>coding</i> that combines a geometric model with a probabilistic model in a principled way. The geometric model is a Riemannian manifold with a Riemannian metric, <span>({g}_{ij}(textbf{x}))</span>, which we interpret as a measure of <i>dissimilarity</i>. The probabilistic model consists of a stochastic process with an invariant probability measure that matches the density of the sample input data. The link between the two models is a potential function, <span>(U(textbf{x}))</span>, and its gradient, <span>(nabla U(textbf{x}))</span>. We use the gradient to define the dissimilarity metric, which guarantees that our measure of dissimilarity will depend on the probability measure. Finally, we use the dissimilarity metric to define a coordinate system on the embedded Riemannian manifold, which gives us a low-dimensional encoding of our original data.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 5","pages":"1197 - 1248"},"PeriodicalIF":1.2,"publicationDate":"2024-03-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140172695","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}