Pub Date : 2024-05-21DOI: 10.1016/j.ijar.2024.109215
Bich Khue Vo , Hung Son Nguyen
Feature importance is a complex issue in machine learning, as determining a superior attribute is vague, uncertain, and dependent on the model. This study introduces a rough-fuzzy hybrid (RAFAR) method that merges various techniques from rough set theory and fuzzy set theory to tackle uncertainty in attribute importance and ranking. RAFAR utilizes an interval-valued fuzzy matrix to depict preference between attribute pairs. This research focuses on constructing these matrices from datasets and identifying suitable rankings based on these matrices. The concept of interval-valued weight vectors is introduced to represent attribute importance, and their additive and multiplicative compatibility is examined. The properties of these consistency types and the efficient algorithms for solving related problems are discussed. These new theoretical findings are valuable for creating effective optimization models and algorithms within the RAFAR framework. Additionally, novel approaches for constructing pairwise comparison matrices and enhancing the scalability of RAFAR are suggested. The study also includes experimental results on benchmark datasets to demonstrate the accuracy of the proposed solutions.
{"title":"An attribute ranking method based on rough sets and interval-valued fuzzy sets","authors":"Bich Khue Vo , Hung Son Nguyen","doi":"10.1016/j.ijar.2024.109215","DOIUrl":"https://doi.org/10.1016/j.ijar.2024.109215","url":null,"abstract":"<div><p>Feature importance is a complex issue in machine learning, as determining a superior attribute is vague, uncertain, and dependent on the model. This study introduces a rough-fuzzy hybrid (RAFAR) method that merges various techniques from rough set theory and fuzzy set theory to tackle uncertainty in attribute importance and ranking. RAFAR utilizes an interval-valued fuzzy matrix to depict preference between attribute pairs. This research focuses on constructing these matrices from datasets and identifying suitable rankings based on these matrices. The concept of interval-valued weight vectors is introduced to represent attribute importance, and their additive and multiplicative compatibility is examined. The properties of these consistency types and the efficient algorithms for solving related problems are discussed. These new theoretical findings are valuable for creating effective optimization models and algorithms within the RAFAR framework. Additionally, novel approaches for constructing pairwise comparison matrices and enhancing the scalability of RAFAR are suggested. The study also includes experimental results on benchmark datasets to demonstrate the accuracy of the proposed solutions.</p></div>","PeriodicalId":13842,"journal":{"name":"International Journal of Approximate Reasoning","volume":"170 ","pages":"Article 109215"},"PeriodicalIF":3.9,"publicationDate":"2024-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141083329","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-18DOI: 10.1016/j.ijar.2024.109212
Ekaterina Kubyshkina, Giuseppe Primiero
The notion of trustworthiness, central to many fields of human inquiry, has recently attracted the attention of various researchers in logic, computer science, and artificial intelligence (AI). Both conceptual and formal approaches for modeling trustworthiness as a (desirable) property of AI systems are emerging in the literature. To develop logics fit for this aim means to analyze both the non-deterministic aspect of AI systems and to offer a formalization of the intended meaning of their trustworthiness. In this work we take a semantic perspective on representing such processes, and provide a measure on possible worlds for evaluating them as trustworthy. In particular, we intend trustworthiness as the correspondence within acceptable limits between a model in which the theoretical probability of a process to produce a given output is expressed and a model in which the frequency of showing such output as established during a relevant number of tests is measured. From a technical perspective, we show that our semantics characterizes the probabilistic typed natural deduction calculus introduced in D'Asaro and Primiero (2021)[12] and further extended in D'Asaro et al. (2023) [13]. This contribution connects those results on trustworthy probabilistic processes with the mainstream method in modal logic, thereby facilitating the understanding of this field of research for a larger audience of logicians, as well as setting the stage for an epistemic logic appropriate to the task.
{"title":"A possible worlds semantics for trustworthy non-deterministic computations","authors":"Ekaterina Kubyshkina, Giuseppe Primiero","doi":"10.1016/j.ijar.2024.109212","DOIUrl":"10.1016/j.ijar.2024.109212","url":null,"abstract":"<div><p>The notion of trustworthiness, central to many fields of human inquiry, has recently attracted the attention of various researchers in logic, computer science, and artificial intelligence (AI). Both conceptual and formal approaches for modeling trustworthiness as a (desirable) property of AI systems are emerging in the literature. To develop logics fit for this aim means to analyze both the non-deterministic aspect of AI systems and to offer a formalization of the intended meaning of their trustworthiness. In this work we take a semantic perspective on representing such processes, and provide a measure on possible worlds for evaluating them as trustworthy. In particular, we intend trustworthiness as the correspondence within acceptable limits between a model in which the theoretical probability of a process to produce a given output is expressed and a model in which the frequency of showing such output as established during a relevant number of tests is measured. From a technical perspective, we show that our semantics characterizes the probabilistic typed natural deduction calculus introduced in D'Asaro and Primiero (2021)<span>[12]</span> and further extended in D'Asaro et al. (2023) <span>[13]</span>. This contribution connects those results on trustworthy probabilistic processes with the mainstream method in modal logic, thereby facilitating the understanding of this field of research for a larger audience of logicians, as well as setting the stage for an epistemic logic appropriate to the task.</p></div>","PeriodicalId":13842,"journal":{"name":"International Journal of Approximate Reasoning","volume":"172 ","pages":"Article 109212"},"PeriodicalIF":3.9,"publicationDate":"2024-05-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0888613X24000999/pdfft?md5=7a0c991c70c70e79ac2349285a1a28c0&pid=1-s2.0-S0888613X24000999-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141143080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-16DOI: 10.1016/j.ijar.2024.109213
Floris Persiau, Gert de Cooman
In a prequential approach to algorithmic randomness, probabilities for the next outcome can be forecast ‘on the fly’ without the need for fully specifying a probability measure on all possible sequences of outcomes, as is the case in the more standard approach. We take the first steps in allowing for probability intervals instead of precise probabilities on this prequential approach, based on ideas borrowed from our earlier imprecise-probabilistic, standard account of algorithmic randomness. We define what it means for an infinite sequence of successive interval forecasts and subsequent binary outcomes to be random, both in a martingale-theoretic and a test-theoretic sense. We prove that these two versions of prequential randomness coincide, we compare the resulting prequential randomness notions with the more standard ones, and we investigate where the prequential and standard randomness notions coincide.
在算法随机性的前序方法中,可以 "即时 "预测下一个结果的概率,而无需像更标准的方法那样,完全指定所有可能结果序列的概率度量。我们借鉴了早先关于算法随机性的不精确概率标准论述,并在此基础上迈出了第一步,允许用概率区间代替精确概率。我们定义了连续区间预测 Ik 的无穷序列(I1,x1,I2,x2,......)和随后的二元结果 xk 在马丁格尔理论和检验理论意义上的随机性。我们证明这两个版本的前序随机性是重合的,我们将由此得出的前序随机性概念与更标准的随机性概念进行比较,并研究前序随机性概念与标准随机性概念的重合之处。
{"title":"Imprecision in martingale- and test-theoretic prequential randomness","authors":"Floris Persiau, Gert de Cooman","doi":"10.1016/j.ijar.2024.109213","DOIUrl":"10.1016/j.ijar.2024.109213","url":null,"abstract":"<div><p>In a prequential approach to algorithmic randomness, probabilities for the next outcome can be forecast ‘on the fly’ without the need for fully specifying a probability measure on all possible sequences of outcomes, as is the case in the more standard approach. We take the first steps in allowing for probability intervals instead of precise probabilities on this prequential approach, based on ideas borrowed from our earlier imprecise-probabilistic, standard account of algorithmic randomness. We define what it means for an infinite sequence <span><math><mo>(</mo><msub><mrow><mi>I</mi></mrow><mrow><mn>1</mn></mrow></msub><mo>,</mo><msub><mrow><mi>x</mi></mrow><mrow><mn>1</mn></mrow></msub><mo>,</mo><msub><mrow><mi>I</mi></mrow><mrow><mn>2</mn></mrow></msub><mo>,</mo><msub><mrow><mi>x</mi></mrow><mrow><mn>2</mn></mrow></msub><mo>,</mo><mo>…</mo><mo>)</mo></math></span> of successive interval forecasts <span><math><msub><mrow><mi>I</mi></mrow><mrow><mi>k</mi></mrow></msub></math></span> and subsequent binary outcomes <span><math><msub><mrow><mi>x</mi></mrow><mrow><mi>k</mi></mrow></msub></math></span> to be random, both in a martingale-theoretic and a test-theoretic sense. We prove that these two versions of prequential randomness coincide, we compare the resulting prequential randomness notions with the more standard ones, and we investigate where the prequential and standard randomness notions coincide.</p></div>","PeriodicalId":13842,"journal":{"name":"International Journal of Approximate Reasoning","volume":"170 ","pages":"Article 109213"},"PeriodicalIF":3.9,"publicationDate":"2024-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141023719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-10DOI: 10.1016/j.ijar.2024.109211
Leonardo Cella
This paper presents a novel distribution-free Inferential Model (IM) construction that provides valid probabilistic inference across a broad spectrum of distribution-free problems, even in finite sample settings. More specifically, the proposed IM has the capability to assign (imprecise) probabilities to assertions of interest about any feature of the unknown quantities under examination, and these probabilities are well-calibrated in a frequentist sense. It is also shown that finite-sample confidence regions can be derived from the IM for any such features. Particular emphasis is placed on quantile regression, a domain where uncertainty quantification often takes the form of set estimates for the regression coefficients in applications. Within this context, the IM facilitates the acquisition of these set estimates, ensuring they are finite-sample confidence regions. It also enables the provision of finite-sample valid probabilistic assignments for any assertions of interest about the regression coefficients. As a result, regardless of the type of uncertainty quantification desired, the proposed framework offers an appealing solution to quantile regression.
本文提出了一种新颖的无分布推理模型(IM)结构,它能在广泛的无分布问题中提供有效的概率推理,即使在有限样本环境中也是如此。更具体地说,所提出的推理模型有能力为所研究的未知量的任何特征的相关断言分配(不精确的)概率,而且这些概率在频数主义意义上是经过良好校准的。研究还表明,对于任何此类特征,都可以从 IM 中推导出有限样本置信区。本文特别强调了量化回归,在这一领域中,不确定性量化通常采用回归系数集合估计的形式。在这种情况下,IM 可以帮助获取这些集合估计值,确保它们是有限样本置信区域。它还能为回归系数的任何相关断言提供有限样本有效概率分配。因此,无论所需的不确定性量化类型如何,所提出的框架都为量化回归提供了一个极具吸引力的解决方案。
{"title":"Distribution-free Inferential Models: Achieving finite-sample valid probabilistic inference, with emphasis on quantile regression","authors":"Leonardo Cella","doi":"10.1016/j.ijar.2024.109211","DOIUrl":"https://doi.org/10.1016/j.ijar.2024.109211","url":null,"abstract":"<div><p>This paper presents a novel distribution-free Inferential Model (IM) construction that provides valid probabilistic inference across a broad spectrum of distribution-free problems, even in finite sample settings. More specifically, the proposed IM has the capability to assign (imprecise) probabilities to assertions of interest about any feature of the unknown quantities under examination, and these probabilities are well-calibrated in a frequentist sense. It is also shown that finite-sample confidence regions can be derived from the IM for any such features. Particular emphasis is placed on quantile regression, a domain where uncertainty quantification often takes the form of set estimates for the regression coefficients in applications. Within this context, the IM facilitates the acquisition of these set estimates, ensuring they are finite-sample confidence regions. It also enables the provision of finite-sample valid probabilistic assignments for any assertions of interest about the regression coefficients. As a result, regardless of the type of uncertainty quantification desired, the proposed framework offers an appealing solution to quantile regression.</p></div>","PeriodicalId":13842,"journal":{"name":"International Journal of Approximate Reasoning","volume":"170 ","pages":"Article 109211"},"PeriodicalIF":3.9,"publicationDate":"2024-05-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140948806","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-08DOI: 10.1016/j.ijar.2024.109210
Jianhua Dai , Zhilin Zhu , Min Li , Xiongtao Zou , Chucai Zhang
The neighborhood rough set model serves as an important tool for handling attribute reduction tasks involving heterogeneous attributes. However, measuring the relationship between conditional attributes and decision in the neighborhood rough set model is a crucial issue. Most studies have utilized neighborhood information entropy to measure the relationship between attributes. When using neighborhood conditional information entropy to measure the relationships between the decision and conditional attributes, it lacks monotonicity, consequently affecting the rationality of the final attribute reduction subset. In this paper, we introduce the concept of neighborhood granularity and propose a new form of relative neighborhood granularity to measure the relationship between the decision and conditional attributes, which exhibits monotonicity. Moreover, our approach for measuring neighborhood granularity avoids the logarithmic function computation involved in neighborhood information entropy. Finally, we conduct comparative experiments on 12 datasets using two classifiers to compare the results of attribute reduction with six other attribute reduction algorithms. The comparison demonstrates the advantages of our measurement approach.
{"title":"Attribute reduction for heterogeneous data based on monotonic relative neighborhood granularity","authors":"Jianhua Dai , Zhilin Zhu , Min Li , Xiongtao Zou , Chucai Zhang","doi":"10.1016/j.ijar.2024.109210","DOIUrl":"https://doi.org/10.1016/j.ijar.2024.109210","url":null,"abstract":"<div><p>The neighborhood rough set model serves as an important tool for handling attribute reduction tasks involving heterogeneous attributes. However, measuring the relationship between conditional attributes and decision in the neighborhood rough set model is a crucial issue. Most studies have utilized neighborhood information entropy to measure the relationship between attributes. When using neighborhood conditional information entropy to measure the relationships between the decision and conditional attributes, it lacks monotonicity, consequently affecting the rationality of the final attribute reduction subset. In this paper, we introduce the concept of neighborhood granularity and propose a new form of relative neighborhood granularity to measure the relationship between the decision and conditional attributes, which exhibits monotonicity. Moreover, our approach for measuring neighborhood granularity avoids the logarithmic function computation involved in neighborhood information entropy. Finally, we conduct comparative experiments on 12 datasets using two classifiers to compare the results of attribute reduction with six other attribute reduction algorithms. The comparison demonstrates the advantages of our measurement approach.</p></div>","PeriodicalId":13842,"journal":{"name":"International Journal of Approximate Reasoning","volume":"170 ","pages":"Article 109210"},"PeriodicalIF":3.9,"publicationDate":"2024-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140924463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-30DOI: 10.1016/j.ijar.2024.109209
Jesús M. Almendros-Jiménez , Antonio Becerra-Terón , Ginés Moreno , José A. Riaza
During the last years, the study of fuzzy database query languages has attracted the attention of many researchers. In this line of research, our group has proposed and developed FSA-SPARQL (Fuzzy Sets and Aggregators based SPARQL), which is a fuzzy extension of the Semantic Web query language SPARQL. FSA-SPARQL works with fuzzy RDF datasets and allows the definition of fuzzy queries involving fuzzy conditions through fuzzy connectives and aggregators. However, there are two main challenges to be solved for the practical applicability of FSA-SPARQL. The first problem is the lack of fuzzy RDF data sources. The second is how to customize fuzzy queries on fuzzy RDF data sources. Our research group has also recently proposed a fuzzy logic programming language called that offers powerful tuning capabilities that can accept applications in many fields. The purpose of this paper is to show how the tuning capabilities serve to accomplish in a unified framework both challenges in FSA-SPARQL: data fuzzification and query customization. More concretely, from a FSA-SPARQL to transformation, data fuzzification and query customization in FSA-SPARQL become tuning problems. We have validated the approach with queries against datasets from online communities.
{"title":"Tuning fuzzy SPARQL queries","authors":"Jesús M. Almendros-Jiménez , Antonio Becerra-Terón , Ginés Moreno , José A. Riaza","doi":"10.1016/j.ijar.2024.109209","DOIUrl":"https://doi.org/10.1016/j.ijar.2024.109209","url":null,"abstract":"<div><p>During the last years, the study of fuzzy database query languages has attracted the attention of many researchers. In this line of research, our group has proposed and developed <span>FSA-SPARQL</span> (<em>Fuzzy Sets and Aggregators based SPARQL</em>), which is a fuzzy extension of the Semantic Web query language SPARQL. <span>FSA-SPARQL</span> works with fuzzy RDF datasets and allows the definition of fuzzy queries involving fuzzy conditions through fuzzy connectives and aggregators. However, there are two main challenges to be solved for the practical applicability of <span>FSA-SPARQL</span>. The first problem is the lack of fuzzy RDF data sources. The second is how to customize fuzzy queries on fuzzy RDF data sources. Our research group has also recently proposed a fuzzy logic programming language called <span><math><mi>F</mi><mi>A</mi><mi>S</mi><mi>I</mi><mi>L</mi><mi>L</mi></math></span> that offers powerful tuning capabilities that can accept applications in many fields. The purpose of this paper is to show how the <span><math><mi>F</mi><mi>A</mi><mi>S</mi><mi>I</mi><mi>L</mi><mi>L</mi></math></span> tuning capabilities serve to accomplish in a unified framework both challenges in <span>FSA-SPARQL</span>: data fuzzification and query customization. More concretely, from a <span>FSA-SPARQL</span> to <span><math><mi>F</mi><mi>A</mi><mi>S</mi><mi>I</mi><mi>L</mi><mi>L</mi></math></span> transformation, data fuzzification and query customization in <span>FSA-SPARQL</span> become <span><math><mi>F</mi><mi>A</mi><mi>S</mi><mi>I</mi><mi>L</mi><mi>L</mi></math></span> tuning problems. We have validated the approach with queries against datasets from online communities.</p></div>","PeriodicalId":13842,"journal":{"name":"International Journal of Approximate Reasoning","volume":"170 ","pages":"Article 109209"},"PeriodicalIF":3.9,"publicationDate":"2024-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0888613X24000963/pdfft?md5=d1a2bfe1bba5c82a286b84ddb02fef74&pid=1-s2.0-S0888613X24000963-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140825401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-29DOI: 10.1016/j.ijar.2024.109208
Weihua Li , Yi Sun , Pei Heng
Collapsibility refers to the property that, when marginalizing over some variables that are not of interest from the full model, the resulting marginal model of the remaining variables is equivalent to the local model induced by the subgraph on these variables. This means that when the marginal model satisfies collapsibility, statistical inference results based on the marginal model and the local model are consistent. This has significant implications for small-sample data, modeling latent variable data, and reducing the computational complexity of statistical inference. This paper focuses on studying the conditional independence collapsibility of acyclic directed mixed graph (ADMG) models. By introducing the concept of inducing paths in ADMGs and exploring its properties, the conditional independence collapsibility of ADMGs is characterized equivalently from both graph theory and statistical perspectives.
{"title":"Conditional independence collapsibility for acyclic directed mixed graph models","authors":"Weihua Li , Yi Sun , Pei Heng","doi":"10.1016/j.ijar.2024.109208","DOIUrl":"https://doi.org/10.1016/j.ijar.2024.109208","url":null,"abstract":"<div><p>Collapsibility refers to the property that, when marginalizing over some variables that are not of interest from the full model, the resulting marginal model of the remaining variables is equivalent to the local model induced by the subgraph on these variables. This means that when the marginal model satisfies collapsibility, statistical inference results based on the marginal model and the local model are consistent. This has significant implications for small-sample data, modeling latent variable data, and reducing the computational complexity of statistical inference. This paper focuses on studying the conditional independence collapsibility of acyclic directed mixed graph (ADMG) models. By introducing the concept of inducing paths in ADMGs and exploring its properties, the conditional independence collapsibility of ADMGs is characterized equivalently from both graph theory and statistical perspectives.</p></div>","PeriodicalId":13842,"journal":{"name":"International Journal of Approximate Reasoning","volume":"170 ","pages":"Article 109208"},"PeriodicalIF":3.9,"publicationDate":"2024-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140825400","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-29DOI: 10.1016/j.ijar.2024.109205
Marco Grzegorczyk
In this paper we propose a Bayesian approach for inferring Bayesian network (BN) structures from ordinal data. Our approach can be seen as the Bayesian counterpart of a recently proposed frequentist approach, referred to as the ‘ordinal structure expectation maximization’ (OSEM) method. Like for the OSEM method, the key idea is to assume that each ordinal variable originates from a Gaussian variable that can only be observed in discretized form, and that the dependencies in the latent Gaussian space can be modeled by BNs; i.e. by directed acyclic graphs (DAGs). Our Bayesian method combines the ‘structure MCMC sampler’ for DAG posterior sampling, a slightly modified version of the ‘Bayesian metric for Gaussian networks having score equivalence’ (BGe score), the concept of the ‘extended rank likelihood’, and a recently proposed algorithm for posterior sampling the parameters of Gaussian BNs. In simulation studies we compare the new Bayesian approach and the OSEM method in terms of the network reconstruction accuracy. The empirical results show that the new Bayesian approach leads to significantly improved network reconstruction accuracies.
本文提出了一种从序数数据推断贝叶斯网络(BN)结构的贝叶斯方法。我们的方法可以看作是最近提出的一种频繁主义方法的贝叶斯对应方法,即 "序数结构期望最大化"(OSEM)方法。与 OSEM 方法一样,我们的关键思路是假设每个序变量都来自于一个高斯变量,而这个高斯变量只能以离散的形式被观测到,并且潜在高斯空间中的依赖关系可以用 BN(即有向无环图(DAG))来建模。我们的贝叶斯方法结合了用于 DAG 后验采样的 "结构 MCMC 采样器"、稍作修改的 "具有分数等价性的高斯网络贝叶斯度量"(BGe score)版本、"扩展秩似然法 "概念以及最近提出的用于高斯 BN 参数后验采样的算法。在模拟研究中,我们比较了新贝叶斯方法和 OSEM 方法的网络重建精度。实证结果表明,新贝叶斯方法显著提高了网络重建精度。
{"title":"Being Bayesian about learning Bayesian networks from ordinal data","authors":"Marco Grzegorczyk","doi":"10.1016/j.ijar.2024.109205","DOIUrl":"https://doi.org/10.1016/j.ijar.2024.109205","url":null,"abstract":"<div><p>In this paper we propose a Bayesian approach for inferring Bayesian network (BN) structures from ordinal data. Our approach can be seen as the Bayesian counterpart of a recently proposed frequentist approach, referred to as the ‘ordinal structure expectation maximization’ (OSEM) method. Like for the OSEM method, the key idea is to assume that each ordinal variable originates from a Gaussian variable that can only be observed in discretized form, and that the dependencies in the latent Gaussian space can be modeled by BNs; i.e. by directed acyclic graphs (DAGs). Our Bayesian method combines the ‘structure MCMC sampler’ for DAG posterior sampling, a slightly modified version of the ‘Bayesian metric for Gaussian networks having score equivalence’ (BGe score), the concept of the ‘extended rank likelihood’, and a recently proposed algorithm for posterior sampling the parameters of Gaussian BNs. In simulation studies we compare the new Bayesian approach and the OSEM method in terms of the network reconstruction accuracy. The empirical results show that the new Bayesian approach leads to significantly improved network reconstruction accuracies.</p></div>","PeriodicalId":13842,"journal":{"name":"International Journal of Approximate Reasoning","volume":"170 ","pages":"Article 109205"},"PeriodicalIF":3.9,"publicationDate":"2024-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0888613X24000926/pdfft?md5=399d96b4aa67e5b17c35c12d9efa291d&pid=1-s2.0-S0888613X24000926-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140825402","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-25DOI: 10.1016/j.ijar.2024.109206
Ismaïl Baaj , Zied Bouraoui , Antoine Cornuéjols , Thierry Denœux , Sébastien Destercke , Didier Dubois , Marie-Jeanne Lesot , João Marques-Silva , Jérôme Mengin , Henri Prade , Steven Schockaert , Mathieu Serrurier , Olivier Strauss , Christel Vrain
This paper proposes a tentative and original survey of meeting points between Knowledge Representation and Reasoning (KRR) and Machine Learning (ML), two areas which have been developed quite separately in the last four decades. First, some common concerns are identified and discussed such as the types of representation used, the roles of knowledge and data, the lack or the excess of information, or the need for explanations and causal understanding. Then, the survey is organised in seven sections covering most of the territory where KRR and ML meet. We start with a section dealing with prototypical approaches from the literature on learning and reasoning: Inductive Logic Programming, Statistical Relational Learning, and Neurosymbolic AI, where ideas from rule-based reasoning are combined with ML. Then we focus on the use of various forms of background knowledge in learning, ranging from additional regularisation terms in loss functions, to the problem of aligning symbolic and vector space representations, or the use of knowledge graphs for learning. Then, the next section describes how KRR notions may benefit to learning tasks. For instance, constraints can be used as in declarative data mining for influencing the learned patterns; or semantic features are exploited in low-shot learning to compensate for the lack of data; or yet we can take advantage of analogies for learning purposes. Conversely, another section investigates how ML methods may serve KRR goals. For instance, one may learn special kinds of rules such as default rules, fuzzy rules or threshold rules, or special types of information such as constraints, or preferences. The section also covers formal concept analysis and rough sets-based methods. Yet another section reviews various interactions between Automated Reasoning and ML, such as the use of ML methods in SAT solving to make reasoning faster. Then a section deals with works related to model accountability, including explainability and interpretability, fairness and robustness. Finally, a section covers works on handling imperfect or incomplete data, including the problem of learning from uncertain or coarse data, the use of belief functions for regression, a revision-based view of the EM algorithm, the use of possibility theory in statistics, or the learning of imprecise models. This paper thus aims at a better mutual understanding of research in KRR and ML, and how they can cooperate. The paper is completed by an abundant bibliography.
知识表示与推理(KRR)和机器学习(ML)这两个领域在过去四十年中各自发展,本文对这两个领域的交汇点进行了初步和原创性的调查。首先,确定并讨论了一些共同关注的问题,如所使用的表示类型、知识和数据的作用、信息的缺乏或过剩,或对解释和因果理解的需求。然后,调查分为七个部分,涵盖了 KRR 和 ML 的大部分领域。我们首先讨论学习与推理文献中的原型方法:归纳逻辑编程、统计关系学习和神经符号人工智能将基于规则的推理思想与 ML 相结合。然后,我们将重点讨论在学习中使用各种形式的背景知识,从损失函数中的附加正则化项,到符号和向量空间表征的对齐问题,或在学习中使用知识图谱。然后,下一节将介绍 KRR 概念如何有益于学习任务。例如,在声明式数据挖掘中,可以使用约束条件来影响学习模式;在低射学习中,可以利用语义特征来弥补数据的不足;我们还可以利用类比来达到学习目的。与此相反,另一部分研究了 ML 方法如何服务于 KRR 目标。例如,我们可以学习特殊类型的规则,如默认规则、模糊规则或阈值规则,或特殊类型的信息,如约束或偏好。这一部分还涉及形式概念分析和基于粗糙集的方法。还有一节回顾了自动推理与 ML 之间的各种互动,例如在 SAT 求解中使用 ML 方法来加快推理速度。然后,还有一节涉及与模型责任相关的工作,包括可解释性和可解释性、公平性和稳健性。最后,还有一部分涉及处理不完善或不完整数据的工作,包括从不确定性或粗糙数据中学习的问题、使用信念函数进行回归、基于修正的 EM 算法观点、统计学中可能性理论的使用或不精确模型的学习。因此,本文旨在更好地相互理解 KRR 和 ML 的研究,以及它们之间如何合作。本文最后附有丰富的参考书目。
{"title":"Synergies between machine learning and reasoning - An introduction by the Kay R. Amel group","authors":"Ismaïl Baaj , Zied Bouraoui , Antoine Cornuéjols , Thierry Denœux , Sébastien Destercke , Didier Dubois , Marie-Jeanne Lesot , João Marques-Silva , Jérôme Mengin , Henri Prade , Steven Schockaert , Mathieu Serrurier , Olivier Strauss , Christel Vrain","doi":"10.1016/j.ijar.2024.109206","DOIUrl":"10.1016/j.ijar.2024.109206","url":null,"abstract":"<div><p>This paper proposes a tentative and original survey of meeting points between Knowledge Representation and Reasoning (KRR) and Machine Learning (ML), two areas which have been developed quite separately in the last four decades. First, some common concerns are identified and discussed such as the types of representation used, the roles of knowledge and data, the lack or the excess of information, or the need for explanations and causal understanding. Then, the survey is organised in seven sections covering most of the territory where KRR and ML meet. We start with a section dealing with prototypical approaches from the literature on learning and reasoning: Inductive Logic Programming, Statistical Relational Learning, and Neurosymbolic AI, where ideas from rule-based reasoning are combined with ML. Then we focus on the use of various forms of background knowledge in learning, ranging from additional regularisation terms in loss functions, to the problem of aligning symbolic and vector space representations, or the use of knowledge graphs for learning. Then, the next section describes how KRR notions may benefit to learning tasks. For instance, constraints can be used as in declarative data mining for influencing the learned patterns; or semantic features are exploited in low-shot learning to compensate for the lack of data; or yet we can take advantage of analogies for learning purposes. Conversely, another section investigates how ML methods may serve KRR goals. For instance, one may learn special kinds of rules such as default rules, fuzzy rules or threshold rules, or special types of information such as constraints, or preferences. The section also covers formal concept analysis and rough sets-based methods. Yet another section reviews various interactions between Automated Reasoning and ML, such as the use of ML methods in SAT solving to make reasoning faster. Then a section deals with works related to model accountability, including explainability and interpretability, fairness and robustness. Finally, a section covers works on handling imperfect or incomplete data, including the problem of learning from uncertain or coarse data, the use of belief functions for regression, a revision-based view of the EM algorithm, the use of possibility theory in statistics, or the learning of imprecise models. This paper thus aims at a better mutual understanding of research in KRR and ML, and how they can cooperate. The paper is completed by an abundant bibliography.</p></div>","PeriodicalId":13842,"journal":{"name":"International Journal of Approximate Reasoning","volume":"171 ","pages":"Article 109206"},"PeriodicalIF":3.9,"publicationDate":"2024-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0888613X24000938/pdfft?md5=8e95ec50ba09a214586a09348ae99f54&pid=1-s2.0-S0888613X24000938-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140774538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-24DOI: 10.1016/j.ijar.2024.109207
Sébastien Destercke , Jérôme Mengin , Henri Prade
{"title":"Editorial of the special issue “Synergies Between Machine Learning and Reasoning”","authors":"Sébastien Destercke , Jérôme Mengin , Henri Prade","doi":"10.1016/j.ijar.2024.109207","DOIUrl":"10.1016/j.ijar.2024.109207","url":null,"abstract":"","PeriodicalId":13842,"journal":{"name":"International Journal of Approximate Reasoning","volume":"171 ","pages":"Article 109207"},"PeriodicalIF":3.9,"publicationDate":"2024-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140755910","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}