Pub Date : 2023-07-31DOI: 10.48550/arXiv.2307.16577
Marco Zaffalon, Alessandro Antonucci, Rafael Cabañas, David Huber
We address the problem of integrating data from multiple, possibly biased, observational and interventional studies, to eventually compute counterfactuals in structural causal models. We start from the case of a single observational dataset affected by a selection bias. We show that the likelihood of the available data has no local maxima. This enables us to use the causal expectation-maximisation scheme to approximate the bounds for partially identifiable counterfactual queries, which are the focus of this paper. We then show how the same approach can address the general case of multiple datasets, no matter whether interventional or observational, biased or unbiased, by remapping it into the former one via graphical transformations. Systematic numerical experiments and a case study on palliative care show the effectiveness of our approach, while hinting at the benefits of fusing heterogeneous data sources to get informative outcomes in case of partial identifiability.
{"title":"Approximating Counterfactual Bounds while Fusing Observational, Biased and Randomised Data Sources","authors":"Marco Zaffalon, Alessandro Antonucci, Rafael Cabañas, David Huber","doi":"10.48550/arXiv.2307.16577","DOIUrl":"https://doi.org/10.48550/arXiv.2307.16577","url":null,"abstract":"We address the problem of integrating data from multiple, possibly biased, observational and interventional studies, to eventually compute counterfactuals in structural causal models. We start from the case of a single observational dataset affected by a selection bias. We show that the likelihood of the available data has no local maxima. This enables us to use the causal expectation-maximisation scheme to approximate the bounds for partially identifiable counterfactual queries, which are the focus of this paper. We then show how the same approach can address the general case of multiple datasets, no matter whether interventional or observational, biased or unbiased, by remapping it into the former one via graphical transformations. Systematic numerical experiments and a case study on palliative care show the effectiveness of our approach, while hinting at the benefits of fusing heterogeneous data sources to get informative outcomes in case of partial identifiability.","PeriodicalId":13685,"journal":{"name":"Int. J. Approx. Reason.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89171249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Random sets, copulas and related sets of probability measures","authors":"B. Schmelzer","doi":"10.2139/ssrn.4364268","DOIUrl":"https://doi.org/10.2139/ssrn.4364268","url":null,"abstract":"","PeriodicalId":13685,"journal":{"name":"Int. J. Approx. Reason.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79206276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Attribute reduction based on fusion information entropy","authors":"Xia Ji, Jie Li, Peng Zhao, Sheng Yao","doi":"10.2139/ssrn.4334415","DOIUrl":"https://doi.org/10.2139/ssrn.4334415","url":null,"abstract":"","PeriodicalId":13685,"journal":{"name":"Int. J. Approx. Reason.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87556416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-04-12DOI: 10.48550/arXiv.2304.05641
J. Järvinen, S. Radeleczki
We study the pseudo-Kleene algebras of the Dedekind-MacNeille completion of the ordered set of rough set determined by a reflexive relation. We characterize the cases when PBZ and PBZ*-lattices can be defined on these pseudo-Kleene algebras.
{"title":"Pseudo-Kleene algebras determined by rough sets","authors":"J. Järvinen, S. Radeleczki","doi":"10.48550/arXiv.2304.05641","DOIUrl":"https://doi.org/10.48550/arXiv.2304.05641","url":null,"abstract":"We study the pseudo-Kleene algebras of the Dedekind-MacNeille completion of the ordered set of rough set determined by a reflexive relation. We characterize the cases when PBZ and PBZ*-lattices can be defined on these pseudo-Kleene algebras.","PeriodicalId":13685,"journal":{"name":"Int. J. Approx. Reason.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-04-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84325126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-03-09DOI: 10.48550/arXiv.2303.05067
M. Goerigk, R. Guillaume, A. Kasperski, Pawel Zieli'nski
In this paper, an optimization problem with uncertain objective function coefficients is considered. The uncertainty is specified by providing a discrete scenario set, containing possible realizations of the objective function coefficients. The concept of belief function in the traditional and possibilistic setting is applied to define a set of admissible probability distributions over the scenario set. The generalized Hurwicz criterion is then used to compute a solution. In this paper, the complexity of the resulting problem is explored. Some exact and approximation methods of solving it are proposed.
{"title":"Robust optimization with belief functions","authors":"M. Goerigk, R. Guillaume, A. Kasperski, Pawel Zieli'nski","doi":"10.48550/arXiv.2303.05067","DOIUrl":"https://doi.org/10.48550/arXiv.2303.05067","url":null,"abstract":"In this paper, an optimization problem with uncertain objective function coefficients is considered. The uncertainty is specified by providing a discrete scenario set, containing possible realizations of the objective function coefficients. The concept of belief function in the traditional and possibilistic setting is applied to define a set of admissible probability distributions over the scenario set. The generalized Hurwicz criterion is then used to compute a solution. In this paper, the complexity of the resulting problem is explored. Some exact and approximation methods of solving it are proposed.","PeriodicalId":13685,"journal":{"name":"Int. J. Approx. Reason.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74135193","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-02-01DOI: 10.48550/arXiv.2302.00364
R. Ballester-Ripoll, Manuele Leonelli
Sensitivity analysis measures the influence of a Bayesian network's parameters on a quantity of interest defined by the network, such as the probability of a variable taking a specific value. Various sensitivity measures have been defined to quantify such influence, most commonly some function of the quantity of interest's partial derivative with respect to the network's conditional probabilities. However, computing these measures in large networks with thousands of parameters can become computationally very expensive. We propose an algorithm combining automatic differentiation and exact inference to efficiently calculate the sensitivity measures in a single pass. It first marginalizes the whole network once, using e.g. variable elimination, and then backpropagates this operation to obtain the gradient with respect to all input parameters. Our method can be used for one-way and multi-way sensitivity analysis and the derivation of admissible regions. Simulation studies highlight the efficiency of our algorithm by scaling it to massive networks with up to 100'000 parameters and investigate the feasibility of generic multi-way analyses. Our routines are also showcased over two medium-sized Bayesian networks: the first modeling the country-risks of a humanitarian crisis, the second studying the relationship between the use of technology and the psychological effects of forced social isolation during the COVID-19 pandemic. An implementation of the methods using the popular machine learning library PyTorch is freely available.
{"title":"The YODO algorithm: An efficient computational framework for sensitivity analysis in Bayesian networks","authors":"R. Ballester-Ripoll, Manuele Leonelli","doi":"10.48550/arXiv.2302.00364","DOIUrl":"https://doi.org/10.48550/arXiv.2302.00364","url":null,"abstract":"Sensitivity analysis measures the influence of a Bayesian network's parameters on a quantity of interest defined by the network, such as the probability of a variable taking a specific value. Various sensitivity measures have been defined to quantify such influence, most commonly some function of the quantity of interest's partial derivative with respect to the network's conditional probabilities. However, computing these measures in large networks with thousands of parameters can become computationally very expensive. We propose an algorithm combining automatic differentiation and exact inference to efficiently calculate the sensitivity measures in a single pass. It first marginalizes the whole network once, using e.g. variable elimination, and then backpropagates this operation to obtain the gradient with respect to all input parameters. Our method can be used for one-way and multi-way sensitivity analysis and the derivation of admissible regions. Simulation studies highlight the efficiency of our algorithm by scaling it to massive networks with up to 100'000 parameters and investigate the feasibility of generic multi-way analyses. Our routines are also showcased over two medium-sized Bayesian networks: the first modeling the country-risks of a humanitarian crisis, the second studying the relationship between the use of technology and the psychological effects of forced social isolation during the COVID-19 pandemic. An implementation of the methods using the popular machine learning library PyTorch is freely available.","PeriodicalId":13685,"journal":{"name":"Int. J. Approx. Reason.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86717078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-27DOI: 10.48550/arXiv.2301.11479
Thibault Gauthier, Miroslav Olsák, J. Urban
We introduce a self-learning algorithm for synthesizing programs for OEIS sequences. The algorithm starts from scratch initially generating programs at random. Then it runs many iterations of a self-learning loop that interleaves (i) training neural machine translation to learn the correspondence between sequences and the programs discovered so far, and (ii) proposing many new programs for each OEIS sequence by the trained neural machine translator. The algorithm discovers on its own programs for more than 78000 OEIS sequences, sometimes developing unusual programming methods. We analyze its behavior and the invented programs in several experiments.
{"title":"Alien Coding","authors":"Thibault Gauthier, Miroslav Olsák, J. Urban","doi":"10.48550/arXiv.2301.11479","DOIUrl":"https://doi.org/10.48550/arXiv.2301.11479","url":null,"abstract":"We introduce a self-learning algorithm for synthesizing programs for OEIS sequences. The algorithm starts from scratch initially generating programs at random. Then it runs many iterations of a self-learning loop that interleaves (i) training neural machine translation to learn the correspondence between sequences and the programs discovered so far, and (ii) proposing many new programs for each OEIS sequence by the trained neural machine translator. The algorithm discovers on its own programs for more than 78000 OEIS sequences, sometimes developing unusual programming methods. We analyze its behavior and the invented programs in several experiments.","PeriodicalId":13685,"journal":{"name":"Int. J. Approx. Reason.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77583413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-12-20DOI: 10.48550/arXiv.2212.10208
M. Koyda, Gerd Stumme
This work investigates the factorization of finite lattices to implode selected intervals while preserving the remaining order structure. We examine how complete congruence relations and complete tolerance relations can be utilized for this purpose and answer the question of finding the finest of those relations to implode a given interval in the generated factor lattice. To overcome the limitations of the factorization based on those relations, we introduce a new lattice factorization that enables the imploding of selected disjoint intervals of a finite lattice. To this end, we propose an interval relation that generates this factorization. To obtain lattices rather than arbitrary ordered sets, we restrict this approach to so-called pure intervals. For our study, we will make use of methods from Formal Concept Analysis (FCA). We will also provide a new FCA construction by introducing the enrichment of an incidence relation by a set of intervals in a formal context, to investigate the approach for lattice-generating interval relations on the context side.
{"title":"Factorizing Lattices by Interval Relations","authors":"M. Koyda, Gerd Stumme","doi":"10.48550/arXiv.2212.10208","DOIUrl":"https://doi.org/10.48550/arXiv.2212.10208","url":null,"abstract":"This work investigates the factorization of finite lattices to implode selected intervals while preserving the remaining order structure. We examine how complete congruence relations and complete tolerance relations can be utilized for this purpose and answer the question of finding the finest of those relations to implode a given interval in the generated factor lattice. To overcome the limitations of the factorization based on those relations, we introduce a new lattice factorization that enables the imploding of selected disjoint intervals of a finite lattice. To this end, we propose an interval relation that generates this factorization. To obtain lattices rather than arbitrary ordered sets, we restrict this approach to so-called pure intervals. For our study, we will make use of methods from Formal Concept Analysis (FCA). We will also provide a new FCA construction by introducing the enrichment of an incidence relation by a set of intervals in a formal context, to investigate the approach for lattice-generating interval relations on the context side.","PeriodicalId":13685,"journal":{"name":"Int. J. Approx. Reason.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75127118","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-12-12DOI: 10.48550/arXiv.2212.05990
Yacine Izza, Xuanxiang Huang, Alexey Ignatiev, Nina Narodytska, Martin C. Cooper, Joao Marques-Silva
The most widely studied explainable AI (XAI) approaches are unsound. This is the case with well-known model-agnostic explanation approaches, and it is also the case with approaches based on saliency maps. One solution is to consider intrinsic interpretability, which does not exhibit the drawback of unsoundness. Unfortunately, intrinsic interpretability can display unwieldy explanation redundancy. Formal explainability represents the alternative to these non-rigorous approaches, with one example being PI-explanations. Unfortunately, PI-explanations also exhibit important drawbacks, the most visible of which is arguably their size. Recently, it has been observed that the (absolute) rigor of PI-explanations can be traded off for a smaller explanation size, by computing the so-called relevant sets. Given some positive {delta}, a set S of features is {delta}-relevant if, when the features in S are fixed, the probability of getting the target class exceeds {delta}. However, even for very simple classifiers, the complexity of computing relevant sets of features is prohibitive, with the decision problem being NPPP-complete for circuit-based classifiers. In contrast with earlier negative results, this paper investigates practical approaches for computing relevant sets for a number of widely used classifiers that include Decision Trees (DTs), Naive Bayes Classifiers (NBCs), and several families of classifiers obtained from propositional languages. Moreover, the paper shows that, in practice, and for these families of classifiers, relevant sets are easy to compute. Furthermore, the experiments confirm that succinct sets of relevant features can be obtained for the families of classifiers considered.
{"title":"On Computing Probabilistic Abductive Explanations","authors":"Yacine Izza, Xuanxiang Huang, Alexey Ignatiev, Nina Narodytska, Martin C. Cooper, Joao Marques-Silva","doi":"10.48550/arXiv.2212.05990","DOIUrl":"https://doi.org/10.48550/arXiv.2212.05990","url":null,"abstract":"The most widely studied explainable AI (XAI) approaches are unsound. This is the case with well-known model-agnostic explanation approaches, and it is also the case with approaches based on saliency maps. One solution is to consider intrinsic interpretability, which does not exhibit the drawback of unsoundness. Unfortunately, intrinsic interpretability can display unwieldy explanation redundancy. Formal explainability represents the alternative to these non-rigorous approaches, with one example being PI-explanations. Unfortunately, PI-explanations also exhibit important drawbacks, the most visible of which is arguably their size. Recently, it has been observed that the (absolute) rigor of PI-explanations can be traded off for a smaller explanation size, by computing the so-called relevant sets. Given some positive {delta}, a set S of features is {delta}-relevant if, when the features in S are fixed, the probability of getting the target class exceeds {delta}. However, even for very simple classifiers, the complexity of computing relevant sets of features is prohibitive, with the decision problem being NPPP-complete for circuit-based classifiers. In contrast with earlier negative results, this paper investigates practical approaches for computing relevant sets for a number of widely used classifiers that include Decision Trees (DTs), Naive Bayes Classifiers (NBCs), and several families of classifiers obtained from propositional languages. Moreover, the paper shows that, in practice, and for these families of classifiers, relevant sets are easy to compute. Furthermore, the experiments confirm that succinct sets of relevant features can be obtained for the families of classifiers considered.","PeriodicalId":13685,"journal":{"name":"Int. J. Approx. Reason.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74665823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}