Abstract

We address the problem of two-variable causal inference without intervention. This task is to infer an existing causal relation between two random variables, i.e., or , from purely observational data. As the option to modify a potential cause is not given in many situations, only structural properties of the data can be used to solve this ill-posed problem. We briefly review a number of state-of-the-art methods for this, including very recent ones. A novel inference method is introduced, Bayesian Causal Inference (BCI) which assumes a generative Bayesian hierarchical model to pursue the strategy of Bayesian model selection. In the adopted model, the distribution of the cause variable is given by a Poisson lognormal distribution, which allows to explicitly regard the discrete nature of datasets, correlations in the parameter spaces, as well as the variance of probability densities on logarithmic scales. We assume Fourier diagonal Field covariance operators. The model itself is restricted to use cases where a direct causal relation has to be decided against a relation , therefore we compare it other methods for this exact problem setting. The generative model assumed provides synthetic causal data for benchmarking our model in comparison to existing state-of-the-art models, namely LiNGAM, ANM-HSIC, ANM-MML, IGCI, and CGNN. We explore how well the above methods perform in case of high noise settings, strongly discretized data, and very sparse data. BCI performs generally reliably with synthetic data as well as with the real world TCEP benchmark set, with an accuracy comparable to state-of-the-art algorithms. We discuss directions for the future development of BCI.

Highlights

  • We use the ANM Algorithm [2] with HSIC and Gaussian Process Regression (ANM-HSIC) as well as the ANM-MML approach [12]. The latter uses a Bayesian Model Selection, arguably the closest to the algorithm proposed in this publication, at least to our best knowledge

  • While our Bayesian Causal Inference (BCI) algorithm is affected but still performs reliably with an accuracy of ≥ 90%, the ANM algorithms are remarkably robust in the presence of the noise

  • The problem of purely bivariate causal discovery is a rather restricted one as it ignores the possibility of causal independence or hidden confounders

Read more

Summary

Introduction

Causal Inference regards the problem of drawing conclusions about how some entity we can observe does—or does not—influence or is being influenced by another entity. Having knowledge about such law-like causal relations enables us to predict what will happen One can draw the conclusion that a street will be wet (the effect) whenever it rains (the cause) Knowing that it will rain, or observing the rainfall itself, enables one to predict that the street will be wet. Under ideal conditions the system under investigation can be manipulated Such interventions might allow to set causal variables to specific values which allows to study their effects statistically. There, an identification of causal directions has Entropy 2020, 22, 46; doi:10.3390/e22010046 www.mdpi.com/journal/entropy

Objectives
Results
Conclusion

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.