Distinguishing Cause from Effect Using Observational Data: Methods and Benchmarks
@article{Mooij2014DistinguishingCF, title={Distinguishing Cause from Effect Using Observational Data: Methods and Benchmarks}, author={Joris M. Mooij and J. Peters and Dominik Janzing and Jakob Zscheischler and Bernhard Sch{\"o}lkopf}, journal={ArXiv}, year={2014}, volume={abs/1412.3773} }
The discovery of causal relationships from purely observational data is a fundamental problem in science. The most elementary form of such a causal discovery problem is to decide whether X causes Y or, alternatively, Y causes X, given joint observations of two variables X,Y. An example is to decide whether altitude causes temperature, or vice versa, given only joint measurements of both variables. Even under the simplifying assumptions of no confounding, no feedback loops, and no selection bias…
Figures and Tables from this paper
figure 1 table 1 figure 2 table 2 figure 3 table 3 figure 4 table 4 figure 5 figure 6 figure 7 figure 8 figure 9 figure 10 figure 11 figure 12 figure 13 figure 14 figure 15 figure 16 figure 17 figure 18 figure 19 figure 20 figure 21 figure 22 figure 23 figure 24 figure 25 figure 26 figure 27 figure 28 figure 29 figure 30 figure 31 figure 32 figure 33 figure 34 figure 35 figure 36 figure 37 figure 38 figure 39 figure 40 figure 41 figure 42 figure 43 figure 44 figure 45 figure 46 figure 47 figure 48 figure 49 figure 50 figure 51 figure 52 figure 53 figure 54
370 Citations
Comparing Causal Discovery Methods using Synthetic and Real Data
- Computer Science
- 2020
The goal of this work is to evaluate and to compare a number of state-of-the-art techniques in a joint benchmark and employ synthetic data, where it can give detailed reasoning about the resulting performance of the individual methods given their underlying assumptions.
Bivariate Causal Discovery via Conditional Divergence
- Computer ScienceCLeaR
- 2022
This work proposes Conditional Divergence based Causal Inference (CDCI), a novel algorithm for detecting causal direction in purely observational data that can relax multiple strict assumptions commonly adopted in the causal discovery literature, including functional form and noise model.
Leveraging directed causal discovery to detect latent common causes
- Computer ScienceIEEE transactions on neural networks and learning systems
- 2022
A general heuristic is devised which takes a causal discovery algorithm that can only distinguish purely directed causal relations and modifies it to also detect latent common causes, and demonstrates that both the modified algorithms preserve the performance of the original in distinguishing directed causal Relations.
The Effect of Noise Level on Causal Identification with Additive Noise Models
- Computer ScienceArXiv
- 2021
This work considered bivariate cases, which is the most elementary form of a causal discovery problem where one needs to decide whetherX causes Y or Y causesX, given joint distributions of two variablesX, Y, and found that these methods can fail to capture the true causal direction for some levels of noise.
Joint Causal Inference from Multiple Contexts
- Computer ScienceJ. Mach. Learn. Res.
- 2020
This work introduces Joint Causal Inference, a novel approach to causal discovery from multiple data sets from different contexts that elegantly unifies both approaches and concludes that JCI implementations can considerably outperform state-of-the-art causal discovery algorithms.
Joint Causal Inference from Multiple Datasets
- Computer Science
- 2018
JCI deals with several different types of interventions in a unified fashion, it can learn intervention targets, it pools data across different datasets which improves the statistical power of independence tests, and by exploiting differences in distribution between contexts it improves on the accuracy and identifiability of the predicted causal relations.
Causal Discovery Beyond Conditional Independences
- Computer Science
- 2015
This thesis addresses the problem of causal discovery, that is, recovering the underlying causal structure based on the joint probability distribution of the observed random variables, and estimates a set of Markov equivalent graphs.
Effectiveness of classification approach in recovering pairwise causal relations from data.
- Computer Science
- 2018
This thesis extends the bivariate case to the possibility of having at least one confounder between X and Y and improves the Bayesian network scores for the network structures of some medium to large-sized networks using the standard ordering based search algorithms.
The Randomized Causation Coefficient
- Computer ScienceJ. Mach. Learn. Res.
- 2015
This short paper proposes to learn how to perform causal inference directly from data, without the need of feature engineering, and poses causality as a kernel mean embedding classification problem, where inputs are samples from arbitrary probability distributions on pairs of random variables, and labels are types of causal relationships.
Results of the Cause-Effect Pair Challenge
- Computer ScienceCause Effect Pairs in Machine Learning
- 2019
A challenge in causal discovery from observational data with the aim of devising a “causation coefficient” to score pairs of variables to discover whether the data supports the hypothesis that Y = f(X, noise), which for the purpose of this challenge was the definition of causality.
References
SHOWING 1-10 OF 117 REFERENCES
Inference of Cause and Effect with Unsupervised Inverse Regression
- Computer ScienceAISTATS
- 2015
This work addresses the problem of causal discovery in the two-variable case, given a sample from their joint distribution and proposes an implicit notion of independence, namely that pY|X cannot be estimated based on pX (lower case denotes density), however, it may be possible to estimate pY |X based on the density of the effect, pY.
Distinguishing Cause and Effect via Second Order Exponential Models
- Computer Science
- 2009
This work proposes a method to infer causal structures containing both discrete and continuous variables and describes that sometimes a causal hypothesis must be rejected because P(effect|cause) and P(cause) share algorithmic information (which is untypical if they are chosen independently).
Causal Inference on Discrete Data Using Additive Noise Models
- Computer Science, MathematicsIEEE Transactions on Pattern Analysis and Machine Intelligence
- 2011
It is proved that it almost never occurs that additive noise models can be fit in both directions, and an efficient algorithm is proposed that is able to perform this way of causal inference on finite samples of discrete variables.
Regression by dependence minimization and its application to causal inference in additive noise models
- Computer Science, MathematicsICML '09
- 2009
This work proposes a novel method for regression that minimizes the statistical dependence between regressors and residuals, and proposes an algorithm for efficiently inferring causal models from observational data for more than two variables.
Estimation of causal effects using linear non-Gaussian causal models with hidden variables
- Computer ScienceInt. J. Approx. Reason.
- 2008
On the completeness of orientation rules for causal discovery in the presence of latent confounders and selection bias
- Computer ScienceArtif. Intell.
- 2008
A Linear Non-Gaussian Acyclic Model for Causal Discovery
- Computer ScienceJ. Mach. Learn. Res.
- 2006
This work shows how to discover the complete causal structure of continuous-valued data, under the assumptions that (a) the data generating process is linear, (b) there are no unobserved confounders, and (c) disturbance variables have non-Gaussian distributions of non-zero variances.
Causal discovery with continuous additive noise models
- MathematicsJ. Mach. Learn. Res.
- 2014
If the observational distribution follows a structural equation model with an additive noise structure, the directed acyclic graph becomes identifiable from the distribution under mild conditions, which constitutes an interesting alternative to traditional methods that assume faithfulness and identify only the Markov equivalence class of the graph, thus leaving some edges undirected.
Probabilistic latent variable models for distinguishing between cause and effect
- Computer ScienceNIPS
- 2010
A novel method for inferring whether X causes Y or vice versa from joint observations of X and Y is proposed, which considers the hypothetical effect variable to be a function of the hypothetical cause variable and an independent noise term.
Causal Inference Using Nonnormality
- Business
- 2004
Path analysis, often applied to observational data to study causal structures, describes causal relationship between observed variables. The path analysis is of confirmatory nature and can make…