Alain Vaucher, Philippe Schwaller, et al.
AMLD EPFL 2022
We show how a generic feature-selection algorithm returning strongly relevant variables can be turned into a causal structure-learning algorithm. We prove this under the Faithfulness assumption for the data distribution. In a causal graph, the strongly relevant variables for a node X are its parents, children, and children's parents (or spouses), also known as the Markov blanket of X. Identifying the spouses leads to the detection of the V-structure patterns and thus to causal orientations. Repeating the task for all variables yields a valid partially oriented causal graph. We first show an efficient way to identify the spouse links. We then perform several experiments in the continuous domain using the Recursive Feature Elimination feature-selection algorithm with Support Vector Regression and empirically verify the intuition of this direct (but computationally expensive) approach. Within the same framework, we then devise a fast and consistent algorithm, Total Conditioning (TC), and a variant, TCbw, with an explicit backward feature-selection heuristics, for Gaussian data. After running a series of comparative experiments on five artificial networks, we argue that Markov blanket algorithms such as TC/TC bw or Grow-Shrink scale better than the reference PC algorithm and provides higher structural accuracy. ©2008 Jean-Philippe Pellet and André Elisseeff.
Alain Vaucher, Philippe Schwaller, et al.
AMLD EPFL 2022
Ran Iwamoto, Kyoko Ohara
ICLC 2023
Yehuda Naveli, Michal Rimon, et al.
AAAI/IAAI 2006
Segev Shlomov, Avi Yaeli
CHI 2024