Towards Efficient Local Causal Structure Learning
Abstract
Local causal structure learning aims to discover and distinguish direct causes (parents) and direct effects (children) of a variable of interest from data. While emerging successes have been made, existing methods need to search a large space to distinguish direct causes from direct effects of a target variable T. To tackle this issue, we propose a novel Efficient Local Causal Structure learning algorithm, named ELCS. Specifically, we first propose the concept of N-structures, then design an efficient Markov Blanket (MB) discovery subroutine to integrate MB learning with N-structures to learn the MB of T and simultaneously distinguish direct causes from direct effects of T. With the proposed MB subroutine, ELCS starts from the target variable, sequentially finds MBs of variables connected to the target variable and simultaneously constructs local causal structures over MBs until the direct causes and direct effects of the target variable have been distinguished. Using eight Bayesian networks the extensive experiments have validated that ELCS achieves better accuracy and efficiency than the state-of-the-art algorithms.
Index Terms:
Bayesian network, Markov Blanket, Local causal structure learning.1 Introduction
CAUSAL discovery has always been an important goal in many scientific fields, such as medicine, computer science and bioinformatics [1, 2, 3, 4]. There has been a great deal of recent interest in discovering causal relationships between variables, since it is not only helpful to reveal the underlying data generating mechanism, but also to improve classification and prediction performance in both static and non-static environments [5]. However, in many real-world scenarios, it is difficult to discover causal relationships between variables since true causality can only be identified using controlled experimentation [6].
Statistical approaches are useful in generating testable causal hypotheses which can accelerate the causal discovery process [7]. Learning a Bayesian network (BN) from observational data is the popular method for causal structure learning and causal inference. The structure of a BN takes the form of a directed acyclic graph (DAG) in which nodes of the DAG represent the variables and edges represent dependence between variables. A DAG implies causal concepts, since they code potential causal relationships between variables: the existence of a directed edge XY means that X is a direct cause of Y, and the absence of a directed edge XY means that X cannot be a direct cause of Y [8]. When a directed edge XY in a BN indicates that X is a direct cause of Y, in this case, the BN is known as a causal Bayesian network. Given a set of conditional dependencies from observational data and a corresponding DAG model, we can infer a causal Bayesian network using intervention calculus [9]. Then learning BN structures (i.e. DAGs) from observational data is the most important step for causal structure learning.
In recent years, many causal structure learning (i.e. DAG learning) methods have been designed [10], which can be roughly divided into global causal structure learning and local causal structure learning. The first type of methods aims to learn the casual structure of all variables, such as MMHC [11], NOTEARS [12] and DAG-GNN [13]. However, in many practical scenarios, it is not necessary to waste time to learn a global structure when we are only interested in the causal relationships around a given variable. To tackle this issue, the second type of methods is proposed, with the aim to discover and distinguish the direct causes (parents) and direct effects (children) of a target variable, such as PCD-by-PCD [14] and CMB [15].
PCD-by-PCD (PCD means Parents, Children and some Descendants) [14] and CMB (Causal Markov Blanket) [15] first learn the PCD or MB (Markov Blanket) of a target variable and construct a local structure among the target variable and the variables in the PCD or MB, then sequentially learn PCDs or MBs of the variables connected to the target variable and simultaneously construct local structures among variables in PCDs or MBs until the parents and children of the target variable have been distinguished.
While emerging successes have been made, existing local causal structure learning methods suffer from the following limitations. They need to search a large space to distinguish parents from children of a target variable. That is to say, existing local causal structure learning methods not only need to learn the PCD or MB of the target variable, but also may need to learn PCDs or MBs of the variables connected to the target variable. In the worst case (e.g. the target variable has all single ancestors) all existing methods may be required to learn PCDs or MBs of all variables in a dataset. This leads to that existing local causal structure learning methods are often computationally expensive or even infeasible especially with a large-sized BN. For instance, as shown in Fig. 1, there is an N-structure (see Definition 9) formed by four variables T, A, B, and C. Given the target variable T, in order to determine the causal relationship between T and B: PCD-by-PCD is required to learn the PCD of T and PCDs of A and B. For CMB, if only the MB of T is learnt, the edge direction between T and B cannot be determined since there is no V-structures around B. CMB needs to further learn the MBs of B and A to orient the edge between T and B. In a word, both PCD-by-PCD and CMB need to search a large space to determine the edge direction between T and B. A larger search space will result in performing more conditionally independence (CI) tests for discovering the causal relationships around a given variable. More CI tests not only increase computational time, but also lead to more unreliable tests. It will be beneficial to local causal structure learning if the edge direction between B and T can be determined in learning the PCD or MB of the target variable T without learning PCDs or MBs of the other variables.

Then a question naturally arises: can we reduce the search space in determining the edge directions between a given variable and its children to speed up the local causal structure learning? To address this problem, our main contributions of the paper are summarized as follows.
-
•
We propose the concept of N-structures, a special local structure for edge directions in local causal structure learning. Then we propose a new local causal structure learning, called ELCS. Through leveraging the N-structures, ELCS learns the MBs of the variables as few as possible to distinguish parents from children of a given variable as many as possible, which improves the efficiency of local causal structure learning and simultaneously reduces the impact of unreliable CI tests.
-
•
To integrate MB learning with N-structures to infer edge directions as many as possible during the MB learning procedure, we design an efficient MB discovery subroutine (EMB) and its efficient version EMB-II. EMB not only is able to learn the MB of a variable, but also has an ability to distinguish parents from children of the variable.
-
•
We have conducted extensive experiments on eight benchmark BNs, and have compared ELCS with five existing causal structure learning algorithms, including three state-of-the-art global structure learning and two local structure learning algorithms, to demonstrate the effectiveness and efficiency of the ELCS algorithm.
The remainder of this paper is organized as follows. Section 2 reviews the related work, and Section 3 gives the notations and definitions. Section 4 describes the proposed ELCS algorithm in detail. Section 5 reports and discusses the experimental results. Section 6 summarizes the paper.
2 Related Work
Our work focuses on local causal structure learning and is also related to MB learning and global causal structure learning. So this section briefly introduces the related work in the three areas.
MB learning. Learning Markov Blanket (MB) plays an essential part in the skeleton learning during BN structure learning. Existing MB learning methods can be categorized into two types: constraint-based methods and score-based methods. The former employs independence tests to find the MB of a given variable [16, 17], whereas the latter learns the MB using score-based BN structure learning algorithms [18, 19].
Constraint-based methods can be roughly grouped into simultaneous MB learning and divide-and-conquer MB learning. Given the target variable T, the simultaneous MB learning algorithm aims to learn parents, children, and spouses of T simultaneously, and does not distinguish spouses of T from its PC, such as GSMB [20], IAMB [21], Inter-IAMB [22] and Fast-IAMB [23]. To reduce the sample requirement of the simultaneous MB learning algorithm, the divide-and-conquer MB learning algorithm is proposed, with the aim to find PC and spouses of the target variable separately. The representative divide-and-conquer MB learning algorithms include CCMB [16], BAMB [17], MMMB [24], HITON-MB [25], PCMB [26] and STMB [27]. Recently, a comprehensive review of the state-of-the-art MB learning algorithms are discussed in [28].
However, existing MB learning methods only learn a local skeleton around a target variable and do not distinguish parents from children in the learnt MB of a target variable.
Global causal structure learning. A large amount of methods have been designed for global causal structure learning. Recent methods can be roughly categorized into two types: local-to-global structure learning methods and continuous optimization based learning methods. The local-to-global structure learning approach, such as MMHC [11], SSLC/G [18] and GSBN [29], first learns the MB or PC of each variable, then constructs a skeleton of a DAG using the learnt MBs or PCs, and finally orients edges of the learnt skeleton using score-based or constraint-based causal learning algorithms. Instead of learning the MB of each variable first, GGSL [30] starts with a randomly selected variable, and then uses a score-based MB learning algorithm to gradually expand the learnt structure through a series of local structure learning steps. Based on GGSL, a parallel BN structure learning algorithm (PSL) is designed to improve the efficiency [31].
Recently, several continuous optimization based learning approaches have been proposed for global causal structure learning [13, 12, 32, 33]. Zheng et al. consider a BN structure learning problem as a purely continuous optimization problem and propose the NOTEARS algorithm [12]. DAG-GNN uses a graph neural network based deep generative model to capture the complex data distribution to learn BN structures [13]. RL-BIC uses reinforcement learning to search for a directed acyclic graph (DAG) with the best score [32]. Zhang et al. propose a DAG variational autoencoder (D-VAE) for BN structure learning [33].
However, global causal structure learning methods are time consuming or even infeasible when the number of variables of a BN is large. In fact, in many practical settings, we are only interested in distinguishing parents from children of a variable of interest. In this case, it is unnecessary and wasteful to find an entire BN structure.
Local causal structure learning. Local causal structure learning aims to learn and distinguish the parents and children of a target variable. Although many algorithms have been designed for learning a whole structure, only several algorithms have been proposed for local causal structure learning. PCD-by-PCD first discovers the PCD of a target variable, then sequentially discovers PCDs of the variables connected to the target variable and simultaneously finds V-structures and orients the edges connected to the target variable until all the parents and children of the target variable are identified [14]. CMB first learns the MB of a target variable using HITON-MB and orients edges by tracking the conditional independence changes in the MB of the target variable, then sequentially learns MBs of the variables connected to the target variable and simultaneously construct local structures along the paths starting from the target variable until the parents and children of the target variable have been identified or they cannot be identified further by continuing the process [15].
As we discussed in Section 1, both PCD-by-PCD and CMB encounter the time inefficient problem since they need to learn a large number of PCDs or MBs of the variables for distinguishing parents from children of a target variable. To tackle this issue, in this paper, we aim to develop a new method through learning the MBs of variables as few as possible while orienting edges as many as possible.
3 Notations and Definitions
In this section, we will briefly introduce some basic definitions and notations frequently used in this paper (see Table I for a summary of the notations). Let U denote a set of random variables. represents a joint probability distribution over U, and is a DAG over U. In a DAG, X is a parent of Y and Y is a child of X if there exists a directed edge from X to Y. X is an ancestor of Y (i.e., non-descendant of Y) and Y is a descendant of X if there exists a directed path from X to Y.
Notation | Meanings |
---|---|
U | a set of random variables |
W | a subset of U |
a joint probability distribution over U | |
a direct acyclic graph over U | |
DAG | direct acyclic graph |
X, Y, Z, T | a single variable in U |
Z,S | a conditioning set within U |
XY | X and Y are independent given Z |
XY | X and Y are dependent given Z |
MB | Markov Blanket of T |
PC | a set of parents and children of T |
P | a set of parents of T |
C | a set of children of T |
UN | undistinguished variables in PC |
SP | a set of spouses of T |
SP{X} | a spouses of T with regard to T’s child X |
Sep{X} | a set that d-separates X from T |
Sep | a set that contains the sets Sep{} of all variables |
CSP | a set that contains the candidate spouse sets of all PC variables |
Que | a circular queue(first in fist out) |
the size of a set |
Definition 1 (Conditional Independence [34]).
Given a conditioning set Z, X is conditionally independent of Y if and only if .
Definition 2 (Bayesian Network [34]).
The triplet U,, is called a Bayesian network (BN) if U,, satisfies the Markov condition: each variable is conditionally independent of variables in its non-descendant given its parents in .
Definition 3 (Casual Bayesian Network [9]).
A BN is called a causal Bayesian network (CBN) if a directed edge in has causal interpretation, that is, XY indicates that X is a direct cause of Y.
Definition 4 (Causal Structure Learning).
Global causal structure learning aims to learn a DAG over U from observational data, where edges represent potential causal relationships between variables, that is, X is a direct cause of Y if there exists a directed edge from X to Y [9]. Local causal structure learning aims to discover and distinguish direct causes and direct effects of a variable of interest [15].
Definition 5 (V-structure [34]).
If there is no an edge between X and Y, and Z has two incoming edges from X and Y, respectively, then X, Z and Y form a V-structure ().
In a BN, Z is a collider if there are two directed edges from X to Z and from Y to Z, respectively. V-structures play an important role in determining the edge directions between variables. For example, if there is a V-structure () formed by X, Y and Z, we can identify X and Y as parents of Z using conditional independence (CI) tests.
Definition 6 (D-separation [34]).
Given a set S U{X,Y}, a path between X and Y is blocked, if one of the following conditions is satisfied: 1) there is a non-collider variable within S on , or 2) there is a collier variable Z on , while Z and any its descendants are not in S. Otherwise, between X and Y is unblocked. X and Y are d-separation given S if and only if each path between X and Y is blocked by S.
In a DAG, given a conditioning set, we can determine whether two variables are conditionally independent using Definition 6.
Definition 7 (Faithfulness [35]).
Given a BN U,,, is faithful to if and only if all the conditional independencies appear in are entailed by . is faithful if and only if there is a DAG such that is faithful to .
Definition 7 indicates that in a faithful BN, if X and Y are d-separated given the conditioning set S in , then they will be conditionally independent given S in .
Definition 8 (Markov Blanket [34]).
In a faithful BN, the MB of a target variable T is denoted as MBT, which is uniqueness and consists of parents, children and spouses (other parents of the target’s children) of T. All other variables in U MBT{T} are conditionally independent of T given MBT, X U MBT{T}, X T MBT, where X T MBT denotes X and T are conditionally independent conditioning on MBT.
Definition 9 (N-structure).
In a faithful BN, if there exists four variables T, A, B and C, and T is a parent of A and B, C is a parent of A, there is no an edge between C and T, A is an ancestor of B, the other parents of B are in PC set of T. Then, the local structure formed by the four variables is called an N-structure.

Fig. 2 gives examples of the N-structures. In Fig. 2 (a), there is an N-structure formed by T, A, B and C. In Fig. 2 (b), variables T, A, B and C construct an N-structure, and variables T, A, E and C construct an N-structure. Given the target variable T, we can leverage the N-structures to determine edge directions between T and its children (i.e. B and E) during learning the MB of T without learning MBs of the other variables.
Theorem 1 [35].
In a faithful BN, for any two variables X U and Y U, if there exists an edge between X and Y, then S U {X,Y}, XY S holds.
Theorem 1 demonstrates that if X is a parent (or a child) of Y if X and Y are not conditionally independent conditioning any subsets excluding X and Y.
4 The Proposed Method
4.1 The ELCS Algorithm
We propose the Efficient Local Causal Structure learning algorithm (ELCS) to distinguish parents from children of a target variable, as shown in Algorithm 1. ELCS starts from the target variable, sequentially finds MBs of variables connected to the target variable and simultaneously constructs local causal structures over MBs until all the parents and children of the target variable have been distinguished or it is clear that they cannot be further distinguished by continuing the process. In the following, we first summarize the main idea of ELCS, then give the details of ELCS.
To improve the efficiency of local causal structure learning, in ELCS, we propose the following two acceleration strategies. First, ELCS finds the N-structures, and then leverages those found N-structures to infer edge directions between the target variable T and its children during learning the MB of T. Second, two rules in Lemma 1 are used to further infer edge directions between T and its PC during learning the MB of T.
As described in Algorithm 1, given the target variable T, ELCS first initializes the variable set W and the queue Que to empty sets (line 1 in Algorithm 1), where W is used to store variables that their MBs have been learnt, and Que is utilized to store variables that their MBs need to be learnt in next phase. Then, T enters Que (line 2 in Algorithm 1). Next, lines 4-13 in Algorithm 1 will be executed. At line 4 in Algorithm 1, the header element in Que is out of queue, that is, X = T. Since the MB of T has not been learnt, T is added to the set W (lines 5-6 in Algorithm 1). The EMB (Efficient Markov Blanket discovery) subroutine is executed to learn the MB of T (line 7 in Algorithm 1). Given a variable X, the EMB subroutine not only is able to find the PC (PC) and MB (SP PC), but also has an ability to distinguish parents from children of X. The details of EMB are described in Section 4.2. Let P represent a set of the identified parents of X. C denotes a set of the identified children of X. The set containing undistinguished PC variables of X is denoted as UN. After executing the line 7 in Algorithm 1, the sets of P, C, UN, SP, PC are obtained. If UN is empty, that is, parents and children of T are all distinguished, the learning process will be terminated. Otherwise, the undistinguished variables within UN will be put in Que (lines 9-11 in Algorithm 1). Then, Meek rules [36] are used to orient other edge directions between variables in W (line 13 in Algorithm 1). Lines 4-13 in Algorithm 1 will be repeated until all the parents and children of T have been distinguished or Que is empty or the size of W equals to that of the entire variable set.
4.2 EMB Subroutine
ELCS depends on the MB learning methods for local causal structure learning, but existing MB learning algorithms have the following shortcomings. First, existing MB learning methods cannot be directly combined with the N-structures to infer edge directions between the target variable and its children. Second, existing MB learning methods only focus on learning the MB of the target variable and are not able to distinguish parents from children. Third, existing MB learning methods may be computationally expensive. In order to help ELCS to leverage N-structures and the rules in Lemma 1 for efficiently learning local causal structures, we design an Effective MB discovery subroutine (EMB) to learn the MB of a target variable and distinguish parents from children of the target variable simultaneously.
As shown in Algorithm 2, EMB consists of four steps as follows. Given a target variable T, EMB first learns the PC of T (PC) using an existing PC learning algorithm. Second, EMB obtains spouses of T using a RecogSpouses subroutine. Then, EMB removes false PC from PC. Finally, EMB orients edges between T and its PC as many as possible using a DistinguishPC subroutine. Specifically, to find the N-structures, EMB first determines which variable within U{T}PC is a candidate spouse of T, and obtains the candidate spouse set CSP{Y} of each variable Y within PC, and then obtains the spouses of T. Based on the learnt CSP{Y} and spouses, we can find the N-structures. Through leveraging the found N-structures, EMB can distinguish some children of T with regard to the found N-structures. In addition, two rules in Lemma 1 are used in the DistinguishPC subroutine to further distinguish parents from children of T. In the following, we will give the details of these four steps.
Lemma 1.
The PC (parents and children) set of a given variable T (T U) is denoted as PCT. Let X PCT, Y PCT. We can get the following two dependence relationships between X and Y.
(a) X Y and X Y T X and Y are both parents of T. This shows that there is a V-structure () formed by variables X, Y and T, and T is a collider.
(b) X is a direct cause of T, X Y and X Y T Y is a direct effects of T. This shows that there is only one path () from X to Y, and the path is blocked by T.
Step 1 (line 1 in Algorithm 2): EMB obtains PC and Sep of a target variable T by utilizing an existing PC learning algorithm, where Sep is a set that contains the sets Sep{} of all variables. In this paper, we use HITON-PC [25] to find the PC of T (any other state-of-the-art PC learning algorithms can be used here to instantiate the RecogPC() function at Step 1 in Algorithm 2).
Step 2 (line 2 in Algorithm 2): At this step, EMB learns spouses of T. We design a RecogSpouses subroutine for learning spouses. The details of RecogSpouses are described in Algorithm 3. RecogSpouses first finds candidate spouses from all variables within U{T}PC that are conditionally independent of T. If X and T are conditionally independence, then we construct a set Temp that consists of variables which belong to PC and are dependent of X given an empty set (lines 3-8 in Algorithm 3). If X and T are conditionally independent conditioning on Temp, then X cannot be a spouse of T. Otherwise, X is regarded as a candidate spouse and lines 9-15 in Algorithm 3 will be executed. If X and T are dependent conditioning on Sep{X} Y (Y Temp), then X will be added to CSP{Y} (lines 9-15 in Algorithm 3). Since some non-parent variables of Y will be added to CSP{Y}, non-parent variables of each Y PC will be removed from CSP{Y} and the spouse set SP{Y} will be obtained after executing lines 17-24 in Algorithm 3.
Step 3 (lines 3-8 in Algorithm 2): At this step, EMB removes false positives from the candidate set of PC of T. For each variable Y within PC, if there exists a subset Z of the union SP{Y} PC such that Y and T are conditionally independent conditioning on Z, Y will be removed from PC, and SP{Y} will be set to an empty set.
Step 4 (line 9 in Algorithm 2): At this step, EMB distinguishes parents from children of T as many as possible. We propose a DistinguishPC subroutine to accomplish this goal. DistinguishPC first identifies some children of T with the help of spouses of T. Second, DistinguishPC uses the found N-structures to infer edge directions between T and its children. Finally, DistinguishPC distinguishes parents from children of T using Lemma 1.
The details of DistinguishPC are described in Algorithm 4. First, DistinguishPC uses the learnt spouses to identify some children of T (lines 2-6 in Algorithm 4). For example, in Fig. 1, C and D are spouses of T, SP{A} = {C}, SP{K} = {D}, and DistinguishPC identifies A and K as children of T. There exists an N-structure which is formed by C, A, B and T in Fig. 1, DistinguishPC can determine the edge direction between T and B with the help of SP and CSP. At Step 2, C will be added to CSP{A} and CSP{B}, and C will be removed from SP{B} because C is not a parent of B (lines 17-24 in Algorithm 3). Since C is a spouse of T and C is within the set CSP{B}, DistinguishPC identifies B as a child of T (lines 8-12 in Algorithm 4). Theorem 2 gives the theoretical analysis. In addition, in order to orient more edges between T and its PC, two rules in Lemma 1 are used. If X Y and X Y T, we can conclude that X and Y are both parents of T. Therefore, DistinguishPC identifies parents of T as many as possible using Lemma 1 (a) (lines 13-19 in Algorithm 4). If X is a parent of T, X Y and X Y T, then Y is a child of T. DistinguishPC uses the identified parents of T to determine edge directions between T and its children using Lemma 1 (b) (lines 21-28 in Algorithm 4).
We also propose a variant of EMB to further improve the efficiency of MB learning, which is referred to as EMB-II. Compared with EMB, in learning MB, instead of directly executing line 20 in Algorithm 3, EMB-II first ranks the variables within SP{Y} in descending order according to the dependency with variable Y, then executes line 20 in Algorithm 3.
We also propose a variant of ECLS to further improve the efficiency of local causal structure learning, which is referred to as ELCS-II. Compared with ECLS, ELCS-II uses EMB-II for MB learning instead of using EMB.
In the following, we will give the details of Theorem 2 and its proof.
Theorem 2.
In a faithful BN, given an N-structure consisting of four variables T, A, B, C. T is a parent of A and B, C is a parent of A, there is no an edge between C and T, A is an ancestor of B, and the parents of B are in PC set of T, then EMB identifies B as a child of T during learning the MB of T.
Proof.
Under the faithfulness assumption, the PC of a target variable T only contains parents and children of T. Since C is a parent of A, there is no an edge between C and T, then EMB identifies C as a spouse of T since there is a V-structure () around A. At step 2 of EMB, C will be added to CSP{B} (lines 2-16 in Algorithm 3) since C and T are conditionally dependent given the conditioning set {B} Sep{C}. If B is a parent of T, then C and T are conditionally independent given the conditioning set {B} Sep{C}, since all paths from T to C are blocked by conditioning set {B} Sep{C}. Therefore, B is a child of T. ∎
4.3 Tracing
We first trace the execution of EMB using the example in Fig. 3, then trace the execution of ELCS using the example in Fig. 4.
4.3.1 Tracing EMB

We utilize the example in Fig. 3 to trace the execution of EMB. Suppose that we have a dataset for variable set U = {A, B, C, D, E, F, X, H, I, J, K, L, T}. The independence relationships between variables can be represented by the Bayesian Network structure in Fig. 3. In the following, we regard T as the target variable, and give the execution process of EMB.
(1) step 1: referring to the simple network, i.e., the left network in Fig. 3. First, HITON-PC is used to find the PC of T. According to Theorem 1, A, B, L, K, E and J will be added to PC. Note that D is conditionally independent of T given an empty set, hence D will not be in any of the conditioning sets for higher order conditional independence tests. As a result, I will be added to PC since T and I are conditionally dependent given conditioning set {K}. Then, as shown in Fig. 3 (a), PC = {A, B, L, K, E, J, I}.
(2) step 2: as shown in Fig.3 (b), EMB discovers the spouses of T. X and each variable within Temp = {A, B, L, K, E, I} are conditionally dependent given an empty set, while X is conditionally independent of T given the conditioning set Temp, so that X cannot be a candidate spouse of T since each path from X to T is blocked by Temp. Similarly, both F and H are not candidate spouses. C and each variable within Temp = {A, B, L, K, E, I} are conditionally dependent given an empty set, and C is dependent of T given Temp, and we need to conduct further tests. Owning to C T E, C T {E, A}, C T {E, B}, C T {E, L}, C T {E, K} and C T {E, I}, hence C is added to CSP{A} and CSP{B}, CSP{A} = {C}, CSP{B} = {C}. Similarly, D is added to CSP{K} and CSP{I}, CSP{K} = {D}, CSP{I} = {D}. In the following, C will be removed from SP{B} since C B {A, T} (lines 17-24 in Algorithm 3). After this step, SP{A} = {C}, SP{K} = {D}, SP{I} = {D}.
(3) step 3: as shown in Fig. 3 (c), after checking at line 4 in Algorithm 2, I will be removed from PC since I T {K, D}. After this step, PC = {A, B, L, K, E, J}, SP = {C, D}, and MB ={A, B, L, K, E, J, C, D}.
(4) step 4: as shown in Fig. 3 (d), EMB orients the edge directions between T and its PC as many as possible. Since C is a spouse of T, and C has been added to SP{B} at step 2, based on Theorem 2, B is a child of T. In addition, according to Lemma 1, E and J are parents of T since E J and E J T. L is a child of T since E is a parent of T, E L and E L T.
4.3.2 Tracing ELCS
We use the example in Fig. 4 to trace the execution of ELCS.

Suppose that we have a dataset for variable set U = {A, B, C, D, E, F, X, H, I, J, K, L, T, Y}. The independence relationships between variables can be represented by the BN structure in Fig.4. In the following, we regard T as the target variable, and give the execution process of distinguishing parents from children of T using ELCS. We use the = -1 to represent that X is a parent of Y, = 1 represents that X is adjacent to Y, = 0 represents that there is no an edge between X and Y.
(1) step 1: referring to the simple network, i.e., the left network in Fig. 4. We first use EMB to distinguish parents from children of T. After learning the MB of T, as shown in Fig. 4 (a), the PC = {A, B, L, K, E, J, Y} and SP = {C, D} are obtained. Then, = -1, = -1, = -1, = -1, = -1, = -1 and = 1. The edge direction between Y and T is unsure.
(2) step 2: to resolve , as shown in Fig. 4 (b), we need to make a further search. In the following, the MB of Y is extracted using EMB, and = -1, = -1. After updating the current local structure using Meek rules, we can learn that Y is a parent of T, that is, = -1.
4.4 Theoretical Analysis
In the following, we first theoretically analyze the correctness of EMB, then theoretically analyze the correctness of ELCS.
Theorem 3 (Correctness of EMB).
Under the faithfulness assumption, and all CI tests are reliable, EMB finds all and only the MB of a given variable.
Proof.
At step 1, EMB finds all the true PC variables. According to Theorem 1, the variables which are dependent with the target variable T will be added to PC. PC contains the true parents and children of T, since the true PC variables are always dependent of T. In addition, PC also contains some descendants of T [27]. Then, based on the results of step 1, EMB finds all the true spouses of T at step 2. If Y is a collider, X is regarded as a candidate spouse if there exists a V-structure () formed by X, Y and T, and X will be added to CSP{Y} (lines 1-16 in Algorithm 3). Owning to an exhaustive search, EMB will not miss any true spouses of T. In the following, the variable X SP{Y} that is a non-parent of Y will be removed from SP{Y} (lines 17-24 in Algorithm 3). According to the Markov condition, the variable X SP{Y} will be removed if it is not a parent of Y, since conditioning set SP{Y} T PC {X,Y} contains all true parents of Y. Therefore, SP contains all the true spouses of T. The learnt PC may contain some false PC variables. T and each false PC variable are conditionally independent given the spouses of the false PC variable and PC. At the step 3, the false PC variables found at step 1 and false spouses found at step 2 will be removed (lines 3-8 in Algorithm 2). Then, EMB contains all and only the true PC variables PC and spouses SP after executing Algorithm 2, and PC and SP together form all and only true MB variables. At step 4, based on Theorem 2 and Lemma 1, parents and children of T are learnt by EMB are correct.
∎
Theorem 4 describes the correctness of the proposed ELCS algorithm. In the following, we will introduce Theorem 4 and its proof in detail.
Theorem 4 (Correctness of ELCS).
Under the faithfulness assumption, and all CI tests are reliable, ELCS distinguishes all parents from children of a given variable.
Proof.
Under the causal faithfulness assumption, given a target variable, EMB finds all and only the true MB variables and the true PC variables of the target variable. The learnt PC set contains all and only the parents and children of the target variable. Based on Definition 5, the children identified by the learnt spouses are correct. Base on Theorem 2, the children identified by the found N-structures are correct. All the parents and children identified by Lemma 1 are correct. ELCS updates the local causal structures until the parents and children of the target variable have been distinguished. Meek rules [36] are used to orient other undirected edges between the target variable and the variables that are adjacent to the target variable during learning the local causal structure, and all the edge directions determined by Meek rules are correct. Thus, all the parents and children of a given target variable distinguished by ELCS are correct. ∎
4.5 Computational Complexity
The computational complexity of ELCS algorithm depends on the steps of discovering MB. In the following, we will give the computational complexity of EMB and ELCS.
The computational complexity of EMB: The computational complexity of EMB depends on its four steps. Given a target variable T, at step 1, the computational cost is dominated by HITON-PC which takes at most CI tests to find the PC. Step 2 takes CI tests, step 3 takes CI tests, step 4 takes CI tests. Thus, the complexity of EMB is = . Let represent the largest size of the PC sets of all the variables, the complexity of EMB is .
We summarize the computational complexity of EMB and its rivals in Table II. From the table, IAMB is the fastest among all MB learning algorithms. MMMB and HINTON-MB are the slowest two algorithms, since they need to find the PC sets of all target variable’s parents and children. The computational complexity of EMB is lower than STMB. In general, most of the BNs have a large number of variables but a small-sized PC set of each variable, so that EMB is faster than STMB. The computational complexity of BAMB is the same with EMB.
The computational complexity of ELCS: In the best case, the complexity of ELCS is . But in the worst case (e.g. the target variable has all single ancestors), ELCS needs to learn a whole structure, hence the complexity of ELCS is .
Table III reports the computational complexity of ELCS and its rivals. Note that is the memory size of L-BFGS [12]. and are the number of samples in data and iterations, respectively. is the number of neurons in the hidden layer [13]. The computational complexity of global causal structure learning algorithms is higher than that of local causal structure learning algorithms, since they learn the whole structure of all variables. PCD-by-PCD uses MMPC for PC learning, and CMB uses HITON-MB for MB learning. In the best case, the complexity of PCD-by-PCD is consistent with that of MMMB, it is . The complexity of CMB is consistent with that of HITON-MB, it is . In the worst case, since MMPC can discover the separating sets while learning PC, the complexity of PCD-by-PCD is consistent with that of using MMPC to find PCs of all variables, it is . The complexity of CMB is consistent with that of using HITON-MB to find MBs of all variables, it is . ELCS is the fastest algorithm in both best and worst cases, since the complexity of ELCS relies on that of EMB which only takes CI tests to find the MB.
Algorithms | IAMB | MMMB | HITON-MB | STMB | BAMB | EMB |
---|---|---|---|---|---|---|
Complexity |
Algorithm | Worst Case | Best Case |
---|---|---|
PCD-by-PCD | ||
CMB | ||
ELCS | ||
MMHC | ||
NOTEARS | ||
DAG-GNN | ) |
5 Experiments
In this section, we evaluate the performance of the proposed ELCS algorithm, and this section is organized as follows. Section 5.1 gives the experimental settings, Section 5.2 summarizes and discusses the experimental results, Section 5.3 analyses the reason why ELCS is efficient and effective.
5.1 Experimental Settings
5.1.1 Datasets
We use eight benchmark BNs to evaluate ELCS against its rivals. Each benchmark BN contains two groups of data, one group containing 10 data sets with 5000 data examples, and the other one including 10 data sets with 1000 data examples. The number of variables of these BNs ranges from 20 to 801. A brief description of the eight benchmark BNs is listed in Table IV.
5.1.2 Comparison Methods
5.1.3 Implementation Details
PCD-by-PCD and CMB algorithms are implemented by ourselves in MATLAB (https://github.com/kuiy/CausalLearner). For MMHC, we use the implementation in the software tool of Causal Explorer [37]. For NOTEARS and DAG-GNN, we use the source codes provided by the authors. In the experiments, -test with the significance level of 0.01 is utilized to measure the conditional independence between variables. All experimental results are conducted on Windows 10 with Intel(R) i7-8700, 3.19 GHz CPU, and 16GB memory.
5.1.4 Evaluation Metrics
In the experiments, we evaluate the performance using the following metrics.
-
•
ArrP: The number of true directed edges in the output (i.e., the variables in the output belonging to the true parents and children of a target variable in a test DAG) divided by the number of edges in the output of an algorithm.
-
•
ArrR: The number of true directed edges in the output divided by the number of true directed edges (i.e., the number of parents and children of a target variable in a test DAG).
-
•
SHD: SHD is the number of total error edges, which contains undirected edges, reverse edges, missing edges and extra edges. The smaller value of SHD is better.
-
•
FDR: The number of false edges in the output divided by the number of edges in the output of an algorithm.
-
•
Efficiency: The number of CI tests and the running time (in seconds) are utilized to measure the efficiency.
In the following Tables, the results are reported in the format of , where denotes the average results, and represents the standard deviation. The best results in each setting have been marked in bold. ”-” means that the output of the corresponding BN cannot be generated in two days by the algorithm. Note that NOTEARS and DAG-GNN do not perform CI tests.
5.2 Experimental Results of ELCS and Its Rivals
|
|
|
|
|
|
|||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Alarm | 37 | 46 | 4 / 5 | 1 / 6 | 2 - 4 | |||||||||||
Insurance | 27 | 52 | 3 / 7 | 1 / 9 | 2 - 5 | |||||||||||
Child | 20 | 25 | 2 / 7 | 1 / 8 | 2 - 6 | |||||||||||
Alarm10 | 370 | 570 | 4 / 7 | 1 / 9 | 2 - 4 | |||||||||||
Insurance10 | 270 | 556 | 5 / 8 | 1 / 11 | 2 - 5 | |||||||||||
Child10 | 200 | 257 | 2 / 7 | 1 / 8 | 2 - 6 | |||||||||||
Pigs | 441 | 592 | 2 / 39 | 1 / 41 | 3 - 3 | |||||||||||
Gene | 801 | 972 | 4 / 10 | 0 / 11 | 3 - 5 |
We compare ELCS with MMHC, NOTEARS, DAG-GNN, PCD-by-PCD, CMB and ELCS-II on the eight BNs as shown in Table IV. The average results of ArrP, ArrR, SHD, FDR, CI tests and running time of each algorithm are reported in Tables V-VI. Table V summarizes the experimental results on the eight BNs with 5,000 data examples, and Table VI reports the experimental results on the eight BNs with 1,000 data examples. From the experimental results, we have the following observations.
Network | Algorithm | ArrP() | ArrR() | SHD() | FDR() | CI Tests() | Time() |
---|---|---|---|---|---|---|---|
Alarm | MMHC | 0.190.02 | 0.080.02 | 4.580.02 | 0.600.01 | 138604971 | 7.032.74 |
NOTEARS | 0.610.01 | 0.740.04 | 2.840.17 | 0.480.02 | - | 541.9527.18 | |
DAG-GNN | 0.660.03 | 0.540.04 | 2.020.17 | 0.360.05 | - | 1.1e31.1e2 | |
PCD-by-PCD | 0.770.03 | 0.640.05 | 0.940.15 | 0.270.04 | 211092 | 0.810.04 | |
CMB | 0.770.05 | 0.720.06 | 0.760.14 | 0.220.04 | 2111207 | 0.690.07 | |
ELCS | 0.860.01 | 0.810.01 | 0.440.06 | 0.070.02 | 64855 | 0.200.02 | |
ELCS-II | 0.860.01 | 0.810.01 | 0.440.06 | 0.070.02 | 60752 | 0.190.02 | |
Insurance | MMHC | 0.210.02 | 0.030.02 | 5.870.17 | 0.670.02 | 2603271 | 1.180.12 |
NOTEARS | 0.460.01 | 0.240.01 | 4.640.12 | 0.720.02 | - | 420.8922.92 | |
DAG-GNN | 0.540.03 | 0.210.01 | 4.350.25 | 0.670.05 | - | 518.8437.24 | |
PCD-by-PCD | 0.680.02 | 0.450.02 | 2.070.09 | 0.340.03 | 3038300 | 1.480.16 | |
CMB | 0.700.04 | 0.540.04 | 2.310.25 | 0.370.05 | 115534827 | 5.442.29 | |
ELCS | 0.850.04 | 0.690.04 | 1.610.06 | 0.180.05 | 1686276 | 0.750.12 | |
ELCS-II | 0.850.04 | 0.690.04 | 1.610.06 | 0.180.05 | 1637275 | 0.750.12 | |
Child | MMHC | 0.220.03 | 0.190.07 | 3.630.25 | 0.480.03 | 8600632 | 5.320.46 |
NOTEARS | 0.520.02 | 0.390.03 | 2.990.17 | 0.700.03 | - | 140.7436.59 | |
DAG-GNN | 0.500.04 | 0.290.06 | 2.080.10 | 0.440.06 | - | 384.7330.76 | |
PCD-by-PCD | 0.710.02 | 0.590.04 | 0.860.09 | 0.260.04 | 243278 | 1.240.04 | |
CMB | 0.820.05 | 0.750.08 | 0.720.18 | 0.250.08 | 94244106 | 4.581.96 | |
ELCS | 0.710.12 | 0.610.16 | 1.080.36 | 0.090.08 | 2093287 | 0.930.10 | |
ELCS-II | 0.710.12 | 0.610.16 | 1.080.36 | 0.090.08 | 2087287 | 0.930.10 | |
Alarm10 | MMHC | 0.19+0.00 | 0.02+0.00 | 5.63+0.05 | 0.63+0.00 | 9.7e7+8.9e6 | 4.6e4+4.8e3 |
NOTEARS | 0.730.01 | 0.500.01 | 2.270.04 | 0.280.01 | - | 1.6e41.8e3 | |
DAG-GNN | - | - | - | - | - | - | |
PCD-by-PCD | 0.730.01 | 0.540.01 | 1.480.03 | 0.210.01 | 257951770 | 8.180.57 | |
CMB | 0.720.01 | 0.580.01 | 1.570.04 | 0.340.01 | 14011565 | 3.690.15 | |
ELCS | 0.830.01 | 0.680.02 | 1.260.07 | 0.140.02 | 6893483 | 1.770.12 | |
ELCS-II | 0.830.01 | 0.680.02 | 1.260.07 | 0.140.02 | 6916480 | 1.770.12 | |
Insurance10 | MMHC | 0.220.00 | 0.000.00 | 6.720.04 | 0.700.00 | 1.9e52.0e4 | 81.6610.39 |
NOTEARS | 0.300.01 | 0.200.00 | 8.670.44 | 0.850.01 | - | 1.7e41.5e4 | |
DAG-GNN | - | - | - | - | - | - | |
PCD-by-PCD | 0.680.01 | 0.460.01 | 2.100.05 | 0.410.01 | 9581224 | 4.380.13 | |
CMB | 0.640.01 | 0.490.01 | 2.580.06 | 0.480.02 | 399323898 | 16.041.54 | |
ELCS | 0.800.02 | 0.670.02 | 1.750.11 | 0.230.01 | 108091528 | 3.920.55 | |
ELCS-II | 0.800.02 | 0.670.02 | 1.750.11 | 0.230.01 | 106051499 | 3.910.55 | |
Child10 | MMHC | 0.150.01 | 0.030.01 | 5.290.09 | 0.580.01 | 7.9e52.7e5 | 439.79169.28 |
NOTEARS | 0.610.01 | 0.740.04 | 2.840.17 | 0.480.02 | - | 541.9527.18 | |
DAG-GNN | - | - | - | - | - | - | |
PCD-by-PCD | 0.770.01 | 0.680.02 | 0.820.04 | 0.220.01 | 113411470 | 4.440.57 | |
CMB | 0.750.02 | 0.680.02 | 1.030.07 | 0.310.02 | 228612648 | 8.110.95 | |
ELCS | 0.830.05 | 0.760.07 | 0.730.20 | 0.140.03 | 131292613 | 3.980.79 | |
ELCS-II | 0.830.05 | 0.760.07 | 0.730.20 | 0.140.03 | 131042605 | 3.980.79 | |
Pigs | MMHC | 0.260.00 | 0.000.00 | 6.850.07 | 1.000.00 | 4.3e51.4e4 | 207.966.88 |
NOTEARS | 0.430.00 | 0.260.00 | 2.770.03 | 0.770.00 | - | 3.1e41.2e3 | |
DAG-GNN | - | - | - | - | - | - | |
PCD-by-PCD | - | - | - | - | - | - | |
CMB | - | - | - | - | - | - | |
ELCS | 0.910.00 | 0.990.00 | 0.420.02 | 0.150.01 | 133748660 | 8.916.84 | |
ELCS-II | 0.910.00 | 0.990.00 | 0.420.02 | 0.150.01 | 114675659 | 8.385.12 | |
Gene | MMHC | - | - | - | - | - | - |
NOTEARS | - | - | - | - | - | - | |
DAG-GNN | - | - | - | - | - | - | |
PCD-by-PCD | - | - | - | - | - | - | |
CMB | - | - | - | - | - | - | |
ELCS | 0.760.01 | 0.790.01 | 0.790.03 | 0.320.01 | 369507876 | 11.032.35 | |
ELCS-II | 0.760.01 | 0.790.01 | 0.790.03 | 0.320.01 | 360517696 | 11.022.08 |
Network | Algorithm | ArrP() | ArrR() | SHD() | FDR() | CI Tests() | Time() |
---|---|---|---|---|---|---|---|
Alarm | MMHC | 0.220.03 | 0.120.04 | 4.320.20 | 0.570.03 | 88844471 | 1.640.81 |
NOTEARS | 0.590.03 | 0.720.06 | 3.140.21 | 0.510.04 | - | 232.2822.13 | |
DAG-GNN | 0.480.04 | 0.260.06 | 2.010.10 | 0.170.03 | - | 241.1923.82 | |
PCD-by-PCD | 0.660.07 | 0.490.05 | 1.330.10 | 0.220.04 | 1737265 | 0.390.05 | |
CMB | 0.670.06 | 0.520.06 | 1.320.13 | 0.340.07 | 3171410 | 0.500.07 | |
ELCS | 0.720.07 | 0.610.08 | 1.060.14 | 0.110.04 | 901172 | 0.130.03 | |
ELCS-II | 0.720.07 | 0.610.08 | 1.060.14 | 0.110.04 | 861162 | 0.130.03 | |
Insurance | MMHC | 0.220.02 | 0.040.02 | 5.720.18 | 0.650.03 | 2110293 | 0.440.05 |
NOTEARS | 0.430.02 | 0.240.01 | 4.900.12 | 0.750.03 | - | 220.9744.51 | |
DAG-GNN | 0.490.06 | 0.150.05 | 3.780.19 | 0.390.13 | - | 151.7518.26 | |
PCD-by-PCD | 0.680.04 | 0.400.04 | 2.430.15 | 0.300.06 | 1370104 | 0.330.02 | |
CMB | 0.690.06 | 0.460.05 | 2.550.22 | 0.380.08 | 44571196 | 0.760.20 | |
ELCS | 0.690.12 | 0.440.15 | 2.490.40 | 0.370.19 | 1188566 | 0.170.08 | |
ELCS-II | 0.690.12 | 0.440.15 | 2.490.40 | 0.370.19 | 1106513 | 0.170.08 | |
Child | MMHC | 0.240.02 | 0.180.04 | 3.410.14 | 0.450.03 | 4583898 | 0.900.19 |
NOTEARS | 0.490.02 | 0.370.05 | 3.310.23 | 0.720.04 | - | 66.3225.78 | |
DAG-GNN | 0.340.05 | 0.150.03 | 2.200.05 | 0.290.09 | - | 87.708.96 | |
PCD-by-PCD | 0.520.05 | 0.340.06 | 1.610.12 | 0.330.08 | 2085183 | 0.390.02 | |
CMB | 0.740.09 | 0.590.10 | 1.270.29 | 0.330.12 | 49911145 | 0.650.14 | |
ELCS | 0.820.05 | 0.690.06 | 1.010.18 | 0.210.06 | 2882815 | 0.340.10 | |
ELCS-II | 0.820.05 | 0.690.06 | 1.010.18 | 0.210.06 | 2592723 | 0.330.10 | |
Alarm10 | MMHC | 0.190.00 | 0.030.01 | 5.690.07 | 0.630.00 | 3.9e63.5e5 | 700.6355.38 |
NOTEARS | 0.390.01 | 0.470.01 | 9.270.49 | 0.690.02 | - | 1.6e41.2e3 | |
DAG-GNN | - | - | - | - | - | - | |
PCD-by-PCD | 0.660.01 | 0.440.02 | 1.740.06 | 0.200.02 | 265723414 | 4.870.63 | |
CMB | 0.680.01 | 0.480.02 | 1.900.06 | 0.390.02 | 10827643 | 1.510.08 | |
ELCS | 0.750.02 | 0.530.02 | 1.720.06 | 0.200.03 | 88001218 | 1.180.16 | |
ELCS-II | 0.750.02 | 0.530.02 | 1.720.06 | 0.200.03 | 87451209 | 1.180.16 | |
Insurance10 | MMHC | 0.240.01 | 0.050.01 | 6.570.05 | 0.630.01 | 9.6e51.2e5 | 236.1125.93 |
NOTEARS | 0.200.02 | 0.200.00 | 14.110.92 | 0.910.01 | - | 9.0e31.1e3 | |
DAG-GNN | - | - | - | - | - | - | |
PCD-by-PCD | 0.630.01 | 0.370.01 | 2.660.05 | 0.460.02 | 84611809 | 1.600.35 | |
CMB | 0.620.01 | 0.450.01 | 2.950.05 | 0.450.02 | 208952158 | 3.230.33 | |
ELCS | 0.500.01 | 0.260.00 | 3.180.04 | 0.650.00 | 43331736 | 0.600.25 | |
ELCS-II | 0.500.01 | 0.260.00 | 3.180.04 | 0.650.00 | 39661423 | 0.590.25 | |
Child10 | MMHC | 0.220.01 | 0.190.02 | 4.370.11 | 0.480.01 | 7.7e62.0e6 | 1.5e33.9e2 |
NOTEARS | 0.490.01 | 0.340.02 | 2.990.10 | 0.650.02 | - | 3.3e31.9e2 | |
DAG-GNN | - | - | - | - | - | - | |
PCD-by-PCD | 0.550.02 | 0.360.03 | 1.690.06 | 0.380.03 | 156983819 | 2.820.71 | |
CMB | 0.710.04 | 0.590.03 | 1.580.12 | 0.350.03 | 269863942 | 3.710.54 | |
ELCS | 0.670.03 | 0.550.02 | 1.560.07 | 0.360.02 | 5074658 | 0.670.09 | |
ELCS-II | 0.670.03 | 0.550.02 | 1.560.07 | 0.360.02 | 4889631 | 0.660.09 | |
Pigs | MMHC | 0.260.00 | 0.000.00 | 6.720.02 | 1.000.00 | 4.6e59.9e3 | 90.352.90 |
NOTEARS | 0.420.00 | 0.220.01 | 2.850.03 | 0.800.01 | - | 2.4e41.7e3 | |
DAG-GNN | - | - | - | - | - | - | |
PCD-by-PCD | - | - | - | - | - | - | |
CMB | - | - | - | - | - | - | |
ELCS | 0.910.01 | 0.990.00 | 0.400.03 | 0.150.01 | 117933279 | 0.840.13 | |
ELCS-II | 0.910.01 | 0.990.00 | 0.400.03 | 0.150.01 | 116853272 | 0.840.13 | |
Gene | MMHC | - | - | - | - | - | - |
NOTEARS | - | - | - | - | - | - | |
DAG-GNN | - | - | - | - | - | - | |
PCD-by-PCD | - | - | - | - | - | - | |
CMB | - | - | - | - | - | - | |
ELCS | 0.770.00 | 0.780.01 | 0.780.02 | 0.310.01 | 317533432 | 4.370.47 | |
ELCS-II | 0.770.00 | 0.780.01 | 0.780.02 | 0.310.01 | 314303406 | 4.360.47 |
ELCS versus MMHC. Regardless of the number of samples (5000 or 1000), ELCS is significantly better than MMHC. On the ArrP and ArrR metrics, ELCS is superior to MMHC, which means that ELCS finds more true causal edges and less false casual edges. In addition, on the SHD metric, the value of SHD of ELCS is significantly lower than that of MMHC. On the FDR metric, ELCS performs better than MMHC. Furthermore, ELCS always uses less CI tests than MMHC. To learn the local causal structure of a target variable, MMHC needs to learn the whole DAG over all variables in a dataset, hence MMHC performs much more CI tests than ELCS. Thus, we can conclude that ELCS is more efficient and effective than MMHC.
ELCS versus NOTEARS and DAG-GNN. NOTEARS and DAG-GNN are global causal learning algorithms, they need to learn the global structure over all variables, and then obtain the parents and children of a given variable. ELCS achieves better performance than NOTEARS and DAG-GNN using both 5,000 data samples and 1,000 data samples, especially using 5,000 data samples. On the ArrP, ArrR, SHD and FDR metrics, ELCS is significantly better than NOTEARS and DAG-GNN. The values of ELCS on ArrP and ArrR metrics are higher than that of NOTEARS and DAG-GNN, and lower on the SHD and FDR metrics. Since NOTEARS and DAG-GNN adopt a continuous optimization strategy to obtain the DAG from observational data, and the experimental results are susceptible to the influence of parameters. Additionally, NOTEARS and DAG-GNN need to spend much time in learning the DAG, since they obtain the optimal solution by means of a large number of iterations. In a word, ELCS is superior to NOTEARS and DAG-GNN.
ELCS versus PCD-by-PCD and CMB. Both PCD-by-PCD and CMB are local causal learning algorithms. Using 5,000 data samples, ELCS performs better than PCD-by-PCD and CMB. Except on Child, ELCS achieves highest ArrP and ArrR values, and lowest SHD and FDR values on the other BNs. In addition, ELCS uses less CI tests than PCD-by-PCD and CMB on most of BNs. Using 1,000 data samples, ELCS is superior to PCD-by-PCD and CMB on Alarm, Child, Alarm10, Pigs and Gene. ELCS is better than CMB and worse than PCD-by-PCD on Insurance on the ArrP, ArrR, SHD and FDR metrics, while ELCS has advantages in terms of CI tests and running time. On Insurance10, ELCS is worse than PCD-by-PCD and CMB on the ArrP, ArrR, SHD and FDR metrics. The reason may be that EMB learns inaccurate MBs on small data samples. ELCS is better than PCD-by-PCD and little worse than CMB on Child10 on the ArrP, ArrR, SHD and FDR metrics. Generally, ELCS performs better than PCD-by-PCD and CMB.
ELCS versus ELCS-II. ELCS-II is superior to ELCS. ELCS-II further improves the efficiency of EMB while maintaining the same performance as measured by the ArrP, ArrR, SHD and FDR metrics, which indicates the efficiency of ELCS-II.
In summary, it can be seen from Tables V-VI, ELCS is significantly better than MMHC, NOTEARS and DAG-GNN. Additionally, ELCS outperforms PCD-by-PCD and CMB on the ArrP, ArrR, SHD, FDR metrics. Specifically, compared with PCD-by-PCD and CMB, ELCS not only achieves higher ArrP and ArrR values, but also achieves lower SHD and FDR values. Furthermore, ELCS is the fastest algorithm among all structure learning algorithms. ELCS is significantly better than MMHC and NOTEARS in terms of running time. MMHC, NOTEARS and DAG-GNN are global causal learning algorithms, they need to find the global structure of a BN. In particular, ELCS is 10 times faster than MMHC and 1000 times faster than NOTEARS and DAG-GNN on average. Additionally, ELCS is also superior to PCD-by-PCD and CMB in terms of running times. ELCS is 2 times faster than PCD-by-PCD and 3 times faster than CMB on average. Specifically, MMHC, NOTEARS, PCD-by-PCD and CMB fail to generate the output on several BNs, whereas ELCS can be successful applied in learning the local causal structure of each variable within two days. But beyond that, ELCS uses the smallest number of CI tests. Overall, ELCS is superior to its rivals in both efficiency and accuracy.
Network | Algorithm | ArrP() | ArrR() | SHD() | FDR() | CI Tests() | Time() |
---|---|---|---|---|---|---|---|
Alarm | ELCS | 0.860.01 | 0.810.01 | 0.440.06 | 0.070.02 | 64855 | 0.200.02 |
ELCS w/o N | 0.870.01 | 0.810.01 | 0.400.06 | 0.050.02 | 70199 | 0.220.03 | |
Insurance | ELCS | 0.850.04 | 0.690.04 | 1.610.06 | 0.180.05 | 1686276 | 0.750.12 |
ELCS w/o N | 0.850.04 | 0.680.03 | 1.610.14 | 0.180.05 | 1924257 | 0.840.11 | |
Child | ELCS | 0.710.12 | 0.610.16 | 1.080.36 | 0.090.08 | 2093287 | 0.930.10 |
ELCS w/o N | 0.700.13 | 0.590.16 | 1.130.36 | 0.090.09 | 2143277 | 0.940.09 | |
Alarm10 | ELCS | 0.830.01 | 0.680.02 | 1.260.07 | 0.140.02 | 6893483 | 1.770.12 |
ELCS w/o N | 0.840.01 | 0.680.02 | 1.240.06 | 0.140.02 | 7579524 | 2.120.14 | |
Insurance10 | ELCS | 0.800.02 | 0.670.02 | 1.750.11 | 0.230.01 | 108091528 | 3.920.55 |
ELCS w/o N | 0.860.02 | 0.730.02 | 1.420.10 | 0.170.01 | 113001713 | 4.490.55 | |
Child10 | ELCS | 0.830.05 | 0.760.07 | 0.730.20 | 0.140.03 | 131292613 | 3.980.79 |
ELCS w/o N | 0.830.06 | 0.760.07 | 0.740.21 | 0.130.03 | 134912559 | 4.320.77 | |
Pigs | ELCS | 0.910.00 | 0.990.00 | 0.420.02 | 0.150.01 | 133748660 | 8.916.84 |
ELCS w/o N | 0.910.00 | 0.990.00 | 0.400.02 | 0.150.01 | 143438657 | 9.706.75 | |
Gene | ELCS | 0.760.01 | 0.790.01 | 0.790.03 | 0.320.01 | 369507876 | 11.032.35 |
ELCS w/o N | 0.760.01 | 0.780.01 | 0.790.03 | 0.320.01 | 380618041 | 12.522.62 |
5.3 Why ELCS is Efficient and Effective?
In this section, we analyse the reason why ELCS is efficient and effective from the following two aspects. First, we give a case study to evaluate the effectiveness of utilizing the N-structures to infer edge directions between a given variable and its children. Second, we evaluate the effectiveness of the proposed EMB subroutine, since the proposed ELCS algorithm relies on EMB.
5.3.1 Case Study
To illustrate the benefit of utilizing the N-structures, we do not use the N-structures to distinguish the children of a given variable in learning the MB of the variable, that is, in the DistinguishPC subroutine, we remove lines 8-12 in Algorithm 4, and we denote this version of ELCS as “ECLS w/o N”. Table VII summarizes the experimental results of ECLS and “ECLS w/o N” on the eight BNs with 5,000 data examples. From the experimental results, we note that ELCS achieves comparable performance against “ECLS w/o N” in terms of ArrP, ArrR, SHD and FDR on average but ELCS performs less CI tests. Specifically, on Alarm, for each variable, the number of CI tests is reduced by 53 on average. On Gene, for each variable, the number of CI tests is reduced by 1111 on average. We observe that there are significant differences between the number of CI tests of ELCS and “ECLS w/o N” on Insurance, Alarm10, Insurance10, Pigs and Gene, but there is a little difference between them on the other BNs. The reason may be that ELCS makes use of the N-structures to speed up children identification, and there are few N-structures existing in Alarm, Child and Child10, leading to that the number of CI tests of ELCS and “ECLS w/o N” on these three BNs are close to each other. The efficiency performance of ECLS will improve a little on the BNs with few N-structures, this is a limitation of ELCS. In summary, ELCS is more efficient and provides better local structure learning quality than “ECLS w/o N”, which indicates the effectiveness of leveraging the N-structures for local casual structure learning.
Network | Algorithm | Distance() | F1() | Precision() | Recall() | CI Tests() | Time() |
---|---|---|---|---|---|---|---|
Alarm | IAMB | 0.150.03 | 0.900.02 | 0.940.02 | 0.890.01 | 1422 | 0.050.00 |
MMMB | 0.100.02 | 0.940.02 | 0.920.02 | 0.970.01 | 60426 | 0.240.01 | |
HITON-MB | 0.060.02 | 0.960.01 | 0.970.02 | 0.970.01 | 39412 | 0.130.00 | |
STMB | 0.300.02 | 0.780.02 | 0.730.02 | 0.960.01 | 53115 | 0.190.01 | |
BAMB | 0.090.03 | 0.940.02 | 0.960.03 | 0.950.01 | 35111 | 0.140.00 | |
EMB | 0.060.01 | 0.960.01 | 0.990.01 | 0.950.01 | 3187 | 0.100.00 | |
EMB-II | 0.060.01 | 0.960.01 | 0.990.01 | 0.950.01 | 2985 | 0.090.00 | |
Insurance | IAMB | 0.360.02 | 0.760.01 | 0.940.02 | 0.670.01 | 1042 | 0.040.00 |
MMMB | 0.310.02 | 0.790.02 | 0.880.03 | 0.760.02 | 1186124 | 0.600.07 | |
HITON-MB | 0.330.03 | 0.780.02 | 0.880.03 | 0.740.02 | 67962 | 0.310.04 | |
STMB | 0.490.03 | 0.650.02 | 0.640.04 | 0.770.03 | 70347 | 0.330.02 | |
BAMB | 0.300.02 | 0.800.01 | 0.890.03 | 0.770.02 | 61939 | 0.330.02 | |
EMB | 0.310.01 | 0.790.01 | 0.920.02 | 0.730.01 | 37023 | 0.160.01 | |
EMB-II | 0.310.01 | 0.790.01 | 0.920.02 | 0.730.01 | 36020 | 0.150.01 | |
Child | IAMB | 0.150.02 | 0.900.02 | 0.950.03 | 0.880.01 | 631 | 0.030.00 |
MMMB | 0.050.02 | 0.970.01 | 0.960.02 | 0.990.01 | 89725 | 0.470.01 | |
HITON-MB | 0.040.03 | 0.980.02 | 0.970.03 | 0.990.01 | 49916 | 0.240.01 | |
STMB | 0.170.04 | 0.890.03 | 0.840.04 | 0.980.02 | 37435 | 0.170.02 | |
BAMB | 0.090.03 | 0.950.02 | 0.930.02 | 0.980.02 | 37611 | 0.190.01 | |
EMB | 0.050.02 | 0.970.02 | 0.970.02 | 0.980.02 | 2058 | 0.090.00 | |
EMB-II | 0.050.02 | 0.970.02 | 0.970.02 | 0.980.02 | 2048 | 0.090.00 | |
Alarm10 | IAMB | 0.360.01 | 0.750.01 | 0.830.01 | 0.740.00 | 163714 | 0.590.01 |
MMMB | 0.260.01 | 0.820.01 | 0.880.01 | 0.810.00 | 192645 | 0.600.01 | |
HITON-MB | 0.250.01 | 0.840.01 | 0.900.01 | 0.820.00 | 171411 | 0.440.00 | |
STMB | 0.670.01 | 0.480.01 | 0.410.01 | 0.830.01 | 504939 | 1.890.02 | |
BAMB | 0.300.01 | 0.800.00 | 0.830.01 | 0.820.00 | 180212 | 0.570.01 | |
EMB | 0.250.01 | 0.830.01 | 0.910.01 | 0.810.00 | 19247 | 0.500.02 | |
EMB-II | 0.250.01 | 0.830.01 | 0.910.01 | 0.810.00 | 19087 | 0.490.02 | |
Insurance10 | IAMB | 0.420.01 | 0.710.01 | 0.890.01 | 0.660.00 | 12108 | 0.500.01 |
MMMB | 0.330.01 | 0.780.01 | 0.820.01 | 0.800.00 | 327445 | 1.530.03 | |
HITON-MB | 0.320.01 | 0.780.01 | 0.840.01 | 0.800.00 | 234818 | 0.930.01 | |
STMB | 0.770.01 | 0.400.01 | 0.300.01 | 0.790.00 | 6781118 | 3.360.07 | |
BAMB | 0.340.01 | 0.770.00 | 0.800.01 | 0.800.00 | 254122 | 1.170.01 | |
EMB | 0.280.01 | 0.810.00 | 0.910.01 | 0.780.00 | 218915 | 0.780.01 | |
EMB-II | 0.280.01 | 0.810.00 | 0.910.01 | 0.780.00 | 212214 | 0.750.01 | |
Child10 | IAMB | 0.240.01 | 0.840.01 | 0.870.01 | 0.880.00 | 75010 | 0.310.00 |
MMMB | 0.100.01 | 0.940.01 | 0.910.01 | 0.990.00 | 162221 | 0.700.01 | |
HITON-MB | 0.080.01 | 0.950.01 | 0.920.01 | 0.990.00 | 119411 | 0.430.01 | |
STMB | 0.560.02 | 0.560.02 | 0.450.02 | 0.990.00 | 288124 | 1.260.01 | |
BAMB | 0.240.01 | 0.840.00 | 0.760.01 | 0.990.00 | 111110 | 0.460.01 | |
EMB | 0.070.01 | 0.950.01 | 0.940.01 | 0.990.00 | 10628 | 0.330.00 | |
EMB-II | 0.070.01 | 0.950.01 | 0.940.01 | 0.990.00 | 10618 | 0.330.00 | |
Pigs | IAMB | 0.420.00 | 0.710.00 | 0.620.00 | 0.960.00 | 26167 | 1.280.00 |
MMMB | 0.130.00 | 0.920.00 | 0.870.00 | 1.000.00 | 3.2e52.2e4 | 2.2e21.7e1 | |
HITON-MB | 0.140.00 | 0.920.00 | 0.860.00 | 1.000.00 | 46956454 | 34.940.33 | |
STMB | 0.820.01 | 0.260.00 | 0.180.01 | 1.000.00 | 457702746 | 29.202.05 | |
BAMB | 0.180.01 | 0.890.01 | 0.820.01 | 1.000.00 | 29097201 | 31.160.13 | |
EMB | 0.120.00 | 0.930.00 | 0.880.00 | 1.000.00 | 87843405 | 5.732.43 | |
EMB-II | 0.120.00 | 0.930.00 | 0.880.00 | 1.000.00 | 7335209 | 4.940.26 | |
Gene | IAMB | 0.320.00 | 0.790.00 | 0.760.00 | 0.890.00 | 346310 | 1.360.00 |
MMMB | 0.250.00 | 0.830.00 | 0.770.00 | 0.940.00 | 603548 | 2.210.02 | |
HITON-MB | 0.250.00 | 0.830.00 | 0.770.00 | 0.940.00 | 457622 | 1.440.00 | |
STMB | 0.880.00 | 0.180.00 | 0.130.00 | 1.000.00 | 17282268 | 7.700.18 | |
BAMB | 0.390.00 | 0.730.00 | 0.640.00 | 0.940.00 | 447430 | 1.720.02 | |
EMB | 0.260.00 | 0.820.00 | 0.760.00 | 0.940.00 | 44869 | 1.280.00 | |
EMB-II | 0.260.00 | 0.820.00 | 0.760.00 | 0.940.00 | 441211 | 1.270.00 |
Network | Algorithm | Distance() | F1() | Precision() | Recall() | CI Tests() | Time() |
---|---|---|---|---|---|---|---|
Alarm | IAMB | 0.270.01 | 0.810.01 | 0.930.01 | 0.760.01 | 1202 | 0.020.00 |
MMMB | 0.200.01 | 0.870.01 | 0.910.02 | 0.870.01 | 43733 | 0.090.00 | |
HITON-MB | 0.160.01 | 0.900.01 | 0.950.02 | 0.870.01 | 31512 | 0.050.00 | |
STMB | 0.390.03 | 0.720.02 | 0.710.02 | 0.850.02 | 39212 | 0.060.00 | |
BAMB | 0.210.01 | 0.860.01 | 0.910.02 | 0.860.01 | 28016 | 0.040.00 | |
EMB | 0.180.02 | 0.880.01 | 0.960.02 | 0.850.02 | 29714 | 0.040.00 | |
EMB-II | 0.180.02 | 0.880.01 | 0.960.02 | 0.850.02 | 28513 | 0.040.00 | |
Insurance | IAMB | 0.480.02 | 0.660.01 | 0.920.03 | 0.560.01 | 862 | 0.010.00 |
MMMB | 0.420.03 | 0.710.02 | 0.830.03 | 0.660.02 | 51147 | 0.120.01 | |
HITON-MB | 0.450.03 | 0.690.02 | 0.830.04 | 0.650.01 | 35834 | 0.060.01 | |
STMB | 0.590.03 | 0.580.03 | 0.580.06 | 0.660.04 | 11381277 | 0.150.16 | |
BAMB | 0.450.02 | 0.690.02 | 0.760.04 | 0.680.01 | 40451 | 0.060.01 | |
EMB | 0.460.03 | 0.680.03 | 0.810.08 | 0.640.02 | 39582 | 0.060.01 | |
EMB-II | 0.460.03 | 0.680.03 | 0.810.08 | 0.640.02 | 37176 | 0.050.01 | |
Child | IAMB | 0.270.03 | 0.820.02 | 0.940.03 | 0.760.02 | 541 | 0.010.00 |
MMMB | 0.220.03 | 0.850.02 | 0.890.04 | 0.860.02 | 82385 | 0.130.01 | |
HITON-MB | 0.200.03 | 0.870.02 | 0.900.03 | 0.870.02 | 46953 | 0.060.01 | |
STMB | 0.230.07 | 0.850.05 | 0.860.05 | 0.870.04 | 2217 | 0.040.00 | |
BAMB | 0.230.04 | 0.850.03 | 0.840.04 | 0.910.01 | 44158 | 0.050.01 | |
EMB | 0.17+0.03 | 0.89+0.03 | 0.940.03 | 0.870.02 | 32052 | 0.040.01 | |
EMB-II | 0.17+0.03 | 0.89+0.03 | 0.940.03 | 0.870.02 | 28739 | 0.040.00 | |
Alarm10 | IAMB | 0.520.01 | 0.630.01 | 0.780.01 | 0.600.01 | 135511 | 0.180.00 |
MMMB | 0.390.01 | 0.730.01 | 0.840.01 | 0.700.00 | 157917 | 0.280.00 | |
HITON-MB | 0.370.01 | 0.750.01 | 0.860.01 | 0.700.01 | 147413 | 0.200.00 | |
STMB | 0.760.01 | 0.420.01 | 0.370.02 | 0.700.01 | 366829 | 0.760.00 | |
BAMB | 0.460.00 | 0.680.00 | 0.740.01 | 0.700.00 | 155116 | 0.230.00 | |
EMB | 0.380.01 | 0.740.00 | 0.880.01 | 0.690.00 | 176511 | 0.230.00 | |
EMB-II | 0.380.01 | 0.740.00 | 0.880.01 | 0.690.00 | 175612 | 0.230.00 | |
Insurance10 | IAMB | 0.550.01 | 0.610.01 | 0.850.01 | 0.530.00 | 9635 | 0.130.00 |
MMMB | 0.490.01 | 0.660.01 | 0.700.01 | 0.700.01 | 218048 | 0.400.01 | |
HITON-MB | 0.450.01 | 0.680.01 | 0.740.01 | 0.700.01 | 169828 | 0.250.00 | |
STMB | 0.750.01 | 0.460.01 | 0.330.01 | 0.740.01 | 144314 | 0.210.01 | |
BAMB | 0.530.01 | 0.620.01 | 0.600.01 | 0.740.01 | 224359 | 0.340.01 | |
EMB | 0.650.06 | 0.530.05 | 0.470.07 | 0.710.01 | 43331736 | 0.570.23 | |
EMB-II | 0.650.06 | 0.530.05 | 0.470.07 | 0.710.01 | 39661423 | 0.550.23 | |
Child10 | IAMB | 0.400.01 | 0.720.01 | 0.840.01 | 0.710.01 | 6148 | 0.080.00 |
MMMB | 0.280.02 | 0.810.01 | 0.820.02 | 0.860.01 | 175743 | 0.270.00 | |
HITON-MB | 0.250.02 | 0.830.01 | 0.840.02 | 0.870.01 | 127224 | 0.170.00 | |
STMB | 0.660.02 | 0.480.02 | 0.390.02 | 0.850.01 | 218641 | 0.390.00 | |
BAMB | 0.470.01 | 0.670.01 | 0.580.01 | 0.900.01 | 146043 | 0.200.01 | |
EMB | 0.220.02 | 0.850.01 | 0.870.02 | 0.880.01 | 122510 | 0.160.00 | |
EMB-II | 0.220.02 | 0.850.01 | 0.870.02 | 0.880.01 | 11828 | 0.160.00 | |
Pigs | IAMB | 0.340.00 | 0.790.00 | 0.820.00 | 0.840.00 | 17551 | 0.220.00 |
MMMB | 0.150.01 | 0.910.01 | 0.850.01 | 1.000.00 | 19788417879 | 8.440.71 | |
HITON-MB | 0.120.01 | 0.920.01 | 0.880.01 | 1.000.00 | 470281667 | 4.780.18 | |
STMB | 0.850.00 | 0.250.00 | 0.150.00 | 1.000.00 | 256263234 | 2.240.09 | |
BAMB | 0.310.01 | 0.800.01 | 0.690.01 | 1.000.00 | 404664419 | 11.221.55 | |
EMB | 0.110.01 | 0.930.00 | 0.890.01 | 1.000.00 | 754199 | 0.580.01 | |
EMB-II | 0.110.01 | 0.930.00 | 0.890.01 | 1.000.00 | 7466193 | 0.560.01 | |
Gene | IAMB | 0.390.00 | 0.730.00 | 0.790.00 | 0.790.00 | 288710 | 0.360.00 |
MMMB | 0.280.00 | 0.810.00 | 0.750.00 | 0.930.00 | 456936 | 0.750.01 | |
HITON-MB | 0.250.01 | 0.830.00 | 0.790.00 | 0.930.00 | 391826 | 0.540.01 | |
STMB | 0.860.00 | 0.210.00 | 0.140.00 | 0.990.00 | 1067274 | 2.500.01 | |
BAMB | 0.460.01 | 0.670.00 | 0.570.01 | 0.940.00 | 381752 | 0.610.01 | |
EMB | 0.250.00 | 0.830.00 | 0.780.00 | 0.930.00 | 422818 | 0.570.00 | |
EMB-II | 0.250.00 | 0.830.00 | 0.780.00 | 0.930.00 | 420417 | 0.570.00 |
5.3.2 Experimental Results of EMB and Its Rivals
We evaluate the effectiveness of the proposed EMB by comparing it with five state-of-the-art MB learning algorithms, including BAMB [17], IAMB [21], MMMB [24], HITON-MB [25] and STMB [27].
For MB learning algorithms, we use precision, recall, F1, distance [26] [27], CI tests, and running time (in seconds) as the evaluation metrics.
-
•
Precision: The number of true positives in the output (i.e., the variables in the output belonging to the true MB of a target variable in a test DAG) divided by the number of variables in the output of an algorithm.
-
•
Recall: The number of true positives in the output divided by the number of true positives (the number of the true MB of a target variable in a test DAG).
-
•
F1 = 2 * Precision * Recall / (Precision + Recall): The F1 score is the harmonic average of the precision and recall, where F1 = 1 is the best case (perfect precision and recall) while F1 = 0 is the worst case.
- •
-
•
Efficiency: The number of CI tests and the running time (in seconds) are used to measure the efficiency.
Tables VIII-IX report the experimental results of EMB and its rivals. From the experimental results, we have the following observations.
EMB versus IAMB, MMMB and HITON-MB. IAMB is much faster than EMB, while IAMB is significantly worse than EMB in terms of distance, F1, precision and recall on average. Compared with MMMB and HITON-MB, EMB is more efficient. EMB needs much less CI tests than MMMB and HITON-MB. In addition, using 5,000 data samples, EMB is 2 times faster than MMMB and 1.2 times faster than HITON-MB on average. Moreover, EMB is more accurate than MMMB. In particular, using 5,000 data samples, EMB achieves the lowest distance and the highest F1 values on Alarm, Insurance10, Child10 and Pigs. Using 1,000 data samples, EMB obtains the lowest distance and the highest F1 values on Child, Child10, Pigs and Gene. Overall, EMB is superior to IAMB, MMMB and HITON-MB.
EMB versus STMB and BAMB. From Tables VIII-IX, we note that STMB achieves higher recall values than EMB, but on the distance, F1 and precision metrics, STMB is significantly worse than EMB. Compared with BAMB, EMB achieves lower distance and higher F1 values. Additionally, the number of CI tests of EMB is less than STMB and BAMB. More specifically, using 5,000 data samples, EMB is 3.5 times faster than STMB and 1.5 times faster than BAMB on average. In a word, EMB performs better than BAMB and STMB in both efficiency and accuracy.
EMB versus EMB-II. EMB is inferior to EMB-II. Compared with EMB, EMB-II uses less CI tests for MB learning while achieving the same performance as measured by the distance, F1, precision and recall matrics, which indicates the efficiency of EMB-II.
EMB is able to effectively find the MB of a target variable, and simultaneously distinguish parents and children of the target variable. We note that EMB uses less CI tests for MB learning, which can reduce the impact of unreliable CI tests. In summary, EMB is helpful to learn the local causal structure.
To further demonstrate the effectiveness of EMB, we propose three variants of ELCS, which are referred to as ELCS-M, ECLS-S, ELCS-B, respectively. ELCS-M uses MMMB to replace EMB in ELCS. ECLS-S and ELCS-B use STMB and BAMB to replace EMB in ELCS, respectively. Table X reports the experimental results of ECLS and its three variants on the eight BNs with 5,000 data examples. From the table, we observe that ELCS outperforms these three rivals in terms of both CI tests and running time, which implies the efficiency of ELCS. We also note that ELCS achieves better ArrP, ArrR, SHD and FDR values than that of these three rivals, which shows the effectiveness of ELCS.
EMB has achieved encouraging performance, but it still suffers from the following two drawbacks. First, to improve the efficiency of EMB while maintaining competitive performance, EMB chooses to remove non-spouses within U{T}PC of the target variable T as early as possible at lines 2-16 in Algorithm 3. The size of the conditioning sets Temp (line 9 in Algorithm 3) and {Y} Sep{X} (line 11 in Algorithm 3) may be large, when the size of data samples is finite, the results of CI tests may be unreliable, leading to poor performance of EMB. Second, the performance of EMB is limited by HITON-PC that is used for PC learning. If HITON-PC has a lower quality of PC learning, inaccurate MBs will be learnt by EMB.
Network | Algorithm | ArrP() | ArrR() | SHD() | FDR() | CI Tests() | Time() |
---|---|---|---|---|---|---|---|
Alarm | ELCS-M | 0.710.02 | 0.590.03 | 1.140.06 | 0.260.03 | 1627186 | 0.620.08 |
ELCS-S | 0.790.03 | 0.720.05 | 0.880.14 | 0.130.04 | 140278 | 0.570.03 | |
ELCS-B | 0.780.03 | 0.660.05 | 0.920.09 | 0.260.03 | 77846 | 0.300.02 | |
ELCS | 0.860.01 | 0.810.01 | 0.440.06 | 0.070.02 | 64855 | 0.200.02 | |
ELCS-II | 0.860.01 | 0.810.01 | 0.440.06 | 0.070.02 | 60752 | 0.190.02 | |
Insurance | ELCS-M | 0.760.03 | 0.590.03 | 1.870.15 | 0.310.04 | 69761183 | 3.280.62 |
ELCS-S | 0.670.05 | 0.500.05 | 2.500.25 | 0.370.04 | 2653476 | 1.390.25 | |
ELCS-B | 0.670.02 | 0.440.02 | 2.260.10 | 0.410.01 | 3182447 | 1.780.27 | |
ELCS | 0.850.04 | 0.690.04 | 1.610.06 | 0.180.05 | 1686276 | 0.750.12 | |
ELCS-II | 0.850.04 | 0.690.04 | 1.610.06 | 0.180.05 | 1637275 | 0.750.12 | |
Child | ELCS-M | 0.680.11 | 0.560.12 | 1.180.31 | 0.060.04 | 88971247 | 4.430.62 |
ELCS-S | 0.810.07 | 0.720.09 | 0.750.18 | 0.180.07 | 2451516 | 1.280.28 | |
ELCS-B | 0.750.03 | 0.650.05 | 0.930.14 | 0.260.05 | 2252195 | 1.160.10 | |
ELCS | 0.710.12 | 0.610.16 | 1.080.36 | 0.090.08 | 2093287 | 0.930.10 | |
ELCS-II | 0.710.12 | 0.610.16 | 1.080.36 | 0.090.08 | 2087287 | 0.930.10 | |
Alarm10 | ELCS-M | 0.770.01 | 0.590.02 | 1.600.07 | 0.190.02 | 105701207 | 3.170.39 |
ELCS-S | 0.690.01 | 0.460.01 | 1.650.03 | 0.320.01 | 8429237 | 3.660.15 | |
ELCS-B | 0.680.01 | 0.440.00 | 1.850.02 | 0.460.01 | 7223170 | 2.090.03 | |
ELCS | 0.830.01 | 0.680.02 | 1.260.07 | 0.140.02 | 6893483 | 1.770.12 | |
ELCS-II | 0.830.01 | 0.680.02 | 1.260.07 | 0.140.02 | 6916480 | 1.770.12 | |
Insurance10 | ELCS-M | 0.750.01 | 0.610.01 | 1.850.04 | 0.330.01 | 144612927 | 6.351.27 |
ELCS-S | 0.600.01 | 0.380.01 | 2.440.02 | 0.510.01 | 10338533 | 5.780.55 | |
ELCS-B | 0.640.01 | 0.420.01 | 2.310.06 | 0.470.01 | 7556224 | 3.550.11 | |
ELCS | 0.800.02 | 0.670.02 | 1.750.11 | 0.230.01 | 108091528 | 3.920.55 | |
ELCS-II | 0.800.02 | 0.670.02 | 1.750.11 | 0.230.01 | 106051499 | 3.910.55 | |
Child10 | ELCS-M | 0.800.02 | 0.750.03 | 0.750.08 | 0.160.02 | 134381388 | 5.800.58 |
ELCS-S | 0.660.01 | 0.480.02 | 1.170.03 | 0.450.02 | 152492335 | 6.860.70 | |
ELCS-B | 0.700.01 | 0.520.02 | 1.050.05 | 0.470.02 | 138801187 | 4.280.05 | |
ELCS | 0.830.05 | 0.760.07 | 0.730.20 | 0.140.03 | 131292613 | 3.980.79 | |
ELCS-II | 0.830.05 | 0.760.07 | 0.730.20 | 0.140.03 | 131042605 | 3.980.79 | |
Pigs | ELCS-M | - | - | - | - | - | - |
ELCS-S | - | - | - | - | - | - | |
ELCS-B | - | - | - | - | - | - | |
ELCS | 0.910.00 | 0.990.00 | 0.420.02 | 0.150.01 | 133748660 | 8.916.84 | |
ELCS-II | 0.910.00 | 0.990.00 | 0.420.02 | 0.150.01 | 114675659 | 8.385.12 | |
Gene | ELCS-M | - | - | - | - | - | - |
ELCS-S | - | - | - | - | - | - | |
ELCS-B | - | - | - | - | - | - | |
ELCS | 0.760.01 | 0.790.01 | 0.790.03 | 0.320.01 | 369507876 | 11.032.35 | |
ELCS-II | 0.760.01 | 0.790.01 | 0.790.03 | 0.320.01 | 360517696 | 11.022.08 |
6 Conclusion
A new local causal structure learning algorithm (ELCS) has been proposed in this paper, which reduces the search space in distinguishing parents from children of a target variable of interest. Specifically, ELCS makes use of the N-structures to distinguish parents from children of the target variable during learning the MB of the target variable. Furthermore, to combine MB learning with the N-structures to infer edge directions between the target variable and its PC, we design an effective MB discovery subroutine (EMB). We theoretically analyze the correctness of ELCS. Extensive experimental results on benchmark BNs indicate that ELCS not only improves the efficiency for learning the local causal structure, but also achieves better performance in accuracy. In future, we plan to extend the ELCS algorithm for global causal structures learning and robust machine learning.
Acknowledgments
This work was supported in part by the National Key Research and Development Program of China (under Grant 2020AAA0106100), the National Natural Science Foundation of China (under Grant 61876206), and Open Project Foundation of Intelligent Information Processing Key Laboratory of Shanxi Province (under Grant CICIP2020003).
References
- [1] K. Yu, L. Liu, and J. Li, “A unified view of causal and non-causal feature selection,” ACM Transactions on Knowledge Discovery from Data (TKDD), in press, 2020.
- [2] R. Cai, J. Qiao, K. Zhang, Z. Zhang, and Z. Hao, “Causal discovery with cascade nonlinear additive noise model,” in Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI, Macao, China, 10-16 August, 2019, pp. 1609–1615.
- [3] B. Huang, K. Zhang, M. Gong, and C. Glymour, “Causal discovery and forecasting in nonstationary environments with state-space models,” in Proceedings of the 36th International Conference on Machine Learning, ICML, 9-15 June, Long Beach, California, USA, 2019, pp. 2901–2910.
- [4] A. Marx and J. Vreeken, “Identifiability of cause and effect using regularized regression,” in Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2019, Anchorage, AK, USA, 4-8 August, 2019, pp. 852–861.
- [5] K. Yu, L. Liu, J. Li, W. Ding, and T. D. Le, “Multi-source causal feature selection,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 42, no. 9, pp. 2240–2256, 2020.
- [6] B. Gourévitch, R. L. Bouquin-Jeannès, and G. Faucon, “Linear and nonlinear causality between signals: methods, examples and neurophysiological applications,” Biol. Cybern., vol. 95, no. 4, pp. 349–369, 2006.
- [7] J. Choi, R. S. Chapkin, and Y. Ni, “Bayesian causal structural learning with zero-inflated poisson bayesian networks,” in Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems, December 6-12, virtual, 2020.
- [8] M. H. Maathuis, M. Kalisch, P. Bühlmann et al., “Estimating high-dimensional intervention effects from observational data,” The Annals of Statistics, vol. 37, no. 6A, pp. 3133–3164, 2009.
- [9] J. Pearl, Causality. Cambridge university press, 2009.
- [10] A. Marx and J. Vreeken, “Testing conditional independence on discrete data using stochastic complexity,” in The 22nd International Conference on Artificial Intelligence and Statistics, AISTATS, 16-18 April, Naha, Okinawa, Japan, 2019, pp. 496–505.
- [11] I. Tsamardinos, L. E. Brown, and C. F. Aliferis, “The max-min hill-climbing bayesian network structure learning algorithm,” Machine Learning, vol. 65, no. 1, pp. 31–78, 2006.
- [12] X. Zheng, B. Aragam, P. Ravikumar, and E. P. Xing, “Dags with NO TEARS: continuous optimization for structure learning,” in Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, 3-8 December, Montréal, Canada., 2018, pp. 9492–9503.
- [13] Y. Yu, J. Chen, T. Gao, and M. Yu, “DAG-GNN: DAG structure learning with graph neural networks,” in Proceedings of the 36th International Conference on Machine Learning, ICML, 9-15 June, Long Beach, California, USA, 2019, pp. 7154–7163.
- [14] J. Yin, Y. Zhou, C. Wang, P. He, C. Zheng, and Z. Geng, “Partial orientation and local structural learning of causal networks for prediction,” in Causation and Prediction Challenge at WCCI, Hong Kong, June 1-6, 2008, pp. 93–105.
- [15] T. Gao and Q. Ji, “Local causal discovery of direct causes and effects,” in Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems, 7-12 December, Montreal, Quebec, Canada, 2015, pp. 2512–2520.
- [16] X. Wu, B. Jiang, K. Yu, C. Miao, and H. Chen, “Accurate markov boundary discovery for causal feature selection,” IEEE Trans. Cybern., vol. 50, no. 12, pp. 4983–4996, 2020.
- [17] Z. Ling, K. Yu, H. Wang, L. Liu, W. Ding, and X. Wu, “BAMB: A balanced markov blanket discovery approach to feature selection,” ACM Transactions on Intelligent Systems and Technology, vol. 10, no. 5, pp. 52:1–52:25, 2019.
- [18] T. Niinimaki and P. Parviainen, “Local structure discovery in bayesian networks,” in Proceedings of the Twenty-Eighth Conference on Uncertainty in Artificial Intelligence, Catalina Island, CA, USA, 14-18 August, 2012, pp. 634–643.
- [19] T. Gao and Q. Ji, “Efficient score-based markov blanket discovery,” Int. J. Approx. Reason., vol. 80, pp. 277–293, 2017.
- [20] D. Margaritis and S. Thrun, “Bayesian network induction via local neighborhoods,” in Advances in Neural Information Processing Systems 12, [NIPS Conference, Denver, Colorado, USA, November 29 - December 4, 1999], S. A. Solla, T. K. Leen, and K. Müller, Eds., 1999, pp. 505–511.
- [21] I. Tsamardinos and C. F. Aliferis, “Towards principled feature selection: Relevancy, filters and wrappers,” in Proceedings of the Ninth International Workshop on Artificial Intelligence and Statistics, AISTATS 2003, Key West, Florida, USA, January 3-6, 2003, C. M. Bishop and B. J. Frey, Eds., 2003.
- [22] I. Tsamardinos, C. F. Aliferis, A. R. Statnikov, and E. Statnikov, “Algorithms for large scale markov blanket discovery.” in FLAIRS conference, vol. 2, 2003, pp. 376–380.
- [23] S. Yaramakala and D. Margaritis, “Speculative markov blanket discovery for optimal feature selection,” in Proceedings of the 5th IEEE International Conference on Data Mining (ICDM 2005), 27-30 November 2005, Houston, Texas, USA, 2005, pp. 809–812.
- [24] I. Tsamardinos, C. F. Aliferis, and A. R. Statnikov, “Time and sample efficient discovery of markov blankets and direct causal relations,” in Proceedings of the Ninth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 24-27 August, 2003, pp. 673–678.
- [25] C. F. Aliferis, I. Tsamardinos, and A. R. Statnikov, “HITON: A novel markov blanket algorithm for optimal variable selection,” in AMIA, American Medical Informatics Association Annual Symposium, Washington, DC, USA, November 8-12, 2003.
- [26] J. M. Peña, R. Nilsson, J. Björkegren, and J. Tegnér, “Towards scalable and data efficient learning of markov boundaries,” Int. J. Approx. Reason., vol. 45, no. 2, pp. 211–232, 2007.
- [27] T. Gao and Q. Ji, “Efficient markov blanket discovery and its application,” IEEE Trans. Cybernetics, vol. 47, no. 5, pp. 1169–1179, 2017.
- [28] K. Yu, X. Guo, L. Liu, J. Li, H. Wang, Z. Ling, and X. Wu, “Causality-based feature selection: Methods and evaluations,” ACM Computing Surveys (CSUR), vol. 53, no. 5, pp. 1–36, 2020.
- [29] D. Margaritis and S. Thrun, “Bayesian network induction via local neighborhoods,” in Advances in Neural Information Processing Systems 12, [NIPS Conference, Denver, Colorado, USA, November 29-December 4], 1999, pp. 505–511.
- [30] T. Gao, K. P. Fadnis, and M. Campbell, “Local-to-global bayesian network structure learning,” in Proceedings of the 34th International Conference on Machine Learning, ICML, Sydney, NSW, Australia, 6-11 August, 2017, pp. 1193–1202.
- [31] T. Gao and D. Wei, “Parallel bayesian network structure learning,” in Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, 10-15 July, 2018, 2018, pp. 1671–1680.
- [32] S. Zhu, I. Ng, and Z. Chen, “Causal discovery with reinforcement learning,” in 8th International Conference on Learning Representations, ICLR 2020, April 26-30 Addis Ababa, Ethiopia, 2020.
- [33] M. Zhang, S. Jiang, Z. Cui, R. Garnett, and Y. Chen, “D-VAE: A variational autoencoder for directed acyclic graphs,” in Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, 8-14 December 2019, Vancouver, BC, Canada, 2019, pp. 1586–1598.
- [34] J. Pearl, Probabilistic reasoning in intelligent systems: networks of plausible inference. Morgan Kaufmann series in representation and reasoning, 1988.
- [35] P. Spirtes, C. Glymour, and R. Scheines, Causation, Prediction, and Search. MIT Press, 2000.
- [36] C. Meek, “Causal inference and causal explanation with background knowledge,” Proc. Conf. on Uncertainty in Artificial Intelligence (UAI-95), pp. 403–410, 1995.
- [37] C. F. Aliferis, I. Tsamardinos, A. R. Statnikov, and L. E. Brown, “Causal explorer: A causal probabilistic network learning toolkit for biomedical discovery,” in Proceedings of the International Conference on Mathematics and Engineering Techniques in Medicine and Biological Scienes, METMBS, June 23-26, Las Vegas, Nevada, USA, 2003, pp. 371–376.
![]() |
Shuai Yang received his B.S. and M.S. degrees in computer science from the Hefei University of Technology, Hefei, China, in 2016 and 2019, respectively, where he is currently pursuing the Ph.D. degree with the School of Computer Science and Information Engineering. His main research interests include domain adaptation and causal discovery. |
![]() |
Hao Wang received the B.S. degree from the Department of Electrical Engineering and Automation, Shanghai Jiao Tong University, Shanghai, China, in 1984, and the M.S. and Ph.D. degrees in Computer Science from the Hefei University of Technology, Hefei, China, in 1989 and 1997, respectively. He is a Professor with the School of Computer Science and Information Engineering, Hefei University of Technology. His current research interests include artificial intelligence and robotics and knowledge engineering. |
![]() |
Kui Yu received the Ph.D. degree in computer science from the Hefei University of Technology, Hefei, China, in 2013. He is currently a Professor with the School of Computer Science and Information Engineering, Hefei University of Technology. From 2015 to 2018, he was a Research Fellow of Computer Science with STEM, University of South Australia, Adelaide SA, Australia. From 2013 to 2015, he was a Postdoctoral Fellow with the School of Computing Science, Simon Fraser University, Burnaby, BC, Canada. His main research interests include causal discovery and machine learning. |
![]() |
Fuyuan Cao received the M.S. and Ph.D. degrees in computer science from the Shanxi University, Taiyuan, China, in 2004 and 2010, respectively. He is currently a professor with the School of Computer and Information Technology, Shanxi University. His current research interests include machine learning and clustering analysis. |
![]() |
Xindong Wu (Fellow, IEEE) received the Ph.D. degree in artificial intelligence from the University of Edinburgh, Edinburgh, U.K., in 1993. He is currently a Chang Jiang Scholar with the School of Computer Science and Information Engineering, Hefei University of Technology, China, and also the Chief Scientist with the Mininglamp Academy of Sciences, Mininglamp Technology, Beijing, China. His research interests include data mining, knowledge-based systems, and web information exploration. He is a Fellow of the AAAS. He is also the Steering Committee Chair of the IEEE International Conference on Data Mining (ICDM), the Editor-in-Chief of Knowledge and Information Systems and of Springer book series, Advanced Information and Knowledge Processing. |