Discovering Cyclic Causal Mo dels by Indep endent Comp onents Analysis Gustavo Lacerda Machine Learning Department School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 Peter Spirtes Joseph Ramsey Department of Philosophy Carnegie Mellon University Pittsburgh, PA 15213 1.1 Patrik O. Hoyer Dept. of Computer Science University of Helsinki Helsinki, Finland Abstract We generalize Shimizu et al's (2006) ICA-based approach for discovering linear non-Gaussian acyclic (LiNGAM) Structural Equation Models (SEMs) from causally sufficient, continuous-valued observational data. By relaxing the assumption that the generating SEM's graph is acyclic, we solve the more general problem of linear non-Gaussian (LiNG) SEM discovery. LiNG discovery algorithms output the distribution equivalence class of SEMs which, in the large sample limit, represents the population distribution. We apply a LiNG discovery algorithm to simulated data. Finally, we give sufficient conditions under which only one of the SEMs in the output class is "stable". The mo del, with an illustration Let x be the random vector of substantive variables, e be the vector of error terms, and B be the matrix of linear coefficients for the substantive variables. Then the following equation describes the linear SEM model: x = Bx + e For example, consider the model defined by: x1 = e1 x2 = 1.2x1 - 0.3x4 + e2 x3 = 2x2 + e3 x4 = -x3 + e4 x5 = 3x2 + e5 Note that the coefficient of each variable on the lefthand-side of the equation is 1. (2) (1) 1 Linear SEMs Linear structural equation models (SEMs) are statistical causal models widely used in the natural and social sciences (including econometrics, political science, sociology, and biology) [1]. The variables in a linear SEM can be divided into two sets, the error terms (typically unobserved), and the substantive variables. For each substantive variable xi , there is a linear equation with xi on the left-handside, and the direct causes of xi plus the corresponding error term on the right-hand-side. Each SEM with jointly independent error terms can be associated with a directed graph (abbreviated as DG) that represents the causal structure of the model and the form of the linear equations. The vertices of the graph are the substantive variables, and there is a directed edge from xi to xj just when the coefficient of xi in the structural equation for xj is non-zero. 1 1 Fig. 1: Example 1 x can also be expressed directly as a linear combination of the error terms, as long as I - B is invertible. Solving for x in Eq. 1 gives x = (I - B )-1 e. If we let A = (I - B )-1 , then x = Ae. A is called the reduced form matrix (in the terminology of Independent cursive", and SEMs with cyclic graphs "non-recursive"[15]. We avoid this usage, and use "acyclic" or "cyclic" instead. Traditionally, SEMs with acyclic graphs are called "re- Components Analysis (see Section 3.1), it is called the "mixing matrix"). The distributions over the error terms in a SEM, together with the linear equations, entail a joint distribution over the substantive variables. This joint distribution can be interpreted in terms of physical processes, as shown next. 1.2 Interpretating linear SEMs that do not contain any "self-loops" (edges from a vertex to itself ) 2 , i.e. the B matrices output by our LiNG discovery algorithms have all zeros in the diagonal. This is because it is impossible to determine the values of the diagonal entries of the B matrix from equilibrium data alone. In the underlying dynamical equations, it may be that for some index a, xa [t + (k - 1)] affects xa [t + k ] (i.e. ba,a = 0). Our goal is to recover the coefficients ¯ that both represent the distribution of x and correctly predict the effects of manipulations. A manipulation of a variable xi to a distribution P is modeled by replacing the dynamical equation for xi by a new dynamical equation xi [t + k ] = e i , where e i has distribution P [10]. For these purposes, the following argument sketches why the underdetermination of the diagonal of Bequil by the equilibrium data is not a problem, as long as ba,a = 1 in the underlying dynamical equations. ¯ The equation for xa has the form: n These equations (contained in matrix equation (1)) can be given several different interpretations. Under one class of interpretations, they are a set of equations satisfied by a set of variables x in equilibrium. With some further assumptions, the B matrix in the simultaneous equations (a.k.a. "equilibrium equations") also represents the coefficients in a set of dynamical equations describing a deterministic dynamical system. Fisher [5] gave one such interpretation as follows: There is a relatively long observation period of length 1, and a much shorter "reaction" lag of length = ¯ 1/n. The observed variable is the vector x[t], defined as the average of x over the observation period starting at t: n 1k ¯ x[t + k ] (3) x[t] n =1 ¯ ¯ xa = ba,a xa + k =a,k=1 ¯ ba,k xk + ea (6) If ba,a = 1, it is possible to rewrite this as: n Suppose that the underlying dynamical equations are: x[t + k ] = Bdyn x[t + (k - 1)] + e (4) ¯ ¯ xa - ba,a xa = k =a,k=1 ¯ ba,k xk + ea (7) where e is constant over the observation period (but may differ for different units in the population, e.g. different observation periods). Fisher showed that, in the limit as approaches 0, there is a Bequil = Bdyn such that: ¯ ¯ x[t] = Bequil x[t] + e (5) k 1 ¯ xa = 1 - ba,a a n ¯ ba,k xk + ea = kn =1 ¯ ba,k xk +ea (8) =a,k=1 if and only if the modulus of each eigenvalue of Bdyn is less than or equal to 1, and no eigenvalue is equal to 1. The assumptions underlying this model are fairly strong, but commonly made in econometrics, and defended by Fisher [5]. A simpler, but similar interpretation with similar consequences is the one in which the observed value x is the state in which the dynamical system converged, rather than its average over an observation period. 1.2.1 Dealing with self-lo ops where b ,a = 0. The modified system of equations containing Equation 8 is represented by a graph that has no self-loops, and has a different underlying dynamical equation in which the coefficient for xa [t + (k - 1)] in the equation for xa [t + k ] is zero. Note that in the second equation, the error term ea has been rescaled by 1/(1 - ba,a ) to form a new error term ea and when (I - B )-1 is taken to form the reduced form coefficients, the coefficients corresponding to ea in the first set of equations will be multiplied by (1 - ba,a ), and the two changes cancel each other out. Now, if we consider the original dynamical system and the one that results from setting the diagonal of B to zero (as above), it is sometimes the case that one dynamical system satisfies the conditions for the dynamical equations to approach the simultaneous equations 2 Fisher argues that self-loops are not realistic, but these arguments are not entirely convincing. The LiNG discovery algorithms presented in this paper (described in section 4) output a set of directed graphs in the limit, while the other one does not, because the magnitude of the coefficients in the equation for xa [t] are different. If both forms satisfy Fisher's conditions, then the act of manipulating any variable to a fixed distribution for all t makes the two sets of dynamical equations have equivalent limiting simultaneous equations. 1.2.2 Self-lo ops with co efficient 1 Unfortunately, the case where ba,a = 1 cannot be handled in the same way, since 1/(1 - ba,a ) is infinite. If ba,a = 1, then there may be no equivalent form without a self-loop (or more precisely, the corresponding equations without a self-loop may require setting the variance of some error terms to zero). The case where ba,a = 1 is a genuine problem that we do not currently have a solution for. For the purposes of this paper, we assume that no self-loops have a coefficient of 1. As Dash has pointed out [4], there are cases where the simultaneous equations have a different graph than the underlying dynamical equations, and hence the graph that represents the simultaneous equations cannot be used to predict the effects of a manipulation of the underlying dynamical system. In [4], Dash presents two such examples. In both of them, in effect, Bdyn has a 1 in the diagonal. the linear coefficients, and features common to those directed graphs (such as ancestor relations). The algorithm performs a series of statistical tests of zero partial correlations to construct the PAG. The set of zero partial correlations that is entailed by a linear SEM with uncorrelated errors depends only upon the linear coefficients, and not upon the distribution of the error terms. Under some assumptions4 , in the large sample limit, CCD outputs a PAG that represents the true graph. There are a number of limitations to this algorithm. First, the set of DGs contained in a PAG can be large, and while they all entail the same zero partial correlations (viz., those judged to hold in the population), they need not entail the same joint distribution or even the same covariances. Hence in some cases, the set represented by the PAG will include cyclic graphs that do not fit the data well. Therefore, even assuming that the errors are all Gaussian, it is possible to reduce the size of the set of graphs output by CCD, although in practice this can be intractable. For details on the algorithm, see [11]. 3 Shimizu et al's approach for discovering LiNGAM SEMs 2 2.1 The problem and its history The problem of DG causal discovery Using the interpretations from 1.2, we can frame the problem as follows: given samples of the equilibrium distribution of a LiNG process whose observed variables form a causally sufficient set 3 , find the set of SEMs that describe this distribution, under the assumption that it is non-empty. 2.2 Richardson's Cyclic Causal Discovery (CCD) Algorithm The "LiNGAM algorithm"[12], which uses Independent Components Analysis (ICA), reliably discovers a unique correct LiNGAM SEM, under the following assumptions about the data: the structural equations of the generating process are linear and can be represented by an acyclic graph; the error terms have non-zero variance; the samples are independent and identically distributed; no more than one error term is Gaussian; and the error terms are jointly independent.5 3.1 Indep endent Comp onents Analysis (ICA) While many algorithms have been suggested for discovering (equivalence classes of ) directed acyclic graphs (DAGs) from data, for general linear directed graphs (DGs) only one provably correct algorithm was known (until now), namely Richardson's Cyclic Causal Discovery (CCD) algorithm. CCD outputs a "partial ancestral graph" (PAG) that represents both a set of directed graphs that entail the same set of zero partial correlations for all values of 3 A set V of variables is causally sufficient for a population if and only if in the population every common direct cause of any two or more variables in V is in V . (For subtleties regarding this definition, see [13]). Independent components analysis [3, 8] is a statistical technique used for estimating the mixing matrix A in equations of the form x = Ae (e is often called "sources" and written s), where x is observed and e and A are not. ICA algorithms find the invertible linear transformaThe assumptions are: the samples are independent and identically distributed, no error term has zero variance, the statistical tests for zero partial correlations are consistent, linearity of the equations, the existence of a unique reduced form, faithfulness (i.e. there are no zero partial correlations in the population that are not entailed for all values of the free parameters of the true graph), and that the error terms are uncorrelated. 5 The error terms are typically not jointly independent if the set of variables is not causally sufficient. 4 tion W = A-1 of the data X that makes the error distributions corresponding to the implied samples E of e maximally non-Gaussian (and thus, maximally independent). The matrix A can be identified up to scaling and permutation as long as the observed distribution is a linear, invertible mixture of independent components, at most one of which is Gaussian [3]. There are computationally efficient algorithms for estimating A [8]. 3.2 The LiNGAM discovery algorithm After the algorithm finds the correct permutation, it finds the correct scaling, i.e. "normalizing" W by dividing each row by its diagonal element, so that the diagonal of the output matrix is all 1s (i.e. the coefficient of each error term is 1, as specified in Section 1). Bringing it all together, the algorithm computes B by using B = I - W , where W = normaliz e(W ), W = RowP ermute(WI C A ) and WI C A = I C A(X ). Besides the fact that it determines the direction of every causal arrow, another advantage of LiNGAM over conditional-independence-based methods [13] is that the correctness of the algorithm does not require the faithfulness assumption. For more details on the LiNGAM approach, see [12]. If we run an ICA algorithm on data generated by a linear SEM, the matrix WI C A obtained will be a rowscaled, row-permuted version of I - B , where B is the coefficient matrix of the true model (this is a consequence of the derivation in Section 1.1). We are now left with the problem of finding the proper permutation and scale for the W matrix so that it equals I - B . 4 Discovering LiNG SEMs Fig. 2: After removing the edges whose coefficients are statistically indistinguishable from zero: (a) the raw WI C A matrix output by ICA on a SEM whose graph is x2 x1 x3 (b) the corresponding W matrix, obtained by permuting the error terms in WI C A The assumptions of the family of LiNG discovery algorithms described below (abbreviated as "LiNG-D") are the same as the LiNGAM assumptions, replacing the assumption that the SEM is acyclic with the weaker assumption that the diagonal of Bdyn contains no 1s. In this more general case, as in the acyclic case, candidate models are generated by finding all admissible matches of the error terms (ei 's) to the observed variables (xi 's). In other words, each candidate corresponds to a row-permutation of the WI C A matrix that has a zeroless diagonal. As in LiNGAM, the output is the set of admissible models. In LiNGAM, this set is guaranteed to contain a single model, thanks to the acyclicity assumption. If the true model has cycles, however, more than one model will be admissible. The remainder of this section addresses the problem of finding the admissible models, given that ICA has finite data to work with. 4.1 Prune and solve Constrained n-Ro oks Since the order of the error terms given by ICA is arbitrary, the algorithm needs to correctly match each error term ei to its respective substantive variable xi . This means finding the correct permutation of the rows of WI C A . We know that the row-permutation of WI C A corresponding to the correct model cannot have a zero in the diagonal (we call such permutations "inadmissible") because W = I - B , and the diagonal of B is zero. Since, by assumption, the data was generated by a DAG, there is exactly one row-permutation of WI C A that is admissible [12]. To visualize this, this constraint says that there is exactly one way to reorder the error terms so that every ei is the target of a vertical arrow.6 In this example, swapping the first and second error terms is the only permutation that produces an admissible matrix, as seen in Fig. 2(b). Another consequence of acyclicity is that there will be no right-pointing arrows in this representation, provided that the xs are topologically sorted w.r.t. the DAG. 6 These algorithms generate candidate models by testing which entries of WI C A are zero (i.e. pruning), and finding all admissible permutations based on that (i.e. solving Constrained n-Rooks, see Section 4.1.2). We call an algorithm "local" if, for each entry wi,j of WI C A , it makes a decision about whether wi,j is zero using only wi,j . 4.1.1 Deciding which entries are zero There are several methods for deciding which entries of WI C A to set to zero: · Thresholding: the simplest method for estimating which entries of WI C A are zero is to simply choose a threshold value, and set every entry of WI C A smaller than the threshold (in absolute value) to zero. This method fails to account for the fact that different coefficients may have different spreads, and will miss all coefficients smaller than the threshold. · Test the non-zero hyp othesis by b o otstrap sampling: another method for estimating which entries of WI C A are actually zero is to do bootstrap sampling. Bootstrap samples are created by resampling with replacement from the original data. Then ICA is run on each bootstrap sample, and each coefficient wi,j is calculated for each bootstrap sample. This leads to a real-valued distribution for each coefficient.7 Then, for each one, a non-parametric quantile test is performed in order to decide whether 0 is an outlier. If it isn't, the coefficient is set to 0 (i.e. the corresponding edge is pruned.)8 · Use sparse ICA: Use an ICA algorithm that returns a sparse (i.e. pre-pruned) mixture, such as the one presented by Zhang and Chan [16]. Unlike the other methods above, this is not a local algorithm. 4.1.2 Constrained n-Ro oks: the problem and an algorithm 4.2 A non-lo cal algorithm Local algorithms work under the assumption that the estimates of the wi,j are independent of each other ­ which is in general false when estimating with finite samples. This motivates the use of non-local methods. In the LiNGAM (acyclic) approach [12], a non-local algorithm is presented for finding the single best rowpermutation of WI C A , which minimizes a loss function that heavily penalizes entries in the diagonal that are close to zero (such as x |1/x|). This is written as a linear assignment problem (i.e. finding the best match between the ei s and xi s), which can be solved using the Hungarian algorithm [9] or others. For general LiNG discovery, however, algorithms that find the best linear assignment do not suffice, since there may be multiple admissible permutations. One idea is to use a k -th best assignment algorithm [2] (i.e. the k -th permutation with the least penalty on the diagonal), for increasing k . With enough data, all permutations corresponding to inadmissible models will score poorly, and there should be a clear separation between admissible and inadmissible models. The non-local method presented above, like the thresholding method, fails to account for differences in spread among estimates of the entries of WI C A . It would be straightforward to fix this by modifying the loss function to penalize diagonal entries for which the test fails to reject the null hypothesis (as described in the part about bootstrap sampling in Section 4.1.1), instead of penalizing them for merely being close to zero. 4.3 Sample run Once it is decided which entries are zero, the algorithm searches for every row-permutation of WI C A that has a zeroless diagonal. Each such row-permutation corresponds to a placement of n rooks onto the non-zero entries on an n × n chessboard such that no two rooks threaten each other. Then the rows are permuted so that all the rooks end up on the diagonal, thus ensuring that the diagonal has no zeros. To solve this problem, we use a simple depth-first search that prunes search paths that have nowhere to place the next rook. In the worst case, every permutation is admissible, and the search must take O(n!). One needs to be careful when doing this, since each run of ICA may return a WI C A in a different row-permutation. This means that we first need to row-permute each bootstrap WI C A to match with the original WI C A . 8 One could ob ject that, instead of a quantile test, the correct procedure would be to simulate under the null hypothesis (i.e.: edge is absent) using the estimated error terms, and then compare the obtained distribution of the ICA statistics with their distribution for the bootstrap. However, this raises issues and complexities that are tangential to the current paper. 7 We generated 15000 sample points using the SEM in Example 1 and error terms distributed according to a symmetric Gaussian-squared distribution9 . Fig. 3 shows the output of the local thresholding algorithm with the cut-off set to 0.05. For the sake of reproducibility, our code with instructions is available from: www.phil.cmu.edu/~tetrad/ cd2008.html . 5 5.1 Theory Notions of DG equivalence There are a number of different senses in which the directed graphs associated with SEMs can be "equivalent" or "indistinguishable" given observational data, The distribution was created by sampling from the standard Gaussian(0,1) and squaring it. If the value sampled was negative, it was made negative again. 9 the error terms are assumed to be Gaussian, then distribution equivalence entails (but is not entailed by) covariance equivalence, which entails (but is not entailed by) d-separation equivalence. So for example, given Gaussian error terms, A B and A B are zero partial correlation equivalent, covariance equivalent, and distribution equivalent. But given non-Gaussian error terms, A B and A B are zero-partial-correlation equivalent and covariance equivalent, but not distribution equivalent. So for Gaussian errors and this pair of DGs, no algorithm that relies only on observational data can reliably select a unique acyclic graph that fits the population distribution as the correct causal graph without making further assumptions; but for all (or all except one) nonGaussian errors there will always be a unique acyclic graph that fits the population distribution. While there are theorems about the case of cyclic graphs and Gaussian errors, we are not aware of any such theorems about cyclic graphs with non-Gaussian errors with respect to distribution equivalence. In the case of cyclic graphs with all Gaussian errors, distribution equivalence is equivalent to covariance equivalence, which entails (but is not entailed by) dseparation equivalence [14]. In the case of cyclic graphs in which at most one error term is non-Gaussian, distribution equivalence entails (but is not entailed by) covariance equivalence, which in turn entails (but is not entailed by) d-separation equivalence. However, given at most one Gaussian error term, the important difference between acyclic graphs and cyclic graphs is that no two different acyclic graphs are distribution equivalent, but there are different cyclic graphs that are distribution equivalent. Hence, no algorithm that relies only on observational data can reliably select a unique cyclic graph that fits the data as the correct causal graph without making further assumptions. For example, the two cyclic graphs in Fig. 3 are distribution equivalent. 5.2 The output of LiNG-D is correct and as fine as p ossible Fig. 3: The output of LiNG-D: Candidate #1 and Candidate #2 assuming linearity and no dependence between error terms: · DGs G1 and G2 are zero partial correlation equivalent if and only if the set of zero partial correlations entailed for all values of the free parameters (non-zero linear coefficients, distribution of the error terms) of a linear SEM with DG G1 is the same as the set of zero partial correlations entailed for all values of the free parameters of a linear SEM with G2 . For linear models, this is the same as d-separation equivalence. [13] · DGs G1 and G2 are covariance equivalent if and only if for every set of parameter values for the free parameters of a linear SEM with DG G1 , there is a set of parameter values for the free parameters of a linear SEM with DG G2 such that the two SEMs entail the same covariance matrix over the substantive variables, and vice-versa. · DGs G1 and G2 are distribution equivalent if and only if for every set of parameter values for the free parameters of a linear SEM with DG G1 , there is a set of parameter values for the free parameters of a linear SEM with DG G2 such that the two SEMs entail the same distribution over the substantive variables, and vice-versa. Do not confuse this with the notion of distribution-entailment equivalence between SEMs: two SEMs with fixed parameters are distribution-entailment equivalent iff they entail the same distribution. It follows from well-known theorems about the Gaussian case [13], and some trivial consequences of known results about the non-Gaussian case [12], that the following relationships exist among the different senses of equivalence for acyclic graphs: If all of the error terms are assumed to be Gaussian, distribution equivalence is equivalent to covariance equivalence, which in turn is equivalent to d-separation equivalence. If not all of Theorem 1 The output of LiNG-D is a set of SEMs that comprise a distribution-entailment equivalence class. Pro of: First, we show that any two SEMs in the output of LiNG-D entail the same distribution. The weight matrix output by ICA is determined only up to scaling and row permutation. Intuitively, then, permuting the error terms does not change the mixture. Now, more formally: Let M1 and M2 be candidate models output by LiNGD. Then W1 and W2 are row-permutations of WI C A : W 1 = P1 W I C A , W 2 = P2 W I C A Likewise, for the error terms: E1 = P1 E , E2 = P2 E Then the list of samples X implied by M1 is A1 E1 = -1 (W1 )-1 E1 = (P1 WI C A )-1 (P1 E ) = WI C A P1 -1 P1 E = -1 WI C A E . By the same argument, the list of samples X implied -1 by M2 is also WI C A E . Therefore, any two SEM models output by LiNG-D entail the same distribution. Now, it remains to be shown that if LiNG-D outputs one SEM that entails a distribution P , it outputs all SEMs that entail P . Suppose that there is a SEM S that represents the same distribution as some T , which is output by LiNG-D. Then the reduced-form coefficient matrices for S and T , AS and AT , are the same up to columnpermutation and scaling. Hence, I - BS and I - BT are also the same up to scaling and row-permutation (by I - B = A-1 ). By the assumption that there are no self-loops with coefficient 1, neither I - BT nor I - BS has zeros on the diagonal. Since I - BT is a scaled row-permutation of WI C A that has no zeros on the diagonal, so is I - BS . Thus S is also output by T LiNG-D. heorem 2 If the simultaneous equations are linear and can be represented by a directed graph; the error terms have non-zero variance; the samples are independently and identical ly distributed; no more than one error term is Gaussian; and the error terms are jointly independent, then in the large sample limit, LiNG-D outputs al l SEMs that entail the population distribution. Pro of: ICA gives pointwise consistent estimates of A and W under the assumptions listed [3]. This entails that there are pointwise consistent tests of whether an entry in the W matrix is zero, and hence (by definition) in the large sample limit, the limit of both type I and type II errors of tests of zero coefficients are zero. Given the correct zeroes in the W matrix, the output of the local version of the LiNG-D algorithm is correct in the sense that the simultaneous equation describes the population distribution. I n general, each candidate model B = I - W has the structure of a row-permutation of WI C A . The structures can be generated by analyzing what happens when we permute the rows of W . Remember that edges in B (and thus W ) are read column-torow. Thus, row-permutations of W change the positions of the arrow-heads (targets), but not the arrow- tails (sources). Richardson proved that the operation of reversing a cycle preserves the set of entailed zero partial correlations, but did not consider distribution equivalence [11]. 5.3 Adding the assumption of stability In dynamical systems, "stable" models are ones in which the effects of one-time noise dissipate. For example, a model that has a single cycle whose cycleproduct (product of coefficients of edges in the cycle) is 1 is unstable, while one that has a single cycle whose cycle-product is between -1 and 1 is stable. On the other hand, if a positive feedback loop of cycle-product 2 is counteracted by a negative loop with cycle-product -1.5, then the model is stable, because the effective cycle-product is 0.5. A general way to express stability is lim B t = 0, t which is mathematically equivalent to: for all eigenvalues e of B , |e| < 1, in which |z | means the modulus of z . This eigenvalues criterion is easy to compute. Given only the coefficients between different variables, it is impossible to measure the stability of a SEM without assuming something about the self-loops. Therefore, in this section, it is assumed that the true model has no self-loops. It is often the case that many of the SEMs output by LiNG-D are unstable. Since in many situations, the variables are assumed to be in equilibrium, we are often allowed to rule out unstable models. In the remainder of this section, we will prove that if the SEM generating the population distribution has a graph in which the cycles are disjoint, then among the candidate SEMs output by LiNG-D, at most one will be stable. Theorem 3 SEMs in the form of a simple cycle with a cycle-product such that | | 1 are unstable. Pro of: Let k be the length of the cycle. Then B k = I . Then for all integers i, B ik = i I . So if | | 1, the entries of B ik do not get smaller than the entries of B as i increases. Thus, B t will not converge to 0 as C t . orollary 1: For SEMs in the form of a simple cycle, having a cycle-product 1 is equivalent to having an eigenvalue 1 (in modulus), which is equivalent to being unstable. Theorem 4 Suppose that there is a SEM M with disjoint cycles with coefficient matrix B and graph G that entails a distribution Q, and a SEM M0 = M with graph G0 , coefficient matrix B0 , which is an admissible permutation of M and also entails Q. Then G0 also contains disjoint cycles, at least one of which is a reversal of a cycle C in G, whose cycle-product is the inverse of the cycle-product of C . Pro of: Due to space limitations, the proof is just sketched here. Every permutation can be represented as a product of disjoint cyclic subpermutations of the form a b . . . m n a, where a b means a gets mapped onto b. (Some cyclic subpermutations may be trivial, i.e. contain a single object mapped onto itself ). Hence it suffices to prove the theorem for a single admissible cyclic row permutation of B . It can be shown that if a cyclic row permutation of B , a b . . . m n a is admissible, then G contains the cycle C equal to a b . . . m n a, and G0 contains the reversed cycle C equal to a b . . . m n a. Moreover, if G0 contains two cycles that touch, so does G. Consider BC , the submatrix of B that contains the coefficients of the edges in cycle C . 0 ... 0 bk,1 b1,2 0 ... 0 BC = .. 0 . b2,3 0 .. . 0 0 0 k -1 Note that the cycle-product C = bk,1 i=0 bi,i+1 . WC = I - BC . The "reversal" is the row-permutation in which first row gets "rotated" into the bottom: -b1,2 1 ... 0 . 0 -b2,3 . . 0 RowP ermute(WC ) = .. 0 . 0 1 1 0 ... -bk,1 the case, the eigenvalues of the cycles). Suppose a cycle of G is unstable. Then it has an eigenvalue 1 (in modulus). But since this is also an eigenvalue of BG , it follows that G is unstable. The other direction goes T similarly. heorem 6 If the true SEM is stable and has a graph in which the cycles are disjoint, then no other SEMs in the output of LiNG-D wil l be stable. Pro of: Suppose the true SEM is stable and has a graph in which the cycles are disjoint. Call it G. Since, by Theorem 2, the output of LiNG-D are the admissible distribution-entailment equivalent alternatives to the true SEM, it suffices to show that all other admissible candidates are unstable. By Theorem 5, all cycles in G are stable. Let H be an admissible alternative to G, such that H = G. By Theorem 4, H will have at least one cycle C reversed relative to G and this reversed cycle will have a cycle product that is the inverse of the cycle product of C . By Corollary 1, the reversed matrix is not stable. Thus, by Theorem 5, H is unstable. Therefore, the only stable admissible alternative to G I is G itself. t follows that if the true model's cycles are disjoint, then under the assumption that the true model is stable, we can fully identify it using a LiNG discovery algorithm (at most one SEM in the output of the LiNG discovery algorithm will be stable). For example, consider the two candidate models shown in Fig. 3. By assuming that the true model is stable, one would select candidate #2. Since our simulation used a stable model, this is indeed the correct answer (see Fig. 1). In general, however, there may be multiple stable models, and one cannot reliably select the correct one. When the cycles are not disjoint, it is easy to find examples for which there are multiple stable candidates. The condition of disjoint cycles is sufficient, but not necessary: it is easy to come up with SEMs where we have exactly one stable SEM in the distributionentailment equivalence class, despite intersecting cycles. Normalizing the diagonal to be all 1s, we get WC . Computing BC = I - WC , one can see that the cyclek -1 1 W product C = bk,1 i=0 bi,1+1 = 1/C . i e will now show that for SEMs in which the cycles are disjoint, their stability only depends on the stability of the cycles. Theorem 5 A SEM in which the cycles are disjoint is stable if and only if it has no unstable cycles. Pro of: Let be G be a SEM whose cycles are disjoint. Then BG can be written as a block-triangular matrix where each diagonal block is a cycle. The set of eigenvalues of a block-triangular matrix is the union of the sets of eigenvalues of the blocks in the diagonal (in this 6 Discussion We have presented Shimizu's approach for discovering LiNGAM SEMs, and generalized it to a method that discovers general LiNG SEMs. This improves upon the state-of-the-art on cyclic linear SEM discovery by outputting only the distribution-entailment equivalence class of SEMs, instead of the entire d-separation equiv- alence class; and by relaxing the faithfulness assumption. We have also shown that stability can be a powerful constraint, sometimes narrowing the candidates to a single SEM. There are a number of questions that remain open for future research: · The LiNG-D algorithm generates all admissible permutations. The worst-case time-complexity of n-Rooks is high, but can we do better than depthfirst search for random instances? Is there an algorithm to efficiently search for the stable models, without going through all candidates? In the case where the cycles are disjoint, it is possible to just find the correct permutation for each cycle independently, but no such trick is known in general. · How can prior information be incorporated into the algorithm? · How can the algorithm be modified to allow the assumption of causal sufficiency to be relaxed? For the acyclic case, see [7]. · How can the algorithm be modified to allow for mixtures of non-Gaussian and Gaussian (or almost Gaussian) error terms? Hoyer et al [6] address this problem for the acyclic case. · How could we integrate this method into mainstream dynamical systems research? Can the algorithm handle noisy dynamics and noisy observations? Could it be made to handle non-linear dynamics? What about self-loops of coefficient 1? How could one integrate this with methods that use non-equilibrium time-series data? Acknowledgements The authors wish to thank Anupam Gupta, Michael Dinitz and Cosma Shalizi. GL was partially supported by NSF Award No. REC-0537198. [4] D. Dash (2005) - Restructuring Dynamic Causal Systems in Equilibrium. Proceedings of the Tenth International Workshop on Artificial Intel ligence and Statistics (AIStats 2005) [5] F. Fisher (1970) - A correspondence principle for simultaneous equation models. Econometrica, 38(1):73-92. [6] P. O. Hoyer, A. Hyv¨rinen, R. Scheines, P. a Spirtes, J. Ramsey, G. Lacerda, S. Shimizu (2008) - Causal discovery of linear acyclic models with arbitrary distributions. Proceedings of the 24th Conference on Uncertainty in Artificial Intel ligence. [7] P. O. Hoyer, S. Shimizu, and A. J. Kerminen (2006) - Estimation of linear, non-gaussian causal models in the presence of confounding latent variables Proc. Third European Workshop on Probabilistic Graphical Models (PGM'06), pp. 155-162 [8] A. Hyv¨rinen, J. Karhunen, E. Oja (2001) - Indea pendent Component Analysis. Wiley Interscience. [9] H. W. Kuhn (1955) - The Hungarian Method for the assignment problem. Naval Research Logistics Quarterly, 2:83-97, 1955. [10] J. Pearl (2000) - Causality: Models, Reasoning, and Inference. Cambridge University Press. [11] T. Richardson (1996) - A Polynomial-Time Algorithm for Deciding Markov Equivalence of Directed Cyclic Graphical Models. Proceedings of the 12th Conference on Uncertainty in Artificial Intel ligence [12] S. Shimizu, P. Hoyer, A. Hyv¨rinen, A. Kermia nen (2006) - A linear, non-gaussian acyclic model for causal discovery. Journal of Machine Learning Research 7:2003-2030. [13] P. Spirtes, C. Glymour, R. Scheines (1993) - Causation, Prediction, Search. Springer-Verlag Lecture Notes in Statistics 81. [14] P. Spirtes (1995) - Directed Cyclic Graphical Representation of Feedback Models, Proceedings of the Eleventh Conference on Uncertainty in Artificial Intel ligence [15] R. Strotz and H. Wold (1960) - Recursive versus nonrecursive systems: an attempt at synthesis. Econometrica 28:417-427. [16] K. Zhang, L-W. Chan (2006) - ICA with Sparse Connections Intel ligent Data Engineering and Automated Learning IDEAL 2006. References [1] K. Bollen (1989) - Structural Equations with Latent Variables, John Wiley & Sons, New York. [2] C. R. Chegireddy, H. W. Hamacher (1987) - Algorithms for finding K-best perfect matchings Discrete Applied Mathematics, 18:155-165. [3] P. Comon (1994) - Independent component analysis - a new concept? Signal Processing, 36:287314.