183
Views
0
CrossRef citations to date
0
Altmetric
Original Articles

Learning linear non-Gaussian networks: A new view from matrix identification

, &
Pages 251-271 | Received 04 Jun 2012, Accepted 20 Jun 2012, Published online: 22 Oct 2012
 

Abstract

Identifying causal structures from observations is fundamental in many applications. Probabilistic graphical models provide a unifying framework for capturing complex causal dependencies among random variables. Recently, a linear non-Gaussian acyclic model (LiNGAM) and some smart algorithms (ICA-LiNGAM, DirectLiNGAM) have been proposed, which outperform previous graphical models and learning methods in identifying variable orders. We propose new solutions (TMFLiNGAM, SchurLiNGAM and RCLiNGAM) to the LiNGAM learning task from the perspective of matrix identification. TMFLiNGAM and SchurLiNGAM recover orders more directly, and RCLiNGAM can improve the accuracy of previous algorithms on uniform and sparse structures. The perspective also facilitates the learning of sparse models where the performance of all independent component analysis-based algorithms can be improved by reconstructing variable orders from the inverse of separation matrices. Experimental results under various settings provide average evaluations over the learning methods, and verify the effectiveness of our perspective and algorithms.

Acknowledgements

We would like to thank the anonymous editors and reviewers for their constructive suggestions and comments. This work was supported by the National Natural Science Foundation of China (Project Nos 60772050 and 61071131), No. 2 Important National Science and Technology Specific Projects (Project No. 2009ZX02001), the National Key Basic Research and Development Program of China (No. 2009CB320602) and United Technologies Research Center (UTRC).

Notes

Notes

1. For example, it is difficult to determine whether x 1 → x 2 (x 1 causes x 2) or x 2 → x 1 (x 2 causes x 1) from a bivariate Gaussian distribution since x 1 and x 2 are symmetrical in p(x 1, x 2) (Daniušis et al. Citation2010).

2. c implies that the expectation of disturbances can always be assumed to 0, and the LiNGAM equation can be simplified as x = B x + e by a translation transform x ← x − (I − B)−1 c.

3. Learning structures are more important and difficult in our tasks, so we focus on estimating the variable orders. In fact, the parameters of LiNGAM, i.e. the coefficients between observed variables, can be easily determined by regression after the variable orders are correctly identified.

4. P x incorporates the variable information, and is equivalent in describing the variable order as K. Representing the variable order from K to P x brings clearer description of the LiNGAM learning problem since all the variables can now be written as matrices.

5. This statement holds in general situations such as all the connection strength coefficients are non-zero, but does not hold in some situations such as B is sparse (many coefficients are zero). In the latter case, Q is not unique since the LiNGAM learning problem is underdetermined.

6. In automatic control, signal processing and machine learning, the system state vector x is usually hidden and unobservable, but can be measured as a ‘observation vector’ y by various kinds of sensors, which is usually represented by a linear mapping y = T x.

7. We get Q from in ICA-LiNGAM, and find a permutation matrix P x closest to Q by linear assignment algorithm. This modification of ICA-LiNGAM can be called SchurLiNGAM. Our numerical experiments in Section 7 shows that determining P x from by Schur factorisation requires more observational data to achieve similar accuracies as previous ICA-LiNGAM and DirectLiNGAM algorithms.

8. Identifying the permutation and scaling can remove the unknown scaling of each row in the connection matrix B, and can improve the performance of RCLiNGAM algorithm. However, this step increases computational complexity since it calls the linear assignment algorithm (Hoyer et al. Citation2006).

Log in via your institution

Log in to Taylor & Francis Online

PDF download + Online access

  • 48 hours access to article PDF & online version
  • Article PDF can be downloaded
  • Article PDF can be printed
USD 61.00 Add to cart

Issue Purchase

  • 30 days online access to complete issue
  • Article PDFs can be downloaded
  • Article PDFs can be printed
USD 373.00 Add to cart

* Local tax will be added as applicable

Related Research

People also read lists articles that other readers of this article have read.

Recommended articles lists articles that we recommend and is powered by our AI driven recommendation engine.

Cited by lists all citing articles based on Crossref citations.
Articles with the Crossref icon will open in a new tab.