Abstract
Efficiently inducing precise causal models accurately reflecting given data sets is the ultimate goal of causal discovery. The algorithm proposed by Wallace et al. [10] has demonstrated its ability in discovering Linear Causal Models from data. To explore the ways to improve efficiency, this research examines three different encoding schemes and four searching strategies. The experimental results reveal that (1) specifying parents encoding method is the best among three encoding methods we examined; (2) In the discovery of linear causal models, local Hill climbing works very well compared to other more sophisticated methods, like Markov Chain Monte Carto (MCMC), Genetic Algorithm (GA) and Parallel MCMC searching.
Chapter PDF
Similar content being viewed by others
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
References
K. A. Bollen. Structural Equations with Latent Variables. Wiley, New York, 1989.
Honghua Dai, Kevin Korb, Chris Wallace, and Xindong Wu. A study of causal discovery with small samples and weak links. In Proceedings of the 15th International Joint Conference On Artificial Intelligence IJCAI’97, pages 1304–1309. Morgan Kaufmann Publishers, Inc., 1997.
Honghua Dai and Gang Li. An improved approach for the discovery of causal models via MML. In Proceedings of The 6th Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD-2002), pages 304–315, Taiwan, 2002.
C. Holmes and B. Mallick. Parallel markov chain monte carlo sampling: an evolutionary based approach. Technical report, Imperial College, London, 1998.
Michael I. Jordan. Learning in Graphical Models. MIT Press, Cambridge, MA, 1 edition, 1998.
P. Larranaga, M. Poza, Y. Yurramendi, R. H. Murga, and C. M. H. Kuijpers. Structure learning of bayesian networks by genetic algorithms: A performance analysis of control parameters. IEEE Transactions on Pattern analysis and Machine Intelligence, 18(9):912–926, 1996.
D. Madigan and J. York. Bayesian graphical models for discrete data. Technical Report TR-259, University of Washington Department of Statistics, Seattle, WA, 1993.
Judea Pearl. Probabilistic Reasoning in Intelligent Systems. Morgan Kaufmann, San Mateo, California, 1988.
Chris Wallace and D. M. Boulton. An information measure for classification. Computer Journal, 11(2):185–194, 1968.
Chris Wallace, Kevin B. Korb, and Honghua Dai. Causal discovery via MML. In Proceedings of the 13th International Conference on Machine learning (ICML’96), pages 516–524, San Francisco, 1996. Morgan Kauffmann Publishers.
W. L. Wong, W. Lam, and K. S. Leung. Using evolutionary computation and minimum description length principle for data mining of probabilistic knowledge. IEEE Transactions on Pattern Analysis and Machine Intelligence, 21(2):174–178, 1999.
Sewall Wright. The method of path coefficients. Annals of Mathematical Statistics, 5:161–215, 1934.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2002 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Dai, H., Li, G., Tu, Y. (2002). An Empirical Study of Encoding Schemes and Search Strategies in Discovering Causal Networks. In: Elomaa, T., Mannila, H., Toivonen, H. (eds) Machine Learning: ECML 2002. ECML 2002. Lecture Notes in Computer Science(), vol 2430. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-36755-1_5
Download citation
DOI: https://doi.org/10.1007/3-540-36755-1_5
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-44036-9
Online ISBN: 978-3-540-36755-0
eBook Packages: Springer Book Archive