Abstract
What can be predicted when the causal structure and the joint distribution among a set of random variables is known that cannot be predicted when only the joint distribution over the set of random variables is known? The answer is that with the former we can predict the effects of intervening in a system by manipulating the values of certain variables, while with only the latter we cannot. For example, if we know only the joint distribution of smoking and lung cancer, we cannot determine whether stopping people from smoking will reduce the rates of lung cancer. On the other hand, if we also know that smoking causes lung cancer (and that there is no common cause of smoking and lung cancer), we can predict that stopping smoking will reduce lung cancer, and by how much. As the quotations at the beginning of the article show, there is a debate within the statistics community about whether predicting the effects of interventions from passive observations is possible at all. In this paper we will describe some recent work for predicting the effects of interventions and policies given passive observations and some background knowledge. While addressing some of the claims just considered, the results we describe unify two traditions in statistics-one, beginning at least as early as Sewell Wright ([Wright 34]; [Simon 77]; [Blalock 61]; [Kiiveri 82]; [Wermuth 83]; [Lauritzen 84]; [Kiiveri 84]; [Wright 34]; [Glymour 87]; [Pearl 88]), connects directed acyclic graphs with probability distributions through constraints on conditional independence relations, while the other ([Neyman 35]; [Rubin 77]; [Rubin 78]; [PearVerm 91]) connects causal claims with “counterfactual distributions” and offers rules for predicting the distribution of a variable that will result if other variables are deliberately manipulated ([Rubin 77]; [Pratt 88]). We consider the following two cases.
Article Note
It is with data affected by numerous causes that Statistics is mainly concerned. Experiment seeks to disentangle a complex of causes by removing all but one of them, or rather by concentrating on the study of one and reducing the others, as far as circumstances permit, to comparatively small residium. Statistics, denied this resource, must accept for analysis data subject to the influence of a host of causes, and must try to discover from the data themselves which causes are the important ones and how much of the observed effect is due to the operation of each.
-G. U. Yule and M. G. Kendall,1950
George Box has [almost] said “The only way to find out what will happen when a complex system is disturbed is to disturb the system, not merely to observe it passively.” These words of caution about “natural experiments” are uncomfortably strong. Yet in today’s world we see no alternative to accepting them as, if anything, too weak.
-G. Mosteller and J. Tukey, 1977
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Blalock, H. (1961) Causal Inferences in Nonexperimental Research. University of North Carolina Press, Chapel Hill, NC.
Cooper, G. and Herskovitz, E. (1991) “A Bayesian Method for the Induction of Probabilistic Networks from Data,” Journal of Machine Learning 9, 309 - 347.
Glymour, C., Scheines, R., Spirtes, P., and Kelly, K. (1987) Discovering Causal Structure. Academic Press, San Diego, CA.
Holland, P. (1986) “Statistics and Causal Inference,” JASA 81, 945–960.
Kiiveri, H. and Speed, T. (1982) “ Structural Analysis of Multivariate Data: A Review,” in Sociological Methodology, S. Leinhardt, ed. Jossey-Bass, San Francisco.
Kiiveri, H., Speed, T., and Carlin, J. (1984) “Recursive Causal Models,” Journal of the Australian Mathematical Society 36, 30–52.
Lauritzen, S., and Wermuth, N. (1984) “Graphical Models for Associations Between Variables, Some of Which are Quantitative and Some of which are Qualitative,” Ann. Stat. 17 31–57.
Mosteller, F., and Tukey, J. (1977) Data Analysis and Regression, A Second Course in Regression. Addison-Wesley, Massachusetts.
Neyman, J. (1935) “Statistical Problems with Agricultural Experimentation,” J. Roy. Stat. Soc. Suppl. 2 107–180.
Pearl, J. (1988) Probabilistic Reasoning in Intelligent Systems. Morgan Kaufman, San Mateo, CA.
Pratt, J. and Schlaifer, R. (1988) “On the Interpretation and Observation of Laws,” Journal of Econometrics 39, 23–52.
Rubin, D. (1977) “Assignment to Treatment Group on the Basis of a Covariate,” Journal of Educational Statistics 2, 1–26.
Rubin, D. (1978) “Bayesian Inference for Causal Effects: The Role of Randomizations” Ann. Stat. 6, 34–58.
Simon, H. (1977) Models of Discovery. D. Reidel, Dordrecht, Holland.
Spirtes, P., Glymour, C., and Scheines, R. (1993) Causation, Prediction and Search. Springer-Verlag, Lecture Notes in Statistics, New York.
Verma, T. and Pearl, J. (1990) “Equivalence and Synthesis of Causal Models,” in Proc. Sixth Conference on Uncertainty in Al. Association for Uncertainty in AI, Inc., Mountain View, CA.
Wermuth, N. and Lauritzen, S. (1983) “Graphical and Recursive Models for Contingency Tables,” Biometrika 72, 537–552.
Wright, S. (1934) “The method of Path Coefficients,” Ann. Math. Stat. 5, 161–215.
Yule, G. and Kendall, M. (1937) An Introduction to the Theory of Statistics. Charles Griffin, London.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1994 Springer-Verlag New York, Inc.
About this paper
Cite this paper
Spirtes, P., Glymour, C. (1994). Inference, Intervention, and Prediction. In: Cheeseman, P., Oldford, R.W. (eds) Selecting Models from Data. Lecture Notes in Statistics, vol 89. Springer, New York, NY. https://doi.org/10.1007/978-1-4612-2660-4_22
Download citation
DOI: https://doi.org/10.1007/978-1-4612-2660-4_22
Publisher Name: Springer, New York, NY
Print ISBN: 978-0-387-94281-0
Online ISBN: 978-1-4612-2660-4
eBook Packages: Springer Book Archive