Skip to main content

Inference, Intervention, and Prediction

  • Conference paper
Selecting Models from Data

Part of the book series: Lecture Notes in Statistics ((LNS,volume 89))

Abstract

What can be predicted when the causal structure and the joint distribution among a set of random variables is known that cannot be predicted when only the joint distribution over the set of random variables is known? The answer is that with the former we can predict the effects of intervening in a system by manipulating the values of certain variables, while with only the latter we cannot. For example, if we know only the joint distribution of smoking and lung cancer, we cannot determine whether stopping people from smoking will reduce the rates of lung cancer. On the other hand, if we also know that smoking causes lung cancer (and that there is no common cause of smoking and lung cancer), we can predict that stopping smoking will reduce lung cancer, and by how much. As the quotations at the beginning of the article show, there is a debate within the statistics community about whether predicting the effects of interventions from passive observations is possible at all. In this paper we will describe some recent work for predicting the effects of interventions and policies given passive observations and some background knowledge. While addressing some of the claims just considered, the results we describe unify two traditions in statistics-one, beginning at least as early as Sewell Wright ([Wright 34]; [Simon 77]; [Blalock 61]; [Kiiveri 82]; [Wermuth 83]; [Lauritzen 84]; [Kiiveri 84]; [Wright 34]; [Glymour 87]; [Pearl 88]), connects directed acyclic graphs with probability distributions through constraints on conditional independence relations, while the other ([Neyman 35]; [Rubin 77]; [Rubin 78]; [PearVerm 91]) connects causal claims with “counterfactual distributions” and offers rules for predicting the distribution of a variable that will result if other variables are deliberately manipulated ([Rubin 77]; [Pratt 88]). We consider the following two cases.

Article Note

It is with data affected by numerous causes that Statistics is mainly concerned. Experiment seeks to disentangle a complex of causes by removing all but one of them, or rather by concentrating on the study of one and reducing the others, as far as circumstances permit, to comparatively small residium. Statistics, denied this resource, must accept for analysis data subject to the influence of a host of causes, and must try to discover from the data themselves which causes are the important ones and how much of the observed effect is due to the operation of each.

-G. U. Yule and M. G. Kendall,1950

George Box has [almost] said “The only way to find out what will happen when a complex system is disturbed is to disturb the system, not merely to observe it passively.” These words of caution about “natural experiments” are uncomfortably strong. Yet in today’s world we see no alternative to accepting them as, if anything, too weak.

-G. Mosteller and J. Tukey, 1977

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Blalock, H. (1961) Causal Inferences in Nonexperimental Research. University of North Carolina Press, Chapel Hill, NC.

    Google Scholar 

  2. Cooper, G. and Herskovitz, E. (1991) “A Bayesian Method for the Induction of Probabilistic Networks from Data,” Journal of Machine Learning 9, 309 - 347.

    Google Scholar 

  3. Glymour, C., Scheines, R., Spirtes, P., and Kelly, K. (1987) Discovering Causal Structure. Academic Press, San Diego, CA.

    MATH  Google Scholar 

  4. Holland, P. (1986) “Statistics and Causal Inference,” JASA 81, 945–960.

    MathSciNet  MATH  Google Scholar 

  5. Kiiveri, H. and Speed, T. (1982) “ Structural Analysis of Multivariate Data: A Review,” in Sociological Methodology, S. Leinhardt, ed. Jossey-Bass, San Francisco.

    Google Scholar 

  6. Kiiveri, H., Speed, T., and Carlin, J. (1984) “Recursive Causal Models,” Journal of the Australian Mathematical Society 36, 30–52.

    Article  MathSciNet  MATH  Google Scholar 

  7. Lauritzen, S., and Wermuth, N. (1984) “Graphical Models for Associations Between Variables, Some of Which are Quantitative and Some of which are Qualitative,” Ann. Stat. 17 31–57.

    Article  MathSciNet  Google Scholar 

  8. Mosteller, F., and Tukey, J. (1977) Data Analysis and Regression, A Second Course in Regression. Addison-Wesley, Massachusetts.

    Google Scholar 

  9. Neyman, J. (1935) “Statistical Problems with Agricultural Experimentation,” J. Roy. Stat. Soc. Suppl. 2 107–180.

    Article  Google Scholar 

  10. Pearl, J. (1988) Probabilistic Reasoning in Intelligent Systems. Morgan Kaufman, San Mateo, CA.

    Google Scholar 

  11. Pratt, J. and Schlaifer, R. (1988) “On the Interpretation and Observation of Laws,” Journal of Econometrics 39, 23–52.

    Article  MathSciNet  Google Scholar 

  12. Rubin, D. (1977) “Assignment to Treatment Group on the Basis of a Covariate,” Journal of Educational Statistics 2, 1–26.

    Google Scholar 

  13. Rubin, D. (1978) “Bayesian Inference for Causal Effects: The Role of Randomizations” Ann. Stat. 6, 34–58.

    Article  MATH  Google Scholar 

  14. Simon, H. (1977) Models of Discovery. D. Reidel, Dordrecht, Holland.

    MATH  Google Scholar 

  15. Spirtes, P., Glymour, C., and Scheines, R. (1993) Causation, Prediction and Search. Springer-Verlag, Lecture Notes in Statistics, New York.

    MATH  Google Scholar 

  16. Verma, T. and Pearl, J. (1990) “Equivalence and Synthesis of Causal Models,” in Proc. Sixth Conference on Uncertainty in Al. Association for Uncertainty in AI, Inc., Mountain View, CA.

    Google Scholar 

  17. Wermuth, N. and Lauritzen, S. (1983) “Graphical and Recursive Models for Contingency Tables,” Biometrika 72, 537–552.

    Article  MathSciNet  Google Scholar 

  18. Wright, S. (1934) “The method of Path Coefficients,” Ann. Math. Stat. 5, 161–215.

    Article  MATH  Google Scholar 

  19. Yule, G. and Kendall, M. (1937) An Introduction to the Theory of Statistics. Charles Griffin, London.

    MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1994 Springer-Verlag New York, Inc.

About this paper

Cite this paper

Spirtes, P., Glymour, C. (1994). Inference, Intervention, and Prediction. In: Cheeseman, P., Oldford, R.W. (eds) Selecting Models from Data. Lecture Notes in Statistics, vol 89. Springer, New York, NY. https://doi.org/10.1007/978-1-4612-2660-4_22

Download citation

  • DOI: https://doi.org/10.1007/978-1-4612-2660-4_22

  • Publisher Name: Springer, New York, NY

  • Print ISBN: 978-0-387-94281-0

  • Online ISBN: 978-1-4612-2660-4

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics