Definition
In the human experience, information typically flows from one place to another. By contrast, the notion of mutual information introduced by Shannon (1948) is perfectly symmetric in its arguments and does not distinguish between “input” and “output.” In this sense, it is perhaps surprising that this very measure of information indeed captures the capacity of any communication channel – though we should recall that the proof of this fundamental fact is not merely a simple consequence of Shannon’s definition.
In spite of Shannon’s strong and fundamental results, it has been tempting to define a notion of directed information. This was first proposed in (Marko 1973) for stationary processes. The more general and useful definition was given in the brief and beautiful note by Massey (1990). Moreover, Massey (1990), Kramer (1998) and subsequent work revealed that directed information has a natural place in the study of information transmission with feedback from the output to the...
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Granger CWJ (1969) Investigating causal relations by econometric models and cross-spectral methods. Econometrica 37(3):424438
Kramer G (1998) Directed information for channels with feedback, vol 11, ETH series in information processing. HartungGorre, Konstanz
Marko H (1973) The bidirectional communication theory a generalization of information theory. IEEE Trans Commun 21:1345–1351
Massey JL (1990) Causality, feedback and directed information. In: Proceedings of the 1990 international symposium on information theory and its applications, Hawaii, pp 303–305
Massey JL, Massey PC (2005) Conservation of mutual and directed information. In: Proceedings of the 2005 international symposium on information theory, Adelaide, pp 157–158
Quinn C, Coleman TP, Kiyavash N, Hatsopoulos NG (2011) Estimating the directed information to infer causal relationships in ensemble neural spike train recordings. J Comput Neurosci. doi:10.1007/s10827-010-0247-2
Schreiber T (2000) Measuring information transfer. Phys Rev Lett 85:461–464
Shannon CE (1948) A mathematical theory of communication. Bell Syst Tech J 27:379–423, and 623–656
So K, Koralek AC, Ganguly K, Gastpar MC, Carmena JM (2012) Assessing functional connectivity of neural ensembles using directed information. J Neural Eng 9:026004
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer Science+Business Media New York
About this entry
Cite this entry
Gastpar, M.C. (2015). Directed Information Flow and Causality in Neural Systems. In: Jaeger, D., Jung, R. (eds) Encyclopedia of Computational Neuroscience. Springer, New York, NY. https://doi.org/10.1007/978-1-4614-6675-8_141
Download citation
DOI: https://doi.org/10.1007/978-1-4614-6675-8_141
Published:
Publisher Name: Springer, New York, NY
Print ISBN: 978-1-4614-6674-1
Online ISBN: 978-1-4614-6675-8
eBook Packages: Biomedical and Life SciencesReference Module Biomedical and Life Sciences