Skip to main content

What is Information?

  • Chapter
  • First Online:
  • 1326 Accesses

Abstract

Chapter 2 examines how the most current use of the word ‘information’ can lead to outline an axiomatic definition of information. Its most specific features are that it has no existence unless it is physically inscribed as a sequence of symbols on some medium, which however has no influence on it besides ensuring its existence, and that it can be defined only as an equivalence class, with respect to transformations like alphabet change and coding. It is thus an abstract entity which resides in the physical world. An information meets Barbieri’s concept of ‘nominable entity’, which refers to a singular object. This concept is explicated and illustrated. A natural number can be used, besides its usual meanings of representing a quantity (cardinal number) or a rank in a sequence (ordinal number), as a label uniquely representing a nominable entity. The uniqueness of nominable entities entails that their representatives do not suffer any change and thus must be protected against any perturbation. A short history of communication engineering, which developed the means of such a protection referred to as ‘error-correcting codes’, is briefly presented. It is also stated that the theoretical tools needed in order to deal with communication at a distance can be used as well for communication over time such as biological heredity.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.00
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD   109.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    A symbol is an element of some given finite set of distinct objects, referred to as an alphabet.

  2. 2.

    In the information-theoretic literature. Physicists do not even ask the question whether information is a physical entity but deal with it as such, following Schrödinger and Brillouin.

  3. 3.

    In any possible meaning of the word: there should not be a causal relation between them and, if they are random, they should be mutually independent in the probabilistic meaning of the word, i.e., their joint probability should be the product of their individual probabilities.

  4. 4.

    The progress of mathematics has refuted Zeno’s paradoxes like that of Achilles and the tortoise.

  5. 5.

    It is redundant: fortunately, less than \(10^{13}\) people live in France.

  6. 6.

    An error probability of 1/2 prevents any communication by means of the binary alphabet; see Sect. 5.2.2.

References

  • Arikan, E. (2009). Channel polarization: A method for constructing capacity-achieving codes for symmetric binary-input memoryless channels. IEEE Transactions on Information Theory, 55(7), 3051–3073. doi:10.1109/TIT.2009.2037044. July 2009.

    Article  Google Scholar 

  • Barbieri, M. (2007). Is the cell a semiotic system? In M. Barbieri, (Ed.), Introduction to Biosemiotics (pp. 179–207). Dordrecht: Springer.

    Chapter  Google Scholar 

  • Berrou, C., & Glavieux, A. (1996). Near optimum error-correcting coding and decoding: turbo-codes. IEEE Transactions on Computers, 44(10), 1261–1271.

    Google Scholar 

  • Berrou, C., Glavieux, A., & Thitimajshima, P. (1993). Near Shannon limit error-correcting coding and decoding: turbo-codes. Proceeding of ICC'93, Geneva, Switzerland, pp. 1064–1070.

    Google Scholar 

  • Brillouin, L. (1956). Science and information theory. New York: Academic Press.

    Google Scholar 

  • Brillouin, L. (1959). Vie, matiere et observation. Paris: Albin Michel.

    Google Scholar 

  • Chaitin, G. J. (2005). Meta Math!. New York: Pantheon Books.

    Google Scholar 

  • Gàcs, P., & Vitànyi, P. M. B. (2011). Raymond J. Solomonoff 1926–2009. IEEE Information Theory Society Newsletter, 61(1), 11–16.

    Google Scholar 

  • Gallager, R. G. (1962). Low-density parity-check codes. IRE Transactions on Information Theory, IT-8, 21–28.

    Article  Google Scholar 

  • Gallager, R. G. (1963). Low-density parity-check codes. Cambridge: MIT Press.

    Google Scholar 

  • Guizzo, E. (2004). Closing in on the perfect code. IEEE Spectrum, 41(3 INT), 28–34.

    Article  Google Scholar 

  • Kolmogorov, A. N. (1965). Three approaches to the quantitative definition of information. Problems of Information Transmission, 1,  4–7.

    Google Scholar 

  • Kolmogorov, A. N. (1968). Logical basis for information theory and probability theory. IEEE Transactions on Information Theory, IT-14(5), 662–664.

    Article  Google Scholar 

  • Landauer, R. (1996). The physical nature of information. Physics Letters A, 217, 188–193. (Reprinted in Leff and Rex 2003, pp. 335–340).

    Article  CAS  Google Scholar 

  • Leff, H. S., & Rex, A. F. (2003). Maxwell’s demon 2, 2nd edition. Bristol: IoP.

    Google Scholar 

  • Poincaré, H. (1902). La science et l’hypothèse. Paris: Flammarion.

    Google Scholar 

  • Ricard, M., & Trinh, X. Th. (2002). L’infini dans la paume de la main. Paris: Pocket.

    Google Scholar 

  • Rovelli, C., (2004). Relational quantum theory. In N. Kolenda & A. Elitzur (Eds.), Quo Vadis Quantum Mechanics? Dordrecht: Springer.

    Google Scholar 

  • Shannon, C. E. (1948). A mathematical theory of communication. The Bell System Technical Journal, 27, 379–457,  623–656. (Reprinted in Shannon and Weaver 1949, Sloane and Wyner 1993, pp. 5–83 and in Slepian 1974, pp. 5–29).

    Article  Google Scholar 

  • Shannon, C. E. (1956). The bandwagon. IRE Transactions on Information Theory. (Reprinted in Sloane and Wyner 1993, p. 462).

    Google Scholar 

  • Shannon, C. E., & Weaver, W. (1949). The mathematical theory of communication. Urbana: University of Illinois Press.

    Google Scholar 

  • Slepian, D. (Ed.). (1974). Key papers in the development of information theory. Piscataway: IEEE Press.

    Google Scholar 

  • Sloane, N. J. A., & Wyner, A. D. (Eds.). (1993). Claude Elwood Shannon, collected papers. Piscataway: IEEE Press.

    Book  Google Scholar 

  • Trinh, X. Th. (2011). Un non-début de l’univers? Nouvel Observateur Hors-Série No. 77, L’origine du monde.

    Google Scholar 

  • Yockey, H. P. (1974). An application of information theory to the central Dogma and the sequence hypothesis. Journal of Theoretical Biology, 46, 369–406.

    Article  PubMed  CAS  Google Scholar 

  • Yockey, H. P. (1992). Information theory and molecular biology. Cambridge: Cambridge University Press.

    Google Scholar 

  • Yockey, H. P. (2005). Information theory, evolution, and the origin of life. Cambridge: Cambridge University Press.

    Book  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Gérard Battail .

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer Science+Business Media Dordrecht

About this chapter

Cite this chapter

Battail, G. (2014). What is Information?. In: Information and Life. Springer, Dordrecht. https://doi.org/10.1007/978-94-007-7040-9_2

Download citation

Publish with us

Policies and ethics