Skip to main content

Part of the book series: Springer Monographs in Mathematics ((SMM))

  • 562 Accesses

Abstract

The goal here is to encode a sequence of symbols in such a way that it is possible to decode it perfectly (lossless coding), and to decode it sequentially (prefix coding). One may then relate codes and probabilities: this is the essence of the Kraft-McMillan inequalities. If one aims at minimizing the codeword’s length, Shannon’s entropy gives an intrinsic limit, when the word to be encoded is regarded as a random variable. When the distribution of this random variable is known, then the optimal compression rate can be achieved (Shannon’s coding and Huffman’s coding). Moreover, as codeword lengths are identified with probability distributions, for any probability distribution, one may design a prefix code which encodes sequentially. This will be referred to as “coding according to this distribution”. Arithmetic coding, based on a probability distribution which is not necessarily the one of the source, will be particularly detailed. In this way, the algorithmic aspect of coding and the modeling of the source distribution are separated. Here the word “source” is used as a synonym for a random process. We finally point out some essential tools needed to quantify information, in particular the entropy rate of a process. This rate appears as an intrinsic lower bound for the asymptotic compression rate, for almost every source trajectory, as soon as it is ergodic and stationary. This also shows that is is crucial to encode words in blocks. Arithmetic coding has the advantage of encoding in blocks and “online”. If arithmetic coding is devised with the source distribution, then it asymptotically achieves the optimal compression rate. In the following chapters, we will be interested in the question of adapting the code to an unknown source distribution, which corresponds to a fundamentally statistical question.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

eBook
USD 16.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 119.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 119.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. D. Huffman, A method for the construction of minimum redundancy codes. Proc. IRE 40, 1098–1101 (1952)

    Article  Google Scholar 

  2. P. Algoet, T. Cover, A sandwich proof of the Shannon-McMillan-Breiman theorem. Annals of Prob. 16, 899–909 (1988)

    Article  MathSciNet  Google Scholar 

  3. K. Chung, A note on the ergodic theorem of information theory. Annals of Math. Stat. 32, 612–614 (1961)

    Article  MathSciNet  Google Scholar 

  4. R. Dudley, Real analysis and probability, 2nd edn. (Cambridge University Press, New York, 2002)

    Google Scholar 

  5. T.M. Cover, J.A. Thomas, Elements of Information Theory. Wiley series in telecommunications (Wiley, New York, 1991)

    Google Scholar 

  6. C. Shannon, A mathematical theory of communication. Bell Sys. Tech. J. 27(379–423), 623–656 (1948)

    Article  MathSciNet  Google Scholar 

  7. J. Rissanen, Generelized Kraft inequality and arithmetic coding. IBM J. Res. Devl. 20, 20–198 (1976)

    Article  Google Scholar 

  8. R. Pasco. Source coding algorithms for fast data compression. Ph.D. Thesis, Stanford Univ (1976)

    Google Scholar 

  9. A. Garivier. Codage universel: la méthode arithmétique. Texte de préparation à l’agrégation (2006)

    Google Scholar 

  10. B. McMillan, The basic theorems of information theory. Ann. Math. Stat. 24, 196–219 (1953)

    Article  MathSciNet  Google Scholar 

  11. L. Breiman, The individual ergodic theorem of information theory. Ann. Math. Stat. 28, 809–811 (1957)

    Article  MathSciNet  Google Scholar 

  12. A. Barron, The strong ergodic theorem for densities: generalized Shannon-McMillan-Breiman theorem. Annals Probab. 13, 1292–1303 (1985)

    Article  MathSciNet  Google Scholar 

  13. J. Kieffer, A counter-example to Perez’s generalization of Shannon-McMillan theorem. Annals Probab. 1, 362–364 (1973)

    Article  Google Scholar 

  14. J. Kieffer, Correction to “a counter-example to Perez’s generalization of Shannon-McMillan theorem”. Annals of Probab. 4, 153–154 (1976)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Élisabeth Gassiat .

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer International Publishing AG, part of Springer Nature

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Gassiat, É. (2018). Lossless Coding. In: Universal Coding and Order Identification by Model Selection Methods. Springer Monographs in Mathematics. Springer, Cham. https://doi.org/10.1007/978-3-319-96262-7_1

Download citation

Publish with us

Policies and ethics