The AEP Property of Random Sequences and Applications to Information Theory: Part I Basic Principles
Consider a biased coin with probability of heads p. If p does not equal 0 or 1, there is not much that one can say with certainty about a single toss of the coin. However, for a sequence of N independent tosses, if one chooses N large enough one can make statements about the composition of this sequence which will be true with probability as close to 1 as desired.
KeywordsSide Information Broadcast Channel Common Information Binary Digit Transmitted Vector
Unable to display preview. Download preview PDF.
- 2.Wolf, J.K., “Multiple-user communications,” National Telecommunications Conference Atlanta, Georgia, 1973.Google Scholar
- 3.Wyner, A.D., “Recent results in the Shannon theory,” IEEE Transactions on Information Theory vol. IT-20, 2, January 1974.Google Scholar
- 4.Slepian, D. and Wolf, J.K., “Noiseless coding of correlated information sources,” IEEE Transactions on Information Theory vol. IT-19, 471, July 1973.Google Scholar
- 5.Cover, T.M., and Wolf for Theory vol. “A proof of the data compression theorem of Slepian ergodic sources,” IEEE Transactions on Information IT-21, 226, March 1975.Google Scholar
- 6.Wyner, A.D., “The common information of two dependent random variables,” IEEE Transactions on Information Theory vol. IT-21, 163Google Scholar
- 8.Cover, T.M., “An achievable rate region for the broadcast channel,” IEEE Transactions on Information Theory vol. IT-21, 399, July 1975.Google Scholar