Communication Systems Channel Capacity
Communication Systems Channel Capacity
' $ ' $
S-72.2410 Channel Capacity (1) 2 S-72.2410 Channel Capacity (1) 4
Discrete Channels
& % & %
c Patric Östergård
c Patric Östergård
' $ ' $
S-72.2410 Channel Capacity (1) 6 S-72.2410 Channel Capacity (1) 8
& % & %
c Patric Östergård
c Patric Östergård
' $ ' $
S-72.2410 Channel Capacity (1) 9 S-72.2410 Channel Capacity (1) 11
' $ ' $
S-72.2410 Channel Capacity (1) 10 S-72.2410 Channel Capacity (1) 12
A binary channel where bits are lost (rather than corrupted) is called Consider the channel with the transmission matrix
the binary erasure channel and is shown in [Cov, Fig. 8.6]. The
receiver knows which bits have been erased. The mutual information 0.3 0.2 0.5
is now p(y|x) =
0.5 0.3 0.2 ,
I(X; Y ) = H(Y ) − H(Y |X) = H(Y ) − H(α, 1 − α). 0.2 0.5 0.3
Clearly, H(Y ) ≤ log 3, but this bound cannot be achieved. With where the entry in the xth row and the yth column denotes the
Pr(X = 1) = a, we get Pr(Y = 0) = (1 − a)(1 − α), Pr(Y = e) = α, conditional probability p(y|x) that y is received when x is sent. In
and Pr(Y = 1) = a(1 − α), so (after some calculation) this channel, all the rows of the probability transition matrix are
H(Y ) = H(α, 1 − α) + (1 − α)H(a, 1 − a). permutations of each other and so are the columns. Such a channel is
said to be symmetric.
& % & %
c Patric Östergård
c Patric Östergård
' $ ' $
S-72.2410 Channel Capacity (1) 13 S-72.2410 Channel Capacity (1) 15
& % & %
c Patric Östergård
c Patric Östergård
' $ ' $
S-72.2410 Channel Capacity (1) 14 S-72.2410 Channel Capacity (1) 16
We consider a communication system as shown in [Cov, Fig. 8.8]. A The rate of an (n, M ) code is
code for this channel consists of the following:
log M
1. An index set I = {1, 2, . . . , M }. R= bits per transmission.
n
2. An encoding function X n : I → X n .
3. A decoding function g : Y n → I. A rate R is said to be achievable if there exists a sequence of
We say that a code with these parameters is an (n, M ) code. The (n, d2nR e) (for simplicity, we write (n, 2nR ) in the sequel) codes such
conditional probability of error given that index i was sent is denoted that the maximal probability of error λ(n) tends to 0 as n → ∞.
by λi , that is, The capacity of a discrete memoryless channel is the supremum of
λi = Pr(Ŵ 6= i | W = i). all achievable rates.
& % & %
c Patric Östergård
c Patric Östergård
' $ ' $
S-72.2410 Channel Capacity (1) 18 S-72.2410 Channel Capacity (1) 20
The maximal probability of error, λ(n) , for a channel code is defined as (n)
The set A of jointly typical sequences {(xn , y n )} with respect to
the distribution p(x, y) is the set of n-sequences with empirical
λ(n) = max λi .
i∈I entropies -close to the true entropies, that is,
(n)
1
The average probability of error, Pe , for a channel code is defined as (n) n n n
A = (x , y ) ∈ X × Y : − log p(x ) − H(X) < ,
M
n
1 X
Pe(n) =
λi . 1
M i=1 − log p(y n ) − H(Y ) < ,
n
If the index is chosen uniformly on the set I, we have 1
− log p(xn , y n ) − H(X, Y ) < , ,
n
Pe(n) = Pr(Ŵ 6= W ).
& % & %
c Patric Östergård
c Patric Östergård
' $ ' $
S-72.2410 Channel Capacity (1) 21 S-72.2410 Channel Capacity (1) 23
The jointly typical sequences are illustrated in [Cov, Fig. 8.9]. There
We will decode a channel output Y n as index i if the the codeword
are about 2nH(X) typical X sequences and about 2nH(Y ) typical Y
X n (i) is jointly typical with the received signal Y n . For the joint
sequences. However, since there are only 2nH(X,Y ) jointly typical
AEP (asymptotic equipartition property), we have the following
sequences, not all pairs of typical X and Y sequences are also jointly
result.
typical.
Theorem 8.6.1. Let (X n , Y n ) be sequences of length n drawn i.i.d.
Qn The probability that a any randomly chosen pair is jointly typical is
according to p(xn , y n ) = i=1 p(xi , yi ). Then
about 2−nI(X;Y ) . Hence for a fixed Y n , we can consider about
(n)
1. Pr((X n , Y n ) ∈ A ) → 1 as n → ∞. 2nI(X;Y ) such pairs before we are likely to come across a jointly
(n)
2. |A | ≤ 2n(H(X,Y )+) . typical pair. This suggests that there are at most about 2nI(X;Y )
distinguishable signals X n .
& % & %
c Patric Östergård
c Patric Östergård
' $
S-72.2410 Channel Capacity (1) 22
& %
c Patric Östergård