Comm... System CH2-Lec1
Comm... System CH2-Lec1
(EEEg4172)
Chapter 2
Information Theory and Coding
1
2.1 Information Theory Basics
Outline:
▪ Introduction
▪ Information Source
▪ Measure of Information
▪ Discrete Memoryless Channels
▪ Mutual Information
▪ Channel Capacity
2
Introduction
3
Information Source
4
Information Source…
▪ Information sources can be classified as:
➢ Sources with memory, or
➢ Memoryless sources.
▪ A source with memory is one for which a current symbol depends on the
previous symbols.
▪ A memoryless source is one for which each symbol produced is
independent of the previous symbols.
▪ A discrete memoryless source (DMS) can be characterized by the set of it’s
symbols, the probability assignment to the symbols, and the rate by which
the source generates symbols.
5
Measure of Information
6
Measure of Information…
7
Measure of Information…
i. I ( xi ) = 0 for P( xi ) = 1
ii. I ( xi ) 0
iii. I ( xi ) I ( x j ) if P ( xi ) P ( x j )
iv. I ( xi , x j ) = I ( xi ) + I ( x j )
if xi and x j are independent
8
Measure of Information…
Average Information or Entropy of a DMS:
▪ In a practical communication system, we usually transmit long sequences
of symbols from an information source.
▪ Thus, we are more interested in the average information that a source
produces than the information content of a single symbol.
▪ The mean value of I(xi) over the alphabet of source X with m different
symbols is given by:
m
H ( X ) = E[ I ( xi )] = P( xi ) I ( xi )
i =1
m
= − P( xi ) log2
P ( xi )
bits/symbol
i =1
9
Measure of Information…
Average Information or Entropy of a DMS…
▪ The quantity H(X) is called the entropy of the source X.
▪ It a measure of the average information content per symbol for a DMS X.
▪ The source entropy H(X) can be considered the average amount of
uncertainty within source X that is resolved by use of the alphabet.
▪ The source entropy H(X) satisfies the following relations:
0 H ( X ) log2
m
where m is the size of the
alphabet of source X .
▪ The lower bound corresponds to no uncertainty, which occurs when one symbol has
probability p(xi)=1 while p(xj)= 0 for j≠i, so X emits xi at all times. The upper bound
corresponds to the maximum uncertainty which occurs when p(xj)= 1/m for all i (i.e: when
all symbols have equal probability to be generated by X.
10
Measure of Information…
Average Information or Entropy of a DMS…
Information Rate:
▪ If the time rate at which the source X emits symbols is r (symbols/sec), the
information rate R of the source is given by:
𝑅 = 𝑟𝐻 𝑋 𝑏/𝑠
11
Discrete Memoryless Channels
Channel Representation:
▪ A communication channel is the path or medium through which the
symbols flow from a source to a receiver.
▪ A discrete memoryless channel (DMC) is a statistical model with an input
X and output Y as shown in the figure below.
12
Discrete Memoryless Channels…
Channel Representation…
▪ The channel is discrete when the alphabets of X and Y are both finite.
▪ It is memoryless when the current output depends on only the current input
and not on any of the previous inputs.
▪ In the DMC shown above, the input X consists of input symbols x1, x2, …,
xm and the output Y consists of output symbols y1, y2, …, yn.
▪ Each possible input-to-output path is indicated along with a conditional
probability P(yj/xi), which is known as channel transition probability.
13
Discrete Memoryless Channels…
Channel Matrix:
▪ A channel is completely specified by the complete set of transition
probabilities.
▪ The channel matrix , denoted by [P(Y/X)], is given by:
▪ Since each input to the channel results in some output, each row of the
channel matrix must sum to unity, i.e:
14
Discrete Memoryless Channels…
Special Channels:
1. Lossless Channel
▪ A channel described by a channel matrix with only one non-zero element in
each column is called a lossless channel.
▪ An example of a lossless channel is shown in the figure below.
15
Discrete Memoryless Channels…
Special Channels:
2. Deterministic Channel
▪ A channel described by a channel matrix with only one non-zero unity
element in each row is called a deterministic channel.
▪ An example of a deterministic channel is shown in the figure below.
16
Discrete Memoryless Channels…
Special Channels:
3. Noiseless Channel
▪ A channel is called noiseless if it is both lossless and deterministic.
▪ A noiseless channel is shown in the figure below.
▪ The channel matrix has only one element in each in each row and each
column and this element is unity.
17
Discrete Memoryless Channels…
Special Channels:
4. Binary Symmetric Channel
▪ The binary symmetric channel (BSC) is defined by the channel matrix and
channel diagram given below.
18
Mutual Information
Conditional and Joint Entropies:
▪ Using the input probabilities P(xi), output probabilities P(yj), transition
probabilities P(yj/xi), and joint probabilities P(xi, yj), we can define the
following various entropy functions for a channel with m inputs and n
outputs.
19
Mutual Information…
Conditional and Joint Entropies:
▪ The above entropies can be interpreted as the average uncertainties of the
inputs and outputs.
▪ Two useful relationships among the above various entropies are:
20
Mutual Information…
Mutual Information of a Channel:
21
Channel Capacity
22
Channel Capacity…
23
Channel Capacity…
24
Additive White Gaussian Noise(AWGN) Channel
Differential Entropy:
▪ The average amount of information per sample value of x(t) is measured by
25
Additive White Gaussian Noise(AWGN) Channel…
26
Additive White Gaussian Noise(AWGN) Channel…
▪ In an additive white Gaussian noise (AWGN ) channel, the channel output Y is given
by:
Y= X + n
▪ where X is the channel input and n is an additive band-limited white Gaussian noise
with zero mean and variance σ2.
▪ The capacity Cs of an AWGN channel is given by:
27
Additive White Gaussian Noise(AWGN) Channel…
▪ If the channel bandwidth B Hz is fixed, then the output y(t) is also a band-limited
signal completely characterized by its periodic sample values taken at the Nyquist
rate 2B samples/s.
▪ Then the capacity C (b/s) of the AWGN channel is given by:
28
Examples on Information Theory and Coding
Example-1:
a. Calculate H ( X )
b. Find the amount of information contained in the messages
x1 x2 x3 x4 and x4 x3 x3 x2
29
Examples on Information Theory and Coding Cont’d…
Solution:
4
a. H ( X ) = − P( xi ) log2
[ P ( xi )]
i =1
= 1.85 bits/symbol
0.0024
b. P( x1 x2 x3 x4 ) = (0.4)(0.3)(0.2)(0.1) = 0.0096
I ( x1 x2 x3 x4 ) = − log2 = 6.7 bits/symbol
0.0096
8.7 bits/symbol
Similarly,
P( x1 x2 x3 x4 ) = (0.1)(0.2) 2 (0.3) = 0.0012
I ( x1 x2 x3 x4 ) = − log2 = 9.7 bits/symbol
0.0012
30
Examples on Information Theory and Coding Cont’d….
Example-2:
Consider a binary symmetric channel shown below.
b. P (Y ) = P ( X )P (Y / X )
0.9 0.1
= 0.5 0.5
0.2 0 . 8
= 0.55 0.45 = P ( y1 ) P ( y 2 )
P ( y1 ) = 0.55 and P ( y 2 ) = 0.45
32
Examples on Information Theory and Coding Cont’d….
Solution:
33
Examples on Information Theory and Coding Cont’d….
Example-3:
An information source can be modeled as a bandlimited process with a
bandwidth of 6kHz. This process is sampled at a rate higher than the
Nyquist rate to provide a guard band of 2kHz. It is observed that the
resulting samples take values in the set {-4, -3, -1, 2, 4, 7} with
probabilities 0.2, 0.1, 0.15, 0.05, 0.3 and 0.2 respectively. What is the
entropy of the discrete-time source in bits/sample? What is the entropy in
bits/sec?
34
Examples on Information Theory and Coding Cont’d….
Solution:
The entropy of the source is
6
H ( X ) = − P( xi ) log2 = 2.4087 bits/sample
P ( xi )
i =1
35
References:
▪ Simon Haykin, Communication Systems, 4th edition. (chapter-9).
▪ B.P. Lathi, Modern Digital and Analog Communication Systems, 3rd
edition. (Chapter-15)
▪ Henk C.A van Tilborg, Coding Theory, April 03, 1993.
Additional Reading:
▪ https://en.wikipedia.org/wiki/Information_theory
▪ https://en.wikipedia.org/wiki/Coding_theory
▪ https://en.wikipedia.org/wiki/Noisy-channel_coding_theorem
36