Professional Documents
Culture Documents
Itc Slides
Itc Slides
Itc Slides
Digital Modulation
Continuous-wave(CW) modulation (recap):
A parameter of a sinusoidal carrier wave is varied
continuously in accordance with the message signal.
Amplitude
Frequency
Phase
Digital Modulation:
Pulse Modulation: Analog pulse modulation: A periodic
pulse train isused as a carrier. The following parameters of
the pulse are modified in accordance with the message
signal. Signal is transmitted at discrete intervals of time.
Pulse amplitude
Pulse width
& %
Pulse duration
' $
& %
' $
& %
' $
s(t)
m(t)
T t
Ts
+
X
s(t) = m(nTs )h(t nTs )
n=
= 1, 0tT
h(t) = = 12 , t = 0, t = T
= 0, otherwise
+
X
m (t) = m(nTs )(t nTs )
& %
n=
' $
Therefore, we get
Z +
m (t) h(t) = m ( )h(t ) d
Z + +
X
= m(nTs )( nTs )h(t ) d
n=
+
X Z +
= m(nTs ) ( nTs )h(t ) d
n=
+
X
s(t) = m (t) h(t) = m(nTs )h(t nTs )
n=
& %
' $
S() = M () H()
s X
= M ( ks )H()
2
k=
j T2
T2
Since, we use flat top samples, H() = T sinc e .
This results in distortion and a delay of T2 . To correct this
1
the magnitude of the equalizer is chosen as T
.
T sinc 2
The message signal m(t) can be recovered from the PAM
signal s(t) as shown in Figure 2.
& %
' $
& %
The other two types of modulation schemes are shown in
' $
Figure 3.
m(t)
t
0
(a)
t
(b)
(c) t
(d) t
& %
' $
Uniform Quantisers
Quantiser type: The quantiser characteristic can be of midtread
or midrise quantizer. These two types are shown in Figure 4.
output output
level level
4 4
2 2
4 2 0 2 4 4 2 0 2 4 input
input level
2 2
level
4 4
Q=M V
Analysis of error in uniform quantisation:
The input M has zero mean, and the quantiser assumed to
be symmetric = Quantiser output V and therefore the
quantization error Q also will have a zero mean.
& %
' $
2mmax
= L
where, L is the total number of representation levels.
The quantization error Q will have its sample values
bounded by 2 q 2.
1,
q
2 2
fQ (q) =
0, otherwise
& %
the mean square value.
' $
2
Q = E[Q2 ]
Z + 2
= q 2 fQ (q) dq
2
+
1
Z 2
= q 2 dq
2
2
=
12
Typically, the L-ary number k, denotes the kth
representation level of the quantiser,
It is transmitted to the receiver in binary form.
Let R denote the number of bits per sample used in the
construction of the binary code. We may then write
& %
L = 2R
' $
R = log2 (L)
Hence, we get
2mmax
= 2R
also,
2
Q = 13 m2max 22R
Let P denote the average power of the message signal m(t).
We may then express the output signal to noise ratio of a
uniform quantiser as
P
(SN R)O = 2
Q
3P
= 2 .22R
mmax
& %
Example: Sinusoidal Modulating Signal
' $
& %
' $
Non Uniform Quantisers
The main advantages of using non-uniform quantizer are:
colorblack
1. Protection of weak passages over loud passages.
2. Enable uniform precision over the entire range of the voice
signal.
3. Fewer steps are required in comparison with uniform quantizer.
A nonuniform quantiser is equivalent to passing the baseband
signal through a compressor and then applying the compressed
signal to a uniform quantizer. A particular form of compression law
that is used in practice is - law, which is defined as follows
log(1+|m|)
|v| = log(1+)
& %
where, m and v are the normalized input and output voltages and
' $
is a positive constant.
Another compression law that is used in practice is the so called A
- law defined by
A|m| 1
1+log A , 0 |m| A
|v| =
1+log(A|m|) , 1
|m| 1
1+log A A
& %
' $
law A law
1 1
=100 A=100
A=2
=5
0 1 0 1
|m| |m|
& %
' $
& %
' $
m[k]
d[k] d [k] To channel
quantizer q
sum
^
m [k] sum
q
+
Predictor
m [k]
q
q [k] = m
m q [k] + dq [k]
= m[k] d[k] + dq [k]
= m[k] + q[k]
& %
' $
d q[k] m q[k]
m q[k]
Predictor
& %
' $
Delta Modulation
Delta Modulation uses a first order predictor.
It is a one bit DPCM.
DM quantizer uses only two levels(L = 2)
The signal is oversampled(atleast 4 times the Nyquist rate) to
ensure better correlation among the adjacent samples.
& %
' $
m(t)
staircase
T Approximation
s
m q(t)
t
0
m[n] = m(nTs ), n = 0, 1, 2,
& %
The basic principles of delta modulation is the following set
' $
of discrete-time relations:
e[n] = m[n] mq [n 1]
eq = sgn(e[n])
mq [n] = mq [n 1] + eq [n]
& %
The transmitter and receiver of the DM is shown in Figure
' $
9.
+
m(t) + d (t) Sampler
d(t) q
Comparator
m q (t)
Integrator
Transmitter
d q(t)
m(t)
LPF
Integrator
Receiver
& %
' $
Uncertainity, Information, and Entropy
Probabilistic experiment involves the observation of the output
emitted by a discrete source during every unit of time. The source
output is modeled as a discrete random variable, S,which takes on
symbols form a fixed finite alphabet.
S = s0 , s1 , s2 , , sk1
with probabilities
P (S = sk ) = pk , k = 0, 1, , K 1
We assume that the symbols emitted by teh source during
successive signaling intervals are statistically independent. A source
having the jproperties jist described is called discrete memoryless
source, memoryless in the sense that the symbol emitted at any
time is independent of previous choices.
& %
We can define the amount of information contained in each
' $
symbols.
I(sk ) = log( p1k )
Here, generally use log2 since in digital communications we will be
talking about bits. The above expression also tells us that when
there is more uncertainity(less probability) of the symbol being
occured then it conveys more information. Some properties of
information are summarized here:
1. for certain event i.e, pk = 1 the information it conveys is zero,
I(sk ) = 0.
2. for the events 0 pk 1 the information is always I(sk ) 0.
3. If for two events pk > pi , the information content is always
I(sk ) < I(si ).
4. I(sk si ) = I(sk ) + I(si ) if sk and si are statistically independent.
The amount of information I(sk ) produced by the source during an
& %
' $
arbitrary signalling interval depends on the symbol sk emitted by
the source at that time. Indeed, I(sk ) is a discrete random variable
that takes on the values I(s0 ), I(s1 ), , I(sK1 ) with probabilites
p0 , p1 , , pK1 respectively. The mean of I(sk ) over teh source
alphabet S is given by
H(S) = E[I(sk )]
K1
X
= pk I(sk )
k=0
K1
X 1
= pk log2 ( )
pk
k=0
& %
information content per source symbol.
' $
& %
' $
H(S)
=
L
& %
' $
Information Theory and Source Coding
Scope of Information Theory
1. Determine the irreducible limit below which a signal cannot
be compressed.
2. Deduce the ultimate transmission rate for reliable
communication over a noisy channel.
3. Define Channel Capacity - the intrinsic ability of a channel
to convey information.
The basic setup in Information Theory has:
a source,
a channel and
destination.
The output from source is conveyed through the channel and
& %
received at the destination. The source is a random variable S,
' $
which takes symbols from a finite alphabet i.e.,
S = {s0 , s1 , s2 , , sk1 }
with probabilities
P (S = sk ) = pk where k = 0, 1, 2, , k 1
and
k1
X
pk = 1
k=0
& %
Information: Information is defined as
' $
1
I(sk ) = logbase
pk
& %
' $
Properties of Information
1. Information conveyed by a deterministic event is nothing i.e.,
I(sk ) = 0, f or pk = 1a .
I(sk ) 0 f or pk 1
a Give examples about deterministic events, sun rising in the east as opposed
to that of the occurence of the Tsunami
& %
' $
& %
' $
Properties of Information
Entropy: The Entropy(H(s)) of a source is defined as:
the average information generated by a discrete memoryless
source.
Entropy defines the uncertainity in the source.
K1
X
H(s) = pk I(sk )
k=0
Properties of Entropy:
1. If K is the number of symbols generated by a source, then
its entropy is bounded by
0 H(s) log2 K
& %
' $
2. If the source has an inevitable symbol sk for which pk = 1
then
H(s) = 0
Proof:
H(s) 0
If the source has one symbol sk which has
P (s = sk ) = pk = 1, then according to the properties, all
other probabilities should be zero i.e., pi = 0, i, i 6= k.
& %
Therefore, from the definition of entropy
' $
K1
X 1
H(s) = pk log
pk
k=0
1
= pk log
pk
= 1 log(1)
=0
Therefore H(s) 0
H(s) log2 K
Consider a discrete memoryless source which has two
different probability distributions
p0 , p1 , , pk
& %
' $
and
q0 , q1 , , qk
on the alphabet
S = s0 , s1 , , sk
We begin with the equation
K1
1 K1
X qk X qk
pk log = pk loge (1)
pk loge 2 pk
k=0 k=0
loge x x 1 (2)
& %
' $
K1
1
X 1
Suppose, qk = K, k then, qk log = log2 K So,
qk
k=0
& %
' $
in Eq. 3 becomes
K1
X K1
X
pk log qk pk log pk 0
k=0 k=0
K1
X K1
X
= pk log K pk log pk
k=0 k=0
K1 K1
X X 1
= pk log K pk log
pk
k=0 k=0
K1 K1
X X 1
= log K pk pk log
pk
k=0 k=0
K1
X 1
= pk log log2 K
pk
k=0
& %
' $
H(s)
0 1/2 1
p
& %
' $
Channel Coding
Channel Coding is done to ensure that the signal transmitted is
recovered with very low probability of error at the destination.
x y0
0
x y
1 1
Source, X Destination, Y
X Channel Y
x y
J1 K1
P (Y = yk |X = xj ) = p(yk |xj ), j, k
such that
K1
X
p(yk |xj ) = 1, j.
k=0
p(xj , yk ) = P (X = xj , Y = yk )
= P (Y = yk |X = xj )P (X = xj )
= p(yk |xj )p(xj )
& %
' $
& %
' $
1p
x =0 y =0
0 0
x =1 y =1
1 1
1p
Mutual Information
K1
X
H(X |Y) = H(X |Y = yk )p(yk ) (1)
k=0
K1
" #
X J1
X 1
= p(xj |yk )p(yk )log2 (2)
p(xj |yk )
k=0 j=0
K1
" #
X J1
X 1
= p(xj , yk )log2 (3)
p(xj |yk )
k=0 j=0
& %
termed as Mutual Information of the channel and is denoted by
' $
I(X ; Y).
Similarly,
Property 1:
The mutual information of a channel is symmetric, that is
J1
" #
X 1
H(X ) = p(xj ) log2 (7)
j=0
p(xj )
J1
" # K1
X 1 X
= p(xj ) log2 p(yk |xj ) (8)
j=0
p(xj )
k=0
J1 K1
" #
XX 1
= p(yk |xj )p(xj ) log2 (9)
j=0 k=0
p(xj )
J1 K1
" #
XX 1
= p(xj , yk ) log2 (10)
j=0
p(xj )
k=0
& %
we obtain
' $
J1
" #
X K1
X p(xj |yk )
I(X ; Y) = p(xj , yk ) log2 (11)
j=0 k=0
p(xj )
J1
" #
X K1
X p(yk |xj )
I(X ; Y) = p(xj , yk ) log2 = I(Y; X ) (13)
j=0 k=0
p(yk )
Property 2:
& %
The mutual is always non-negative, that is
' $
I(X ; Y) 0
Proof:
We know,
p(xj , yk )
p(xj |yk ) = (14)
p(yk )
Substituting Eq. 14 in Eq. 13, we get
J1
" #
X K1
X p(xj , yk )
I(X ; Y) = p(xj , yk ) log2 = I(Y; X )
j=0 k=0
p(xj )p(yk )
(15)
Using the following fundamental inequality which we
& %
derived discussing the properties of Entropy,
' $
K1
!
X qk
pk log2 0
pk
k=0
I(X ; Y) 0
Property 3:
& %
I(X ; Y) = H(X ) + H(Y) H(X , Y)
' $
where, the joint entropy H(X , Y) is defined as
J1
!
X K1
X 1
H(X , Y) = p(xj , yk ) log2
j=0 k=0
p(xj , yk )
Proof:
J1
!
X K1
X p(xj )p(yk )
H(X , Y) = p(xj , yk ) log2 + (16)
j=0 k=0
p(xj , yk )
J1
!
X K1
X 1
p(xj , yk ) log2 (17)
j=0 k=0
p(xj )p(yk )
J1
!
X K1
X 1
= I(X ; Y) + p(xj , yk ) log2 (18)
j=0 k=0
p(xj )p(yk )
& %
But,
' $
J1
!
X K1
X 1
p(xj , yk ) log2 (19)
j=0 k=0
p(xj )p(yk )
J1
! K1
X 1 X
= log2 p(xj , yk )+ (20)
j=0
p(xj )
k=0
K1
! J1
X 1 X
log2 p(xj , yk ) (21)
p(yk ) j=0
k=0
J1
! K1 !
X 1 X 1
= p(xj ) log2 + p(yk ) log2 (22)
j=0
p(xj ) p(yk )
k=0
= H(X ) + H(Y) (23)
Problems
The following problems may be given as exercises.
1. Show that the mutual information is zero for a deterministic
channel.
2. Prove that I(X ; Y) = min(H(Y), H(X ))
3. Prove that I(X ; Y) = min(log(|Y|), log(|X |))
& %
' $
Source Coding
1. Source symbols encoded in binary
2. The average codelength must be reduced
3. Remove redundancy reduces bit-rate
Consider a discrete memoryless source on the alphabet
S = {s0 , s1 , , sk }
Let the corresponding probabilities be
{p0 , p1 , , pk }
and codelengths be
{l0 , l1 , , lk }.
Then, the average codelength(average number of bits per symbol)
& %
of the source is defined as
' $
K1
=
X
L pk ll
k=0
Lmin
=
L
For an efficient code approaches unity.
The question: What is smallest average codelength that is possible?
The Answer: Shannons source coding theorem
Given a discrete memoryless source of entropy H(s), the average
codeword length L for any distortionless source encoding scheme is
& %
bounded by
' $
H(s)
L
Since, H(s) is the fundamental limit on the average number of
bits/symbol, we can say
Lmin H(s)
H(s)
= =
L
Data Compaction:
1. Removal of redundant information prior to transmission.
2. Lossless data compaction no information is lost.
3. A source code which represents the output of a discrete
memoryless source should be uniquely decodable.
& %
' $
& %
' $
& %
by
' $
K1
X
2lk 1
k=0
s
Initial State 1
0
1 s
2
0
1
1
s
3
& %
can be constructed with an average code-word length l, which is
' $
bounded as follows:
The left hand side of the above equation, the equality is satisfied
owing to the condition that, any symbol sk is emitted with the
probability
pk = 2lk (2)
& %
' $
K1
X K1
X
2lk = pk = 1 (3)
k=0 k=0
K1
lk
=
X
L (4)
2lk
k=0
& %
' $
K1
!
X 1
H(s) = log2 (2lk )
2lk
k=0
(5)
K1
X lk
=
2lk
k=0
& %
' $
Huffman Coding
1. Huffman code is a prefix code
2. The length of codeword for each symbol is roughly equal to the
amount of information conveyed.
3. The code need not be unique (see Figure 3)
A Huffman tree is constructed as shown in Figure. 3, (a) and (b)
represents two forms of Huffman trees. We see that both schemes
have same average length but different variances.
Variance is a measure of the variability in codeword lengths of a
source code. It is defined as follows:
K1
2
X
2 = pk (lk L) (6)
k=0
& %
' $
& %
' $
s0 s1 s2 s3 s4
Symbol code
0 1 0 s0 10
(a) 1
s1 00
Average length, L = .2.2
Variance = 0.160 0 1 s2 01
s3 110
0 s4 111
1
s0 s1 s2 s3 s4 Symbol code
0 1
0 s 0
(b) 0
1
Average length, L = .2.2 0 s1 10
0
Variance = 1.036 1
s2 110
1
(b) s3 1110
& %
s4 1111
Drawbacks:
& %
' $
Lempel-Ziv Coding
1. Overcomes the drawbacks of Huffman coding
2. It is an adaptive and simple encoding scheme.
3. When applied to English text it achieves 55% in contrast to
Huffman coding which achieves only 43%.
& %
' $
000101110010100101.........
We assume that 0 and 1 are known and stored in codebook
subsequences stored : 0, 1
Data to be parsed: 000101110010100101.........
The shortest subsequence of the data stream encountered for the first time and not seen before
is 00
subsequences stored: 0, 1, 00
Data to be parsed: 0101110010100101.........
The second shortest subsequence not seen before is 01; accordingly, we go on to write
We continue in the manner described here until the given data stream has been completely parsed. The
code book is shown below:
Numerical positions: 1 2 3 4 5 6 7 8 9
subsequences: 0 1 00 01 011 10 010 100 101
Numerical Repre 11 12 42 21 41 61 62
sentations:
Binary encoded
blocks: 0010 0011 1001 0100 1000 1100 1101
Example:
For, the binary symmetric channel discussed previously, I(X ; Y
will be maximum when p(x0 ) = p(x1 ) = 12 . So, we have
& %
' $
Since, we know
C = 1 + p log2 p + (1 p) log2 (1 p)
(5)
= 1 H(p)
& %
' $
& %
' $
H(S) C
(6)
Ts Tc
There exists a coding scheme for which the source output
can be transmitted over the channel and be reconstructed
with an arbitrarily small probability of error. The parameter
C
is called critical rate.
Tc
4. Conversly, if
H(S) C
> (7)
Ts Tc
it is not possible to transmit information over the channel
and reconstruct it with an arbitrarily small probability of
error.
& %
Example:
' $
1 C
(8)
Ts Tc
Tc
But the ratio equals the code rate, r of the channel encoder.
Ts
Hence, for a binary symmetric channel, if r C, then there
exists a code capable of achieving an arbitrarily low probability
of error.
& %
' $
Information Capacity Theorem:
The Information Capacity Theorem is defined as
The information capacity of a continuous channel of bandwidth B
hertz, perturbed by additive white Gaussian noise of power spectral
N0
density and limited in bandwidth to B, is given by
2
!
P
C = B log2 1+ (9)
N0 B
K = 2BT (10)
Yk = Xk + Nk (11)
& %
' $
The noise sample Nk is Gaussian with zero mean and variance
given by
2 = N0 B (12)
E[Xk2 ] = P (13)
& %
' $
The mutual information I(Xk ; Yk ) can be expressed as
& %
Hence, the differential entropy of Yk is
' $
1
h(Yk ) = log2 [2e(P + 2 )] (17)
2
2. The variance of the noise sample Nk equals 2 . Hence, the
differential entropy of Nk is given by
1
h(Nk ) = log2 (2e 2 ) (18)
2
3. Now, substituting above two equations into
I(Xk ; Yk ) = h(Yk ) h(Nk ) yeilds
1 P
C = log2 (1 + 2 )bits per transmission (19)
2
With the channel used K times for the transmission of K samples
of the process X(t) in T seconds, we find that the information
& %
capacity per unit time is K/T times the result given above for C.
' $
P
C = Blog2 (1 + )bits per second (20)
N0 B
& %
' $
Error Control Coding
Channel is noisy
Channel output prone to error
we need measure to ensure correctness of the bit stream
transmitted
Error control coding aims at developing methods for coding to
check the correctness of the bit stream transmitted.
The bit stream representation of a symbol is called the codeword of
that symbol.
& %
Convolution Codes
' $
Linear Block Codes
A code is linear if two codes are added using modulo-2 arithmetic
produces a third codeword in the code.
b ,
i i = 0, 1, , n k 1
ci = (1)
mi+kn , i = n k, n k + 1, , n 1
& %
' $
The (n k) parity bits are linear sums of the k message bits.
1, ifb depends onm
i j
pij = (3)
0, otherwise
& %
n = [n0 , n1 , , nn1 ]
' $
p p01 p0,nk1
00
p10 p11 p1,nk1
P=
.. .. .. ..
. . . .
pk1,0 pk1,1 pk1,nk1
& %
' $
c = [b : m]
= c = m[P : Ik ]
G = [P : Ik ]
= c = mG
ci + cj = mi G + mj G
& %
= (mi + mj )G
' $
T
P
HGT = Ink : PT
Ik
= PT + PT
=0
& %
' $
cH T = mGH T
=0
& %
' $
& %
' $
Repetition Codes
This is the simplest of linear block codes. Here, a single message
bit is encoded into a block of n identical bits, producing an (n, 1)
block code. This code allows variable amount of redundancy. It has
only two codewords - all-zero codeword and all-one codeword.
Example
Consider a linear block code which is also a repetition code. Let
k = 1 and n = 5. From the analysis done in linear block codes
G = [1111 : 1] (4)
& %
' $
1 0 0 0 :1
0 1 0 0 : 1
H=
0 0 1 0 : 1
0 0 0 1 :1
& %
' $
Syndrome Decoding
Let c be the original codeword sent and r = c + e be the received
codeword, where e is the error vector.
n
ei = 1, if an error has occured in the ith position0, otherwise
(5)
Decoding of r
A 1 (n k) matrix called error syndrome matrix is calculated
s = rH T (6)
Properties of Syndrome:
& %
' $
1. Syndrome depends only on error.
s = (c + e)H T s = cH T + eH T (7)
Since, cH T is zero,
s = eH T (8)
& %
parity check matrix,
' $
si = (ci + e)H T
ei H T = eH T + ci H T (9)
= eH T
& %
' $
Hamming Distance
Hamming weight, w(c) is defined as the number of nonzero
elements in a codevector.
Hamming distance, d(c1 , c2 ) between two codewords c1 and c2
is defined as the number of bits in which they differ.
Minimum distance, dmin is the minimum hamming distance
between two codewords.
It is of importance in error control coding because, error detection
and correction is possible if
1
t (dmin 1) (10)
2
where, t is the Hamming weight.
& %
' $
& %
' $
c =0 c2 c3 ci c2k
1
e2 c2 + e2 c3 + e2 ci + e2 c2k + e2
e =0 c2 + e3 c3 + e3 ci + e3 c2k + e3
3
.. .. .. .. ..
. . . . .
ej = 0 c2 + ej c3 + ej ci + ej c2k + ej
.. .. .. .. ..
. . . . .
e2nk = 0 c2 + e2nk c3 + e2nk ci + e2nk c2k + e2nk
The 2nk rows of the arrays represents the cosets of the code, and
their first elements e2 , , e2nk are called coset leaders. The
probability of decoding is minimized by choosing the most likely
occuring error pattern as coset leader. For, binary symmetric
channel, the smaller the Hamming weight of an error pattern the
& %
more likely is to occur.
' $
& %
' $
Cyclic Codes
Cyclic property: Any cyclic shift of a codeword in the code is also a
codeword.
Cyclic codes are well suited for error detection. Certain set of
polynomials are chosen for this purpose.
Properties of polynomials:
1. Any polynomial B(x) can be divided by a divisor polynomial
C(x) if B(x) is of higher degree than C(x).
2. Any polynomial B(x) can be divided once by C(x) if both are
of same degree.
3. subtraction uses exclusive OR.
& %
' $
& %
' $
Considering a 3bit CRC, The following process is carried out at encoder and decoder
Remainder is nonzero
& %