Professional Documents
Culture Documents
Source Coding - TELECOMMUNICATION SYSTEMS
Source Coding - TELECOMMUNICATION SYSTEMS
Source Coding - TELECOMMUNICATION SYSTEMS
pk P ( X xk ),
1 k L
where
k 1
P( X xi | Y y j ) P( xi | y j )
P( X xi ) P( xi )
I ( xi ; y j ) log
P ( xi | y j )
P ( xi )
(2-1)
1
I ( xi ; y j ) log
log P( xi )
P( xi )
(2-2)
1
I ( xi ) log
log P ( xi )
P ( xi )
(2-3)
4
I ( X ; Y ) P ( xi , y j ) I ( xi ; y j )
i 1 j 1
n
P ( xi , y j )
P ( xi , y j ) log
i 1 j 1
(2-4)
P ( xi ) P ( y j )
H ( X ) P ( xi ) I ( xi )
i 1
P ( xi ) log P ( xi )
i 1
(2-5)
5
(2-6)
H ( X ) p( x) log p( x)dx
x
H ( X | Y )
x x
p( x, y ) log p( x | y )dxdy
(2-7)
(2-8)
I ( X ;Y ) H ( X ) H ( X | Y )
or alternatively, as:
I ( X ; Y ) H (Y ) H (Y | X )
6
P( y | xi ) P( xi )
I ( xi ; y ) log
p( y ) P( xi )
p( y | xi )
log
p( y )
(2-9)
p ( y | xi )
I ( X ; Y ) p ( y | xi ) P( xi ) log
dy
p( y)
i 1
(2-10)
7
Example
Suppose that X is a discrete random variable with two equally
probable outcomes x1 = A and x2 = -A . Let the conditional pdfs
p(y|xi), i = 1, 2 be Gaussian with mean
xi and variance 2. That is,
2
( y A)
1
2
p ( y | A)
e 2
(2-11)
2
( y A) 2
1
2
p ( y | A)
e 2
(2-11)
2
The average mutual information obtained from (2-10) becomes:
I ( X ;Y )
1 x
p( y | A)
p( y | A)
p
(
y
|
A
)
log
p
(
y
|
A
)
log
dy (2-12)
x
2
p( y)
p( y)
1
p( y | A) p( y | A)
(2-13)
2
the average mutual information I(X;Y) given by (2-12) represents
the channel capacity of a binary-input additive white Gaussian
noise channel.
p( y)
R = log2L
(3-1)
R = log2L + 1
(3-2)
where x denotes the largest integer less than x. The code rate R
in bits per symbol is now R and, since H(X) log2L , it follows that
R H(X).
N
R H (X )
J
(3-3)
R H (X )
(3-4)
10
nk
(3-5)
k 1
H ( X ) R H ( X ) 1
(3-6)
12
An example
of variable
lengh source
coding
for a DMS
13
14
Lemple-Ziv Algorithm
The Huffman coding algorithm yields optimal source codes
in the sense that code words satisfy the prefix condition
and the average code length is minimum but needed to
know the probabilities of occurrence of all the source letters.
Lempel-Ziv source coding algorithm s designed to be
independent of the source statistic.
Lempel-Ziv belongs to the class of universal source coding
algorithm
It is a variable-to-fixed length algorithm.
Lempel-Ziv algorithm is widely used in the compression of
computer files
15
Lemple-Ziv Algorithm
The sequence of the discrete source is parsed into
variable-length blocks, called phrases. A new phrase is
introduced every time a block of letters from the source
differs from previous phrase in the last letter. The
phrases are listed in a dictionary, which stores the
location of the existing phrases. In encoding a new
phrase, we simply specify the location of the existing
phrase in the dictionary and append the new letter.
Consider the binary sequence:
10101101001001110101000011001110101100011011
Dictionary
For Lempel-Ziv
Algorithm
17
LEMPEL-ZIV DECODER
The decoder is just as simple as the encoder.
Specially, it uses the pointer to identify the
root subsequence and then appends the
innovation symbol. Consider, for example, the
binary encoded block 01010 in position 9. The
last bit, 0, is the innovation symbol. The
remain bits, 0101, point to the root
subsequence 01 in position 5. Hence, the
block 01010 is decoded into 010, which is
correct.
18
d (x , ~
x ) (x ~
x )
1
~
d ( X n , X n ) d ( xk , ~
xk )
n k 1
(4-3)
20
1 n
~
D E[d ( X n , X n )] E[d ( xk , ~
xk )] E[d ( x, ~
x )]
n k 1
(4-4)
where the last step follows from the assumption that the source
output process is stationary.
Now suppose we have a memoryless source with a continuousamplitude output~ X that has a pdf p(x), a quantized amplitude
~
X
d
(
x
,
x) ,
output alphabet
, and a per letter distortion measure
~
where x X and ~
x X . Then, the minimum rate in bits per source
output that is required to represent the output X of the
memoryless source with a distortion less than or equal to D is
called the rate-distortion function R(D) and is defined as:
~
(4-5)
R( D)
min
I(X , X )
~
p(~
x | x ):E [ d ( X , X )] D
21
log 2
Rg ( D )
0
1
2
x2
D
(0 D x2 )
2
x
(D )
(4-6)
23
D f (~
x x) p( x)dx
(4-7)
~ 1
For a uniform quantizer, the output levels are specifies xk 2 (2k 1)
as corresponding to an input signal amplitude in the range (k1)
x k, where is the step size. L is number of levels. When the
uniform quantizer is symmetric with an even number of levels, the
average distortion in (4-7) may be expressed as:
L 1
2
D 2
k 1
( k 1)
f ( 2 k 1) x p ( x ) dx
2
2 L
f ( 2 k 1) x p ( x ) dx
2 1 2
(4-8)
25
X Q( X )
~
X is the output of the vector quantizer when input-vector is X.
In general, quantization of the n-dimensional vector X into an ndimensional vector X introduces a quantization error or a distortion
~
d ( X , X ). The average distortion over the set of input vectors X is:
L
~
D P( X Ck ) E[ d ( X , X ) | X Ck ]
k 1
L
P( X Ck )
k 1
~
d
(
X
,
X
) p( X ) dX
X Ck
27
Example: Let x1 and x2 be two random variables with a uniform joint pdf
1 / ab
p ( x1 , x2 ) p ( X )
0
( X C)
otherwise
Where C is the rectangular region illustrated in the figure. Note that the
rectangle is rotated 450 relative to the horizontal axis. If we quantize x1
and x2 by using uniform interval of length D, the number of levels is:
L1 L2
a b
, hence then number needed for coding the vector X x1 , x2
2
is : Rx R1 R2 log 2 L1 log 2 L2
2
a b
22
each component is equivalent to vector quantizati on with the total levels
a b 2 ,
2ab
example a 4b, the difference i bits rate is :1.64bits / vector , thus vector
quantizati on is 0.82bits / sample better for the same distortion
Differential PCM
The predicted value of xn
And the MSE
(b) Decoder