Lecture 5 Continuous-Valued Sources and Channels
I-Hsiang Wang
Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw
November 4, 2016
1 / 63 I-Hsiang Wang IT Lecture 5
Lecture 5 Continuous-Valued Sources and Channels I-Hsiang Wang - - PowerPoint PPT Presentation
Lecture 5 Continuous-Valued Sources and Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 4, 2016 1 / 63 I-Hsiang Wang IT Lecture 5 From Discrete to Continuous So far we have
1 / 63 I-Hsiang Wang IT Lecture 5
2 / 63 I-Hsiang Wang IT Lecture 5
1 First we investigate basic information measures – entropy, mutual information, and KL
2 Then, we introduce differential entropy as a continuous r.v.'s counterpart of Shannon entropy,
3 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Entropy and Mutual Information
4 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Entropy and Mutual Information
Suppose X has the probability density function (p.d.f.) fX (·).
k=−∞ [k∆, (k + 1)∆).
∆
k∆
5 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Entropy and Mutual Information
6 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Entropy and Mutual Information
7 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Entropy and Mutual Information
8 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Entropy and Mutual Information
∞
k=−∞
∞
k=−∞
∞
∞ f (x) log f (x) dx = E
1 f(X)
9 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Entropy and Mutual Information
k,j=−∞ I∆ k × I∆ j , where I∆ k = [k∆, (k + 1)∆).
k × I∆ j such that
1 ∆2
I∆
k ×I∆ j fX,Y (x, y) dx dy.
k × I∆ j , with p.m.f.
10 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Entropy and Mutual Information
k and
j such that
I∆
k fX (x) dx = ∆fX (
I∆
j fY (y) dy = ∆fY (
k,j=−∞ PX,Y (xk, yj) log PX,Y (xk,yj) PX(xk)PY (yj)
k,j=−∞
fX,Y (xk,yj)✟
∆2 fX( xk)fY ( yj)✟
∆2
k,j=−∞ fX,Y (xk, yj) log fX,Y (xk,yj) fX( xk)fY ( yj)
−∞
−∞ fX,Y (x, y) log fX,Y (x,y) fX(x)fY (y) dx dy as ∆ → 0
f(X,Y ) f(X)f(Y )
11 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Entropy and Mutual Information
P,Q
12 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Differential Entropy
13 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Differential Entropy
1 fX(X)
1 fX|Y (X|Y )
fX,Y (X,Y ) fX(X)fY (Y )
14 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Differential Entropy
g(X)
x∈suppf f (x) log f(x) g(x)dx
15 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Differential Entropy
i=1 h
16 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Differential Entropy
1 b−a1 {a ≤ x ≤ b}, its differential entropy
1 √ 2πe− x2
2 , its differential entropy
2 log (2πe).
17 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Differential Entropy
18 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Differential Entropy
19 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Differential Entropy
2 log (2πe)n (det k).
n
i=1
2 log (det k)
2 log (2πe) + 1 2 log (det k) = 1 2 log (2πe)n (det k) .
20 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Differential Entropy
2 log (2πe)n (det k) .
21 / 63 I-Hsiang Wang IT Lecture 5
Measures of Information for Continuous Random Variables Differential Entropy
22 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels
23 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels
24 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels
X: E[b(X)]≤B
Channel Encoder Channel Decoder Channel
X: E[b(X)]≤BI (X ; Y ) .
Channel Encoder Channel Decoder Channel
25 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels
1 Discretization: Discretize the source and channel input/output to create a discrete system, and
2 New typicality: Extend weak typicality for continuous r.v. and repeat the arguments in a similar
26 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels
27 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels
1 First, we formulate the channel coding problem over continuous memoryless channels (CMC),
2 Second, we introduce additive Gaussian noise (AGN) channel, derive the Gaussian channel
28 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Continuous Memoryless Channel
29 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Continuous Memoryless Channel
1 Input/output alphabet X = Y = R. 2 Continuous Memoryless Channel (CMC):
3 Average input cost constraint B:
1 N
k=1 b (xk) ≤ B, where b : R → [0, ∞) is the
30 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Continuous Memoryless Channel
X: E[b(X)]≤B
How to compute h (Y |X ) when X has no density? Recall h (Y |X ) = EX
suppY f (y|X) log f (y|X) dy
, where f (y|x) is the conditional density of Y given X.
31 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Continuous Memoryless Channel
ENC
DEC
32 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Continuous Memoryless Channel
ENC
DEC
33 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Continuous Memoryless Channel
ENC
DEC
New ENC Equivalent DMC
34 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Continuous Memoryless Channel
Equivalent DMC New ENC
DEC
d
d
1 Random codebook generation: Generate the codebook randomly based on the original
2 Choice of discretization: Choose Qin such that the cost constraint will not be violated after
3 Achievability in the equivalent DMC: By the achievability part of the channel coding theorem
4 Achievability in the original CMC: Prove that when the discretization in Qin and Qout gets
35 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
36 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
Channel Encoder Channel Decoder
1 Input/output alphabet X = Y = R. 2 AWGN Channel:
3 Average input power constraint P :
1 N
k=1|xk|2 ≤ P .
37 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
X: E[|X|2]≤P
2 log
σ2
1 √ 2πP e− x2
2P , as shown in the next slide.
38 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
2 log (2πe) σ2 (a)
2 log (2πe)
2 log (2πe) σ2 = 1 2 log
σ2
2 log (2πe) Var [Y ] and Var [Y ] = Var [X] + Var [Z] ≤ P + σ2, since
39 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
p N(P + σ2)
40 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
p N(P + σ2) √ Nσ2
41 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
p N(P + σ2) √ Nσ2
42 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
p N(P + σ2) √ Nσ2
N(P+σ2)
N
√ Nσ2N
N log
N(P+σ2)
N
√ Nσ2N
1 2 log
σ2
2 log
σ2
43 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
√ NP x1 x2
44 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
√ NP x1 αy x2
P P +σ2 (MMSE coeff.)
Nearest Neighbor
45 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
√ NP r N Pσ2 P + σ2 x1 αy x2
P P +σ2 (MMSE coeff.)
Nearest Neighbor
P+σ2
46 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
√ NP r N Pσ2 P + σ2 x1 αy x2
NPσ2/(P+σ2)
N
√ NP
N
σ2 P+σ2
47 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
√ NP r N Pσ2 P + σ2 x1 αy x2
σ2 P+σ2
N ( R+ 1
2 log
(
1 1+ P σ2
))
2 log
σ2
2 log
σ2
48 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
1 Due to CLT, Gaussian well models noise that is the aggregation of many small perturbations. 2 Analytically Gaussian is highly tractable. 3 Consider a input-power-constrained channel with independent additive noise. Within the family
2 log
σ2
49 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
2 log
σ2
2 log
σ2
2 log
σ2
50 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
2 log
log e 2(P+σ2)EY G
2 log
log e 2(P+σ2)EY
51 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
fY G(Y )
fZG(Z)
fY (Y )fZG(Z) fY G(Y )fZ(Z)
fY G(Y )fZ(Z)
(Jensen's Inequality)
fY G(Y )fZ(Z)
52 / 63 I-Hsiang Wang IT Lecture 5
Channel Coding over Continuous Memoryless Channels Gaussian Channel Capacity
fY G(Y )fZ(Z)
(∵ Y = XG + Z)
(∵ Y = XG + Z)
53 / 63 I-Hsiang Wang IT Lecture 5
Lossy Source Coding for Continuous Memoryless Sources
54 / 63 I-Hsiang Wang IT Lecture 5
Lossy Source Coding for Continuous Memoryless Sources
f
S|S: E[d(S,
S)]≤D
55 / 63 I-Hsiang Wang IT Lecture 5
Lossy Source Coding for Continuous Memoryless Sources
i.i.d.
1 2 log
σ2 D
56 / 63 I-Hsiang Wang IT Lecture 5
Lossy Source Coding for Continuous Memoryless Sources
2 log
2 log (2πe D) = 1 2 log
σ2 D
57 / 63 I-Hsiang Wang IT Lecture 5
Lossy Source Coding for Continuous Memoryless Sources
2 log
σ2 D
S|S such that
so that h
( S − S
) = h ( S − S )
S and hence f S|S!
I-Hsiang Wang IT Lecture 5
Lossy Source Coding for Continuous Memoryless Sources
2 log
σ2 D
S|S: E[d(S,
S)]≤D I
S|S that yields the desired I
2 log
2 log 2πe
59 / 63 I-Hsiang Wang IT Lecture 5
Lossy Source Coding for Continuous Memoryless Sources
1 The distortion should be D: E
!
2 The induced mutual information is upper bounded by RG(D):
2 log 2πe
2 log 2πe
2 log a2σ2+b2 b2 !
2 log σ2 D
D
σ2D σ2−D =
σ2−D D
60 / 63 I-Hsiang Wang IT Lecture 5
61 / 63 I-Hsiang Wang IT Lecture 5
fX,Y (X,Y ) fX(X)fY (Y )
1 fX(X)
1 fX|Y (X|Y )
g(X)
62 / 63 I-Hsiang Wang IT Lecture 5
X: E[b(X)]≤B
f
S|S: E[d(S,
S)]≤D
2 log
σ2
1 2 log
σ2 D
63 / 63 I-Hsiang Wang IT Lecture 5