Lecture 4 Capacity of Wireless Channels I-Hsiang Wang - - PowerPoint PPT Presentation

lecture 4 capacity of wireless channels
SMART_READER_LITE
LIVE PREVIEW

Lecture 4 Capacity of Wireless Channels I-Hsiang Wang - - PowerPoint PPT Presentation

Lecture 4 Capacity of Wireless Channels I-Hsiang Wang ihwang@ntu.edu.tw 3/20, 2014 What we have learned So far: looked at specific schemes and techniques Lecture 2: point-to-point wireless channel -


slide-1
SLIDE 1

Lecture ¡4 Capacity ¡of ¡Wireless ¡Channels

I-Hsiang Wang ihwang@ntu.edu.tw 3/20, 2014

slide-2
SLIDE 2

What ¡we ¡have ¡learned

  • So far: looked at specific schemes and techniques
  • Lecture 2: point-to-point wireless channel
  • Diversity: combat fading by exploiting inherent diversity
  • Coding: combat noise, and further exploits degrees of freedom
  • Lecture 3: cellular system
  • Multiple access: TDMA, CDMA, OFDMA
  • Interference management: orthogonalization (partial frequency

reuse), treat-interference-as-noise (interference averaging)

2

slide-3
SLIDE 3

Information ¡Theory

  • Is there a framework to …
  • Compare all schemes and techniques fairly?
  • Assert what is the fundamental limit on how much rate can be

reliably delivered over a wireless channel?

  • Information theory!
  • Provides a fundamental limit to (coded) performance
  • Identifies the impact of channel resources on performance
  • Suggest novel techniques to communicate over wireless channels
  • Information theory provides the basis for the modern

development of wireless communication

3

slide-4
SLIDE 4

Historical ¡Perspective

4

  • First radio built 100+ years ago
  • Great stride in technology
  • But design was somewhat ad-hoc

1901 1948

  • G. Marconi
  • C. Shannon

Engineering ¡meets ¡science New ¡points ¡of ¡view ¡arise

  • Information theory: every channel

has a capacity

  • Provides a systematic view of all

communication problems

slide-5
SLIDE 5

Modern ¡View ¡on ¡Multipath ¡Fading

  • Classical view: fading channels are unreliable
  • Diversity techniques: average out the variation
  • Modern view: exploit fading to gain spectral efficiency
  • Thanks to the study on fading channel through the lens of

information theory!

5

Time Channel quality

slide-6
SLIDE 6

Plot

  • Use a heuristic argument (geometric) to introduce the

capacity of the AWGN channel

  • Discuss the two key resources in the AWGN channel:
  • Power
  • Bandwidth
  • The AWGN channel capacity serves as a building block

towards fading channel capacity:

  • Slow fading channel: outage capacity
  • Fast fading channel: ergodic capacity

6

slide-7
SLIDE 7

Outline

  • AWGN Channel Capacity
  • Resources of the AWGN Channel
  • Capacity of some LTI Gaussian Channels
  • Capacity of Fading Channels

7

slide-8
SLIDE 8

8

AWGN ¡Channel ¡Capacity

slide-9
SLIDE 9

9

Channel ¡Capacity

  • Capacity := the highest data rate can be delivered

reliably over a channel

  • Reliably ≡ Vanishing error probability
  • Before Shannon, it was widely believed that:
  • to communicate with error probability → 0
  • ⟹ data rate must also → 0
  • Repetition coding (with M-level PAM) over N time slots
  • n AWGN channel:
  • Error probability
  • Data rate
  • As long as M ≤ N⅓, the error probability → 0 as N → ∞
  • But, the data rate
  • still → 0 as N → ∞

∼ 2Q r 6N M 3 SNR ! = log2 M N = log2 N 3N

slide-10
SLIDE 10

Channel ¡Coding ¡Theorem

  • For every memoryless channel, there is a definite number C

that is computable such that:

  • If the data rate R < C, then there exists a coding scheme that can

deliver rate R data over the channel with error probability → 0 as the coding block length N → ∞

  • Conversely, if the data rate R > C, then no matter what coding

scheme is used, the error probability → 1 as N → ∞

  • We shall focus on the additive white Gaussian noise

(AWGN) channel

  • Give a heuristic argument to derive the AWGN channel capacity

10

slide-11
SLIDE 11

AWGN ¡Channel

  • We consider real-valued Gaussian channel
  • As mentioned earlier, repetition coding yield zero rate if

the error probability is required to vanish as N → ∞

  • Because all codewords are spread on a single dimension

in an N-dimensional space

  • How to do better?

11

y[n] = x[n] + z[n] x[n] z[n] ∼ N(0, σ2)

Power constraint:

N

X

n=1

|x[n]|2 ≤ NP

slide-12
SLIDE 12

Sphere ¡Packing ¡Interpretation

12

y = x + z RN

  • By the law of large numbers,

as N → ∞, most y will lie inside the N-dimensional sphere of radius

p N(P + σ2)

  • Also by the LLN, as N → ∞,

y will lie near the surface of the N-dimensional sphere centered at x with radius √

Nσ2

How many non-overlapping spheres can be packed into the large sphere?

  • Vanishing error probability

⟹ non-overlapping spheres

p N(P + σ2) √ Nσ2

slide-13
SLIDE 13

Why ¡Repetition ¡Coding ¡is ¡Bad

13

y = x + z RN

It only uses one dimension

  • ut of N !

p N(P + σ2)

slide-14
SLIDE 14

Capacity ¡Upper ¡Bound

14

y = x + z RN

p N(P + σ2) √ Nσ2

Maximum # of non-overlapping spheres = Maximum # of codewords that can be reliably delivered = ⇒ R ≤ 1 N log p N(P + σ2)

N

√ Nσ2N ! = 1 2 log ✓ 1 + P σ2 ◆ 2NR ≤ p N(P + σ2)

N

√ Nσ2N This is hence an upper bound of the capacity C. How to achieve it?

slide-15
SLIDE 15

Achieving ¡Capacity ¡(1/3)

  • (random) Encoding: randomly generate 2NR codewords

{x1, x2, ...} lying inside the “x-sphere” of radius

  • Decoding:
  • Performance analysis: WLOG let x1 is sent
  • By the LLN,
  • As long as x1 lies inside the uncertainty sphere centered at αy

with radius

  • , decoding will be correct
  • Pairwise error probability (see next slide) =

15

√ NP y − → MMSE − → αy − → Nearest Neighbor − → b x

α := P P + σ2

||αy − x1||2 = ||αw + (α − 1)x1||2 ≈ α2Nσ2 + (α − 1)2NP = N Pσ2 P + σ2 r N Pσ2 P + σ2 ✓ σ2 P + σ2 ◆N/2

slide-16
SLIDE 16

Achieving ¡Capacity ¡(2/3)

16

√ NP

x-sphere

r N Pσ2 P + σ2 x1

When does an error occur? Ans: when another codeword x2 falls inside the uncertainty sphere of αy What is that probability (pairwise error probability)? Ans: the ratio of the volume of the two spheres

Pr {x1 → x2} = p NPσ2/(P + σ2)

N

√ NP

N

= ✓ σ2 P + σ2 ◆N/2

Union bound: Total error probability ≤ 2NR

✓ σ2 P + σ2 ◆N/2 αy x2

slide-17
SLIDE 17

Achieving ¡Capacity ¡(3/3)

  • Total error probability (by union bound)
  • As long as the following holds,
  • Hence, indeed the capacity is

17

Pr {E} → 0 as N → ∞ R < 1 2 log ✓ 1 + P σ2 ◆ Pr {E} ≤ 2NR ✓ σ2 P + σ2 ◆N/2 = 2

N R+ 1

2 log 1 1+ P σ2

!!

C = 1 2 log ✓ 1 + P σ2 ◆ bits per symbol time

slide-18
SLIDE 18

18

Resources ¡of ¡ AWGN ¡Channel

slide-19
SLIDE 19

Continuous-­‑Time ¡AWGN ¡Channel

  • System parameters:
  • Power constraint: P watts; Bandwidth: W Hz
  • Spectral density of the white Gaussian noise: N0/2
  • Equivalent discrete-time baseband channel (complex)
  • 1 complex symbol = 2 real symbols
  • Capacity:

19

Power constraint:

N

X

n=1

|x[n]|2 ≤ NP

y[n] = x[n] + z[n] x[n] z[n] ∼ CN(0, N0W) CAWGN(P, W) = 2 × 1 2 log ✓ 1 + P/2 N0W/2 ◆ bits per symbol time = W log ✓ 1 + P N0W ◆ bits/s = log (1 + SNR) bits/s/Hz

SNR := P/N0W SNR per complex symbol

slide-20
SLIDE 20

Complex ¡AWGN ¡Channel ¡Capacity ¡

  • The capacity formula provides a high-level way of

thinking about how the performance fundamentally depends on the basic resources available in the channel

  • No need to go into details of specific coding and

modulation schemes

  • Basic resources: power P and bandwidth W

20

CAWGN(P, W) = W log ✓ 1 + P N0W ◆ bits/s = log (1 + SNR) bits/s/Hz Spectral Efficiency

slide-21
SLIDE 21

Power

  • High SNR:
  • Logarithmic growth with power
  • Low SNR:
  • Linear growth with power

21 3 4 5 6 7 20 40 60 80 100 1 2 SNR log (1 + SNR)

C = log(1 + SNR) ≈ log SNR C = log(1 + SNR) ≈ SNR log2 e

SNR = P N0W

Fix W:

slide-22
SLIDE 22

Bandwidth

22

30 5 Bandwidth W (MHz) Capacity Limit for W → ∞ Power limited region 0.2 1 Bandwidth limited region (Mbps) C(W ) 0.4 25 20 15 10 1.6 1.4 1.2 0.8 0.6 P N0log2 e

C(W) = W log ✓ 1 + P N0W ◆ ≈ W P N0W log2 e = P N0 log2 e

Fix P

slide-23
SLIDE 23

Bandwidth-­‑limited ¡vs. ¡Power-­‑limited

  • When SNR ≪ 1: (Power-limited regime)
  • Linear in power; Insensitive to bandwidth
  • When SNR ≫ 1: (Bandwidth-limited regime)
  • Logarithmic in power; Approximately linear in bandwidth

23

CAWGN(P, W) = W log ✓ 1 + P N0W ◆ bits/s SNR = P N0W CAWGN(P, W) ≈ W ✓ P N0W ◆ log2 e = P N0 log2 e CAWGN(P, W) ≈ W log ✓ P N0W ◆

slide-24
SLIDE 24

24

Capacity ¡of ¡Some LTI ¡Gaussian ¡Channels

slide-25
SLIDE 25

25

SIMO ¡Channel

  • MRC is a lossless operation: we can generate y from :
  • Hence the SIMO channel capacity is equal to the

capacity of the equivalent AWGN channel, which is

h x y e y = ||h||x + e w, e w ∼ CN

  • 0, σ2

↓ MRC, h∗ ||h|| ↓ y = hx + w ∈ CL

Power constraint: P w ∼ CN

  • 0, σ2IL
  • e

y y = e y (h/||h||)

CSIMO = log ✓ 1 + ||h||2P σ2 ◆ Power gain due to Rx beamforming

slide-26
SLIDE 26

MISO ¡Channel

  • Goal: maximize the received power ||h*x||2
  • The answer is ||h||2P ! (check. Hint: SVD)
  • Achieved by Tx beamforming
  • Send a scalar symbol x on the direction of h
  • Power constraint on x : still P
  • Capacity:

26

h x y y = h∗x + w ∈ C, x, h ∈ CL h∗ = ⇥h1 h2 ⇤ y = x||h|| + w

↓ Tx Beamforming x = xh/||h|| ↓ Power constraint:

N

X

n=1

||x||2 ≤ NP

CMISO = log ✓ 1 + ||h||2P σ2 ◆

slide-27
SLIDE 27

Frequency-­‑Selective ¡Channel

  • Key idea 1: use OFDM to convert the channel with ISI

into a bunch of parallel AWGN channels

  • But there is loss/overhead due to cyclic prefix
  • Key idea 2: CP overhead → 0 as Nc → ∞
  • First focus on finding the capacity of parallel AWGN

channels of any finite Nc

  • Then take Nc → ∞ to find the capacity of the frequency-

selective channel

27

y[m] =

L−1

X

l=0

hlx[m − l] + w[m]

slide-28
SLIDE 28

Recap: ¡OFDM

28 d[N–1] ˜ y0 x [N + L – 1] = d[N – 1] Cyclic prefix y [N + L – 1] ˜ dN–1 IDFT DFT Remove prefix ˜ yN–1 y[L] y[N + L – 1] y[1] y[L – 1] y[L] x [L – 1] = d[N – 1] x [L] = d[0] x [1] = d[N – L + 1] ˜ d0 d[0] Channel

y := y[L : Nc + L − 1], w := w[L : Nc + L − 1], h := ⇥h0 h1 · · · hL−1 · · · 0⇤T

e yn = e hn e dn + e wn, n = 0, 1, . . . , Nc − 1

e yn := DFT (y)n , e dn := DFT (d)n , e wn := DFT (w)n , e hn := p NcDFT (h)n

Nc parallel AWGN channels

slide-29
SLIDE 29

Parallel ¡AWGN ¡Channels

29

e d0[m] e h0 e y0[m] e w0[m]

. . .

e d1[m] e h1 e y1[m] e w1[m] e dNc−1[m] e hNc−1 e wNc−1[m] e yNc−1[m]

Parallel Channels e yn = e hn e dn + e wn, n ∈ [0 : 1 : Nc − 1] m = 1, 2, . . . , M (M channel uses) Power Constraint

M

X

n=1

||e d[n]||2 ≤ MNcP

Due to Parseval theorem of DFT

Equivalent Vector Channel e y = e He d + e w e H = diag ⇣ e h0, . . . ,e hNc−1 ⌘ e w ∼ CN

  • 0, σ2I
slide-30
SLIDE 30

Independent ¡Uses ¡of ¡Parallel ¡Channels

  • One way to code over such parallel channels (a special

case of a vector channel): treat each channel separately

  • It turns out that coding across parallel channels does not help!
  • Power allocation:
  • Each of the Nc channels get a portion of the total power
  • Channel n gets power Pn , which must satisfy
  • For a given power allocation {Pn}, the following rate can

be achieved:

30

Nc−1

X

n=0

Pn ≤ NcP R =

Nc−1

X

n=0

log 1 + |e hn|2Pn σ2 !

slide-31
SLIDE 31

Optimal ¡Power ¡Allocation

  • Power allocation problem:
  • It can be solved explicitly by Lagrangian methods
  • Final solution: let (x)+ := max(x , 0)

31

max

P0,...,PNc−1 Nc−1

X

n=0

log 1 + |e hn|2Pn σ2 ! , subject to

Nc−1

X

n=0

Pn = NcP, Pn ≥ 0, n = 0, . . . , Nc − 1 P ∗

n =

ν − σ2 |e hn|2 !+ , ν satisfies

Nc−1

X

n=0

ν − σ2 |e hn|2 !+ = NcP

slide-32
SLIDE 32

Water\illing

32

σ2 |e hn|2 ν

Note: e hn = Hb ✓nW Nc ◆ Baseband frequency response at f = nW/Nc

P1 = 0 Subcarrier n P2 P3 * * * 1 λ

slide-33
SLIDE 33

Frequency-­‑Selective ¡Channel ¡Capacity

  • Final step: making Nc → ∞
  • Replace all
  • by Hb(f), summation over [0 : Nc – 1]

becomes integration from 0 to W

  • Power allocation problem becomes
  • Optimal solution becomes

33

e hn = Hb ✓nW Nc ◆

max

P (f)

Z W log ✓ 1 + |H(f)|2P(f) σ2 ◆ d f, subject to Z W P(f) = P, P(f) ≥ 0, f ∈ [0, W] P(f)∗ = ✓ ν − σ2 |H(f)|2 ◆+ , ν satisfies Z W ✓ ν − σ2 |H(f)|2 ◆+ d f = P

slide-34
SLIDE 34

Water\illing ¡over ¡the ¡Frequecy ¡Spectrum

34

P ( f ) Frequency ( f ) 0.4W 0.2W – 0.2W – 0.4W 4 3.5 3 2.5 2 1.5 1 0.5

|2

* λ

ν σ2 |H(f)|2

slide-35
SLIDE 35

35

Capacity ¡of ¡Fading ¡ Channels

slide-36
SLIDE 36

36

Flat ¡vs. ¡Frequency ¡Selective ¡Fading

  • Frequency selectivity induces ISI
  • Frequency selective underspread channel can be

converted to parallel flat fading channels

  • We focus on flat fading channels (single tap):

x[m] h[m] w[m] y[m] Channel Input Channel Output Channel Gain Noise

y[m] = h[m]x[m] + w[m]

E ⇥ |h[m]|2⇤ = 1, ∀ m Power constraint:

N

X

n=1

|x[n]|2 ≤ NP

slide-37
SLIDE 37

Slow ¡vs. ¡Fast ¡Fading

  • Slow fading (quasi-static)
  • h[m] = h for all m
  • h is random, unknown to Tx
  • h is known to the Rx (if not, it is the non-coherent setting)
  • Fast fading
  • h[m] = hl for all m within the l-th coherence time (Tc) period
  • hl : i.i.d. over l, that is, i.i.d. over different coherence time periods
  • hl is random, could be known or unknown to Tx
  • hl is known to the Rx (if not, it is the non-coherent setting)

37

x[m] h[m] w[m] y[m]

slide-38
SLIDE 38

Slow ¡Fading ¡Channel

  • h[m] = h for all m and h is random
  • Conditional on h, channel capacity = log(1+|h|2SNR)
  • Tx does not know the channel h
  • Suppose Tx send at rate R bits/s/Hz:
  • If R < log(1+|h|2SNR), Shannon: “Pe → 0 as N → ∞”
  • If R > log(1+|h|2SNR), Shannon: “Pe → 1 as N → ∞”
  • Total Error probability:

38

P (N)

e

= Pr

  • R < log
  • 1 + |h|2SNR
  • Pr
  • E | R < log
  • 1 + |h|2SNR
  • + Pr
  • R > log
  • 1 + |h|2SNR
  • Pr
  • E | R > log
  • 1 + |h|2SNR
  • → Pr
  • R < log
  • 1 + |h|2SNR
  • 0 + Pr
  • R > log
  • 1 + |h|2SNR
  • 1

= Pr

  • R > log
  • 1 + |h|2SNR
  • as N → ∞
slide-39
SLIDE 39

Shannon ¡Capacity ¡= ¡0

  • Error probability → Pr{R > log(1+|h|2SNR)} as N → ∞
  • Vanishing error probability: impossible due to deep fade!
  • According to Shannon’s definition, capacity = 0 …
  • Outage probability pout(R) := Pr{R > C(h ; SNR)}
  • Determines the reliability level of sending at a particular rate R
  • C(h ; SNR) = log(1+|h|2SNR) for the point-to-point channel

39

slide-40
SLIDE 40

Outage ¡Probability: ¡Computation

40

pout(R) := Pr

  • R > log
  • 1 + |h|2SNR
  • = Pr
  • |h|2 < (2R − 1)SNR−1

for h ∼ CN(0, 1): Rayleigh fading

0.15 0.2 0.25 0.3 0.35 0.4 0.45 1 2 3 4 5 0.05 0.1 R Area = pout (R)

= 1 − e

−(2R−1) SNR

slide-41
SLIDE 41

Outage ¡Capacity

  • Shannon capacity of a slow fading channel is 0
  • Because we insist that “Pe → 0 as N → ∞”!
  • More realistic capacity measure:
  • Set a reliability level ϵ
  • Find the maximum rate R* such that Pout(R) ≤ ϵ
  • ϵ-outage capacity:
  • It’s just the inverse function of outage probability! (why?)

41

C✏ := max {R | pout (R) ≤ ✏} C✏ = log

  • 1 + F −1 (1 − ✏) SNR
  • F (x) := Pr
  • |h|2 > x

: complementary CDF of |h|2

slide-42
SLIDE 42

Outage ¡Capacity: ¡Computation

42

⇐ ⇒ Pr

  • 2C✏ − 1
  • SNR−1 > |h|2

= ✏ ⇐ ⇒ 1 − F

  • 2C✏ − 1
  • SNR−1

= ✏ ⇐ ⇒ 2C✏ − 1 = F −1 (1 − ✏) SNR ⇐ ⇒ C✏ = log

  • 1 + F −1 (1 − ✏) SNR
  • pout (C✏) = ✏

⇐ ⇒ Pr

  • C✏ > log
  • 1 + |h|2SNR
  • = ✏

For Rayleigh fading: h ∼ CN(0, 1) : y = F(x) = e−x ∈ [0, 1], F −1(y) = − ln y ≥ 0 F −1(1 ✏) = ln(1 ✏) ⇡ ✏, for ✏ ⌧ 1 C✏ ⇡ log (1 + ✏SNR) when ✏ ⌧ 1

slide-43
SLIDE 43

Fade ¡Margin

  • AWGN vs. ϵ-outage capacity:
  • Under a reliability level ϵ, to achieve the same capacity

as the AWGN channel, in slow fading channel the Tx has to use an extra 10log10(1/F –1(1-ϵ)) dB of power

  • Fade margin: the extra amount of Tx power to improve

the system when channel is in deep fade

43

C✏ = log

  • 1 + F −1 (1 − ✏) SNR
  • CAWGN = log (1 + SNR)

Typically less than 1 because ϵ ≪ 10% ⟹ Extra Tx power improve deep fade

slide-44
SLIDE 44

Impact ¡of ¡Fading

  • High SNR:
  • Additive loss
  • Smaller when SNR → ∞
  • Low SNR:
  • Multiplicative loss
  • For Rayleigh,
  • Significant loss when SNR is small
  • Impact of fading is also significant at low SNR

44

C✏ ≈ log

  • F −1(1 − ✏)SNR
  • ≈ CAWGN − log

✓ 1 F −1(1 − ✏) ◆ C✏ ≈ F −1(1 − ✏)SNR log2 e ≈ F −1(1 − ✏)CAWGN F −1(1 ✏) = ln(1 ✏) ⇡ ✏, for ✏ ⌧ 1

slide-45
SLIDE 45

AWGN ¡Capactiy ¡vs ¡Outage ¡Capacity

45

1 –10 –5 5 10 15 20 25 30 0.6 0.4 0.2 0.8

= 0.1 = 0.01

awgn

SNR (dB) 35 40

∋ ∋

C✏ CAWGN

slide-46
SLIDE 46

Diversity ¡Order

  • Recall:
  • Error probability → Outage probability pout(R) as N → ∞
  • Recall from Lecture 2:
  • For uncoded transmission and some coding scheme, we see that

Error probability ~ SNR–1 for some fixed N

  • Is this true for the optimal coding scheme and arbitrarily large N ?
  • Outage probability at high SNR:
  • Even for optimal coding scheme and large N, error prob. ~ SNR–1
  • Optimal diversity order is 1

46

pout(R) = 1 e

−(2R−1) SNR

⇡ 2R 1 SNR when SNR 1

slide-47
SLIDE 47

Optimal ¡Diversity ¡Order ¡and ¡pout(R)

  • Hence we are able to define the optimal diversity order

for point-to-point slow fading channels:

  • Note: in taking the limit, we assume that R is a constant
  • This view will be modified in later lectures when we

discuss the diversity-multiplexing tradeoff

47

d := lim

SNR→∞

− log pout(R) log SNR where pout(R) := Pr {C (h; SNR) < R}

slide-48
SLIDE 48

48

Receive ¡Diversity

  • Outage probability
  • Outage capacity

h x y y = hx + w ∈ CL

Power constraint: P w ∼ CN

  • 0, σ2IL
  • SNR := P

σ2

pout(R) := Pr

  • log
  • 1 + ||h||2SNR
  • < R

= Pr ⇢ ||h||2 < 2R − 1 SNR

  • ≈ (2R − 1)L

L! SNR−L for Rayleigh faded h’s

Lecture 2 Slide #25 ✏ = pout (C✏)

C✏ ≈ log ⇣ 1 + (L!)

1 L (✏) 1 L SNR

E ⇥ ||h||2⇤ = L

C (h; SNR) = log

  • 1 + ||h||2SNR
slide-49
SLIDE 49

SIMO ¡Outage ¡Capacity

49

5 10 15 20 25 30 35 40 –10 1 0.8 0.6 0.4 0.2 –5

awgn

L = 2 L = 4 L = 5 L = 3 L = 1 SNR (dB)

✏ = 1% C✏ CSIMO CSIMO = log

  • 1 + ||h||2SNR
  • = log (1 + LSNR)
slide-50
SLIDE 50

Transmit ¡Diversity

  • Tx beamforming is impossible since Tx does not know h
  • For SIMO, Tx does not need to know h to achieve C(h ; SNR)
  • How to find the optimal outage probability?
  • For i.i.d. Rayleigh fading, it can be shown (cf. Appendix B.8 and

Exercise 5.15, 5.16) that the optimal outage probability

50

h x y y = h∗x + w ∈ C, x, h ∈ CL h∗ = ⇥h1 h2 ⇤

Power constraint:

N

X

n=1

||x||2 ≤ NP SNR := P σ2 E ⇥ ||h||2⇤ = L

y = x||h|| + w

↓ Tx Beamforming x = xh/||h|| ↓

pout(R) = Pr ⇢ log ✓ 1 + ||h||2 SNR L ◆ < R

slide-51
SLIDE 51

Impact ¡of ¡CSIT: ¡Loss ¡in ¡Power ¡Gain

  • Comparison of outage probability
  • The same diversity order L
  • SIMO has L-fold power gain over MISO
  • Lack of channel state information at the transmitter

(CSIT) ⟹ Loss in power gain

51

pTx

  • ut(R) = Pr

⇢ log ✓ 1 + ||h||2 SNR L ◆ < R

  • pRx
  • ut(R) = Pr
  • log
  • 1 + ||h||2SNR
  • < R
slide-52
SLIDE 52

Repetition ¡Coding

52

Time 1 Time 2 Equivalent Channel: h1 h2 h1 h2

u u

y1 y2

  • =

h1 h2

  • u +

w1 w2

  • Projection

− − − − − − → e y = h∗ ||h||y = ||h||u + e w Supports rates up to 1 2 log

  • 1 + ||h||2SNR
  • Outage probability:

pRepetition

  • ut

(R) = Pr ⇢1 2 log

  • 1 + ||h||2SNR
  • < R
  • 2 blocks, but just 1 scalar channel
slide-53
SLIDE 53

Alamouti ¡Scheme

53

Time 1 Time 2

u1 u2 −u∗

2

u∗

1

X = u1 −u∗

2

u2 u∗

1

  • u1, u2 ∈ C

space-time codeword Equivalent Channel: y1 y∗

2

  • =

h1 h2 h∗

2

−h∗

1

u1 u2

  • +

w1 w2

  • = u1

h1 h∗

2

  • + u2

 h2 −h∗

1

  • +

w1 w2

  • h1

h2 h1 h2 Projection onto the two column vectors respectively, we can get two clean channels for u1 and u2! e h1 e h2 e h1 ⊥ e h2

slide-54
SLIDE 54

Performance ¡of ¡Alamouti ¡Scheme

  • Projection onto two orthogonal directions
  • Power allocation: Lack of CSIT ⟹ No idea which

channel is better ⟹ Uniform power allocation (P/2 each)

  • Each channel supports rates up to
  • Outage probability: achieves optimal

54

y1 y∗

2

  • = u1

h1 h∗

2

  • + u2

 h2 −h∗

1

  • +

w1 w2

  • e

h1 ⊥ e h2 e y = u1e h1 + u2e h2 + e w Two parallel channels, each for one symbol!

e y1 := e h∗

1

||e h1|| e y = u1||e h1|| + e w1 = u1||h|| + e w1 e y2 := e h∗

2

||e h2|| e y = u2||e h2|| + e w2 = u2||h|| + e w2

log ✓ 1 + ||h||2 SNR 2 ◆ pAlamouti

  • ut

(R) = Pr ⇢ log ✓ 1 + ||h||2 SNR 2 ◆ < R

slide-55
SLIDE 55

Repetition ¡Coding ¡vs. ¡Alamouti

55

Repetition Alamouti C(h;SNR) pout(R) when SNR ≫ 1 Cϵ for ϵ ≪ 1 Diversity Order 2 2

1 2 log

  • 1 + ||h||2SNR
  • log

✓ 1 + ||h||2 SNR 2 ◆ (22R − 1)2 2 SNR−2 1 2 log ⇣ 1 + √ 2✏SNR ⌘ 2(2R − 1)2SNR−2 log ✓ 1 + r ✏ 2SNR ◆

≤ ≥ ≤ =

slide-56
SLIDE 56

Time ¡and ¡Frequency ¡Diversity

  • Recall from Lecture 2:
  • Time diversity is obtained by coding + interleaving across multiple

(L) coherence time

  • Frequency diversity is obtained by coding + hopping across

multiple (L) coherence bandwidth

  • Hence, time and frequency diversity techniques are

equivalent to coding over L parallel channels:

  • Channel l has power constraint Pl , P1+P2+…+PL ≤ LP,

where P is the power constraint of the original channel

  • No CSIT ⟹ cannot do water-filling

56

yl[m] = hlxl[m] + wl[m], l = 1, 2, . . . , L

slide-57
SLIDE 57

Outage ¡Probability

  • Instead, first use uniform power allocation Pl = P, ∀l
  • Given h and SNR, L parallel channels can support up to
  • Subtlety: due to lack of CSIT, coding across parallel

channel is necessary

  • Because Tx does not know for each of the L channels, how high

the rate should be!

  • Outage probability:
  • For i.i.d. Rayleigh fading, it can be shown (cf. Exercise 5.17)

that uniform power allocation is optimal!

57

pout(R) = Pr ( L X

l=1

log

  • 1 + |hl|2SNR
  • < LR

)

L

X

l=1

log

  • 1 + |hl|2SNR
  • bits/s/Hz