Timing Recovery at Low SNR Cramer-Rao bound, and outperforming the - - PowerPoint PPT Presentation

timing recovery at low snr cramer rao bound and
SMART_READER_LITE
LIVE PREVIEW

Timing Recovery at Low SNR Cramer-Rao bound, and outperforming the - - PowerPoint PPT Presentation

I N S T A I T I G U R T O E E G O F E H T T E C F H P R O G R E S S S E R V I C E O N A N D O L L A O E G S Y 1 8 8 5 Timing Recovery at Low SNR


slide-1
SLIDE 1

Timing Recovery at Low SNR Cramer-Rao bound, and outperforming the PLL

Aravind R. Nayak John R. Barry Steven W. McLaughlin {nayak, barry, swm}@ece.gatech.edu Georgia Institute of Technology

N I

  • A

I G R O E G

  • E

H T

  • F

O

  • L

A E S

  • S

T I T U T E

  • O

F

  • T

E C H N O L O G Y

  • 8

5 8 1

N A D P R O G R E S S S E R V I C E
slide-2
SLIDE 2

1

Communication system model

Transmitter Channel Receiver

Source ECC Encoder Modulator Channel Sampler Equalizer ECC Decoder Discrete-time Continuous-time

slide-3
SLIDE 3

2

Continuous-time discrete-time interface

Source ECC Encoder Modulator Channel Sampler Equalizer ECC Decoder Discrete-time Continuous-time to Continuous-time Discrete-time to

slide-4
SLIDE 4

3

Sampling: Timing recovery

T 2T 3T

τ0 τ1 τ2 –τ3 a1 a0 a2 a3

TIME

T – Symbol duration a0, a1, a2,... – Data symbols τ0, τ1, τ2,... – Timing offsets Timing Recovery Problem: Estimate τ0, τ1, τ2, ...

slide-5
SLIDE 5

4

Constant offset: Frequency offset: Random walk: where wi are i.i.d. zero-mean Gaussian random variables of variance . determines the severity of the random walk.

τk τ = τk τ0 k∆T + τk

1 –

∆T + = = τk

1 +

τk wk + τ0 wi

i = k

+ = = σw

2

σw

2

Timing offset models

slide-6
SLIDE 6

5

Acquisition:

  • Estimate τ0
  • Correlation techniques
  • Known preamble sequence at start of packet (Trained mode)
  • Parameter τ0 spans a large range

Tracking:

  • Keep track of τ1, τ2, τ3,...
  • Based on the phase-locked loop (PLL)
  • Data symbols unknown (Decision-directed mode)
  • Sufficient to track small signals τ1 –τ0 , τ2 –τ1 , τ3 –τ2 , ...

Timing recovery in two stages

slide-7
SLIDE 7

6

PLL: Motivation

Consider the simple case of a time-invariant offset: τk = τ Let be the current timing estimate. Timing error: εi = τi – = τ – . With a perfect timing error detector (TED), we get = εi . Update: With imperfect TED: τ ˆi τ ˆi τ ˆi ε ˆi τ ˆi

1 +

τ ˆi ε ˆi + τ = = τ ˆi

1 +

τ ˆi αε ˆi + =

slide-8
SLIDE 8

7

PLL-based timing recovery

PLL UPDATE

f(t) y(t)

  • rcv. filter

T.E.D.

kT +

k

τ ˆ

rk

for further processing

k

ε ˆ

r t ( ) τ ˆk

1 +

τ ˆk αε ˆk + = τ ˆk

1 +

τ ˆk αε ˆk β ε ˆi

i = k 1 –

+ + =

First-order PLL Second-order PLL

slide-9
SLIDE 9

8

Timing Error Detector (TED)

PLL UPDATE

f(t) y(t)

  • rcv. filter

T.E.D.

kT +

k

τ ˆ

rk

k

d ˆ

for further processing

k

ε ˆ

r t ( ) ε ˆk

3T 16

  • rkd

ˆ k

1 –

rk

1 – d

ˆ k – ( ) =

Mueller & Müller Timing Error Detector

  • TED is a decision-directed device
  • Usually, instantaneous hard quantization
  • Better decisions entail delay that destabilizes the loop
slide-10
SLIDE 10

9

  • Improve the quality of decisions (Approach I)

⇒ Need to get around the delay induced by better decisions. ⇒ Feedback from the ECC decoder and equalizer to timing recovery.

  • Dr. Barry’s presentation!
  • Improve the timing recovery architecture (Approach II)

⇒ Need to assume perfect decisions for tractability. ⇒ Methods based on gradient search and projection operation. ⇒ Use Cramer-Rao bound to evaluate competing methods. This presentation!

Improving timing recovery

slide-11
SLIDE 11

10

Overview: Approach II

Questions:

  • How good is the PLL-based system?
  • Can it be improved upon?

Method:

  • Derive fundamental performance limits.
  • Compare the PLL performance with these limits.
  • Develop methods that outperform the PLL.
slide-12
SLIDE 12

11

Problem statement

The uniform samples are: rk = alh(kT – lT – τl) + nk , where σ2 is the noise variance, and h(t) is the impulse response. Problem: Given samples {rk} and knowledge of channel model, estimate

  • the N uncoded i.i.d. data symbols {ak}
  • the N timing offsets {τk}.

l = N 1 –

AWGN

τ

h(t)

We consider the following uncoded system: ak { }0

N 1 – uncoded i.i.d.

rk

LPF

kT

(uniform)

slide-13
SLIDE 13

12

Cramer-Rao bound

  • answers the following question:

“What is the best any estimator can do?”

  • is independent of the estimator itself.
  • is a lower bound on the error variance of any estimator.

Cramer-Rao bound (CRB)

slide-14
SLIDE 14

13

→ fixed, unknown parameter to be estimated r → observations

  • Sensitivity of

to changes in determines quality of estimation.

  • If

is narrow, for a given r, probable s lie in a narrow range. ⇒ can be estimated better, i.e., with lesser error variance.

  • CRB uses

as a measure of narrowness.

θ f r θ ( ) θ f r θ ( ) θ θ ∂ ∂θ

  • f r θ

( ) log

CRB, intuitively

f r θ1 ( ) f r θ1 ( ) f r θ2 ( ) f r θ2 ( )

r r

slide-15
SLIDE 15

14

CRB for a random parameter

If is random as opposed to being fixed and unknown,

  • is characterized by a p.d.f.

and

  • r,

are characterized by the joint p.d.f. . The measure for narrowness in this case is

θ θ f θ ( ) θ f r θ , ( ) ∂ ∂θ

  • f r θ

, ( ) log

slide-16
SLIDE 16

15

For any unbiased estimator , the estimation error covariance matrix is lower bounded by where is the information matrix given by In particular,

θ ˆ r ( ) E θ ˆ r ( ) θ – ( ) θ ˆ r ( ) θ – ( )

T

[ ] J 1

≥ J J E ∂ ∂θ

  • f r θ

, ( ) log ∂ ∂θ

  • f r θ

, ( ) log

T

      = E θ ˆ i r ( ) θi – ( )

2

[ ] J 1

i i , ( ) ≥

CRB is the inverse of Fisher information

slide-17
SLIDE 17

16

  • An estimator that achieves the CRB is called efficient.
  • Efficient estimators do not always exist.

Fixed, unknown θ: ML is efficient Random θ: MAP is efficient

∂ ∂θ

  • f r θ

( ) log = ∂ ∂θ

  • f r θ

, ( ) log =

Efficient estimators

slide-18
SLIDE 18

17

CRB: lower bound on timing error variance

Constant offset: Frequency offset:

σε

2

σ2 NEh'

σε

2

6σ2 N 1 – ( )N 2N 1 – ( )Eh'

slide-19
SLIDE 19

18

CRB for a random walk

The Cramer-Rao bound on the error variance of any unbiased timing estimator: where is the steady state value, , and . E τ ˆk τk – ( )2 [ ] h f k ( ) ⋅ ≥ h σw

2

η η2 1 –

  • =

f k ( ) N 0.5 + ( ) η log ( ) 1 N 0.5 2 k 1 + ( ) – + ( ) η log ( ) sinh N 0.5 + ( ) η log ( ) sinh

tanh = η λ λ2 4 – + 2

  • =

λ 2 2π2 3

  • 1

–     σw

2

σ2T2

  • +

=

slide-20
SLIDE 20

19

Steady-state value becomes more representative as SNR and N increase.

CRB: Steady-state value

5000 0.4% 0.8% 1.2% Time

σε ⁄ T (%)

σw ⁄ T = 0.05%

h

N = 5000 Parameters SNRbit = 5 dB

slide-21
SLIDE 21

20

Trained PLL away from the CRB

2050 4100 1% 2% 3% 4% 5% Time (bit periods) CRLB Trained PLL with α optimized

N = 4095 Parameters SNR = 5 dB σw ⁄ T = 0.7% α = 0.03 10000 sectors

ESTIMATION ERROR JITTER σε ⁄ T (%)

Trained PLL does not achieve the steady-state CRB.

slide-22
SLIDE 22

21

10 20 30 1% 2% 3% 4% 5% ESTIMATION ERROR JITTER σε ⁄ T (%) SNR (dB)

7 dB

α = 0.01 α = 0.02 α = 0.03 α = 0.05 α = 0.1 α = 0.2 α = 0.03 α = 0.05 CRB

N = 500 Parameters σw ⁄ T = 0.5% 1000 trials

Trained PLL vs. Steady-state CRB

slide-23
SLIDE 23

22

  • As in the random walk case, the PLL does not achieve the

CRB in the constant offset and the frequency offset cases.

  • Using Kalman filtering analysis, we can show that PLL is

the optimal causal timing recovery scheme. ⇒ Eliminate causality constraint to improve performance. ⇒ Block processing.

Outperforming the PLL: Block processing

slide-24
SLIDE 24

23

The trained maximum-likelihood (ML) estimator picks to minimize This minimization can be implemented using gradient descent:

  • Initialization using PLL.
  • Without training, use

instead of .

τ J τ ˆ a ; ( ) rk alh kT lT – τ ˆ – ( )

l

–     2

k ∞ – = ∞

= τ ˆi

1 +

τ ˆi µJ' τ ˆi a ; ( ) – = J τ ˆ a ˆ ; ( ) J τ ˆ a ; ( )

Constant offset: Gradient search

slide-25
SLIDE 25

24

Trained ML achieves CRB

α = 0.01 Parameters N = 5000 τ/T = π/20

  • 8
  • 3

2 7 0.2% 2% 20% SNR (dB) 0.3% 1% 10% 3% RMS Timing Error σε / T

Trained ML, CRB Trained PLL Decision-directed PLL Decision-directed ML

Two ways to improve performance over conventional PLL: * Better architecture – ML for example. * Better decisions – exploit error correction codes.

slide-26
SLIDE 26

25

Frequency offset: Least squares estimation

Let , , from PLL. Model: Problem: Find and to minimize Solution: and

k 0 1 … N 1 – , , , [ ]T = τ τ0 τ1 … τN

1 –

, , , [ ]T = τ ˆ τ ˆ0 τ ˆ1 … τ ˆ N

1 –

, , , [ ]T = τ ∆T ( )k τ0 + = ∆T ˆ τ ˆ0 τ ˆ ∆T ( ) ˆ k τ0 ˆ + ( ) –

2

∆T ˆ N kτ ˆk

k τ ˆk

∑ ∑

– N k2

k

( )2 –

  • =

τ ˆ0 1 N

  • τ

ˆk k∆T ˆ – ( )

=

1000 2000 3000 4000 2 4 6 8 k

τ ˆk τk

slide-27
SLIDE 27

26

Least Squares away from CRB

10000 packets Parameters N = 250

4 8 12 16 10–5 10–4 10–3 10–2 SNR (dB) RMS Estimation Error in ∆T / T 4 8 12 16 10–2 10–1 SNR (dB) RMS Estimation Error in τ0 / T

Decision-Directed Trained CRB CRB Trained Decision-Directed

* Trained MM + PLL + LS about 2 dB away from the CRB ⇒ Gradient descent?

∆T/T ∼ unif[0, 0.005] τ0/T ~ unif[0, 0.1] α optimized

slide-28
SLIDE 28

27

Gradient descent not suitable

−0.1 −0.05 0.05 0.1 500 1000 1500

Given uniform samples , pick and to minimize rk { } ∆T ˆ τ ˆ0 J ∆T ˆ τ ˆ0 a ; , ( ) rk alh kT lT – l∆T ˆ – τ ˆ0 – ( )

l

–     2

k

= ∆T ˆ T ⁄ J ∆T ˆ τ ˆ0 , ( ) parabolic in . J ∆T ˆ τ ˆ0 , ( ) τ ˆ0 Gradient descent→ sensitive to initialization → proceeds along greatest gradient: rattling in the bowl

slide-29
SLIDE 29

28

Gradient descent:

  • Moves along the direction of greatest gradient,
  • Long, narrow valley ⇒ this is not a good idea.

Newton’s method:

  • Makes parabolic approximation,
  • Directly computes the location of the minimum,
  • Efficacy depends on how good the parabolic approximation is.

LM combines these two estimates using a weight factor λ.

Levenberg-Marquardt method

slide-30
SLIDE 30

29

The update box * updates the estimate wi → wi+1, * increases λ if error increased; decreases λ if error decreased.

Levenberg-Marquardt (LM) method

PLL LS Uniform Sampler Compute Update

r t ( ) τ ˆ w ˆ 0

Initialization Levenberg-Marquardt

y ˆ w ˆ

slide-31
SLIDE 31

30

* Trained MM + PLL + LS + LM method achieves the CRB.

Trained LM achieves CRB

4 8 12 16 10–5 10–4 10–3 SNR (dB) RMS Estimation Error in ∆T / T 4 8 12 16 10–2 10–1 SNR (dB) RMS Estimation Error in τ0 / T CRB Trained LM CRB Trained LM

10000 packets Parameters N = 250 ∆T/T ∼ unif[0, 0.005] τ0/T ~ unif[0, 0.1] α optimized

slide-32
SLIDE 32

31

  • N-dimensional estimation problem,
  • ML estimation prohibitively complex.

Instead:

  • Linearize the PLL-based system,
  • Apply projection operator.

Random walk: Linearization and Projection

slide-33
SLIDE 33

32

TED equation: Define: Therefore, we get the following linear Gaussian model:

  • Output yk is the sum of the PLL and the TED outputs.
  • Validity of model depends on linearity of TED characteristics.
  • is an estimate based on previous observations (a priori).
  • yk is based on previous and present observsations (a posteriori).

ε ˆk εk nk + τk τ ˆk – nk + = = yk τ ˆk ε ˆk + = yk τk nk + = τ ˆk

Linear Gaussian model from PLL

slide-34
SLIDE 34

33

For the linear Gaussian model, the MAP estimator is where

  • y is the vector of a posteriori observations
  • is the covariance matrix of the timing offset vector
  • is the variance of the noise nk

τ ˆmap y ( ) Kτ σn

2I

+ ( )

1 – Kτy

= Kτ τ σn

2

MAP estimator

slide-35
SLIDE 35

34

5 10 15 20 10

−4

10

−3

SNR (dB) Timing Error Variance CRB Trained MAP Trained PLL

  • 5.5 dB gain over PLL.
  • 1.5 dB away from CRB.
  • CRB not attainable with the timing model chosen. (The a posteriori den-

sity needs to be Gaussian, which is not the case here.)

  • Gap partly due to loss due to linearization of the TED characteristics.

f θ r ( )

1000 packets Parameters N = 500 σw/T = 0.33% α optimized

MAP estimator: Performance

random walk model first order PLL MM TED

slide-36
SLIDE 36

35

MAP estimator takes the form of a matrix operation: Using the structure of the matrices involved, we can rewrite this as where

  • is a convolution matrix,

⇒ implemented as a time-invariant filter,

  • is diagonal matrix with different diagonal entries,

⇒ implemented as time-varying scaling of the filter output.

τ ˆmap y ( ) Kτ σn

2I

+ ( )

1 – Kτy

= τ ˆmap y ( ) A1A2y ≈ A2 A1

MAP estimator: Reduced-complexity Implementation

slide-37
SLIDE 37

36

  • Conventional timing recovery based on the PLL.
  • Cramer-Rao bound gives a bound on performance of any timing estimator.
  • Derived the CRB for different timing offset models.
  • PLL does not achieve the CRB.
  • With constant offset, gradient descent achieves the CRB.
  • With frequency offset, the Levenberg-Marquardt method achieves the CRB.
  • With a random walk, the MAP estimator significantly outperforms the CRB.

(Caveat: With a random walk, the CRB is not achievable.)

Summary

slide-38
SLIDE 38

37

Thank you! Questions?