Order Recursion Introduction Introduction Rc = d Order versus Time - - PowerPoint PPT Presentation

order recursion introduction introduction rc d order
SMART_READER_LITE
LIVE PREVIEW

Order Recursion Introduction Introduction Rc = d Order versus Time - - PowerPoint PPT Presentation

Order Recursion Introduction Introduction Rc = d Order versus Time Updates There are many ways to solve the normal equations Matrix Inversion by Partitioning Lemma Solutions are mathematically equivalent Levinson Algorithm


slide-1
SLIDE 1

Order Updates Rmcm = dm

  • Suppose we wish to solve the normal equations for a range of M

values

  • May help select the final model order
  • If we have solved for the optimal cm ∈ Cm×1, can we more

efficiently solve for cm+1?

  • This is called an order update
  • In this set of slides I have dropped the o subscript to simplify

notation

  • We will only discuss the optimal parameter vector
  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

3

Order Recursion

  • Introduction
  • Order versus Time Updates
  • Matrix Inversion by Partitioning Lemma
  • Levinson Algorithm
  • Interpretations
  • Examples
  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

1

Time Updates R(n)c(n) = d(n)

  • Suppose we are estimating rx(ℓ) and ryx(ℓ) from a finite data

segment, x(n) for n ∈ {0, 1, . . . , N − 1}

  • Let’s say we have solved the normal equations for this data set
  • When a new observation is obtained, can we solve the updated

normal equations more efficiently by using the solution for the first N points?

  • Called time updates
  • The dual problem of order updates
  • These techniques are essentially adaptive filters (chapter 10)
  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

4

Introduction Rc = d

  • There are many ways to solve the normal equations
  • Solutions are mathematically equivalent
  • May not be numerically equivalent
  • May differ in

– Numeric stability – Order of computation – Usefulness of intermediate quantities: partial autocorrelation coefficients – Studying properties of solution: minimum phase,

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

2

slide-2
SLIDE 2

Linear Combiner: Notation In this set of notes and m subscript is used to indicate the size of matrices and vectors Rm+1 = E

  • xm

xm+1 xH

m

x∗

m+1

  • =

Rm rbm rH

bm

ρbm

  • dm+1 = E [xm+1y∗] = E

xm xm+1

  • y∗
  • =

dm dm+1

  • where

rbm E

  • xmx∗

m+1

  • ρbm E
  • |xm+1|2
  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

7

Scope of Order Updates Rmcm = dm

  • The text covers all of the cases we have discussed

– Any linear estimation problem – Optimum FIR filters – Optimum FIR filters for stationary processes

  • While the solutions are insightful for the first two cases, they are

no more efficient computationally

  • We will only discuss the last case
  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

5

Inversion of Partitioned Hermitian Matrices

  • Suppose we know R−1

m and we wish to compute R−1 m+1 efficiently

so that we can solve the normal equations Rm+1cm+1 = dm+1

  • The inverse of a Hermitian matrix is also Hermitian
  • We are not assuming Rm is Toeplitz yet

Qm+1 R−1

m+1 =

Qm qm qH

m

qm

  • = QH

m+1

Rm+1Qm+1 = Rm rbm rH

bm

ρbm Qm qm qH

m

qm

  • =

Im 0m 0H

m

1

  • RmQm + rbmqH

m = Im

rH

bmQm + ρbmqH m = 0H m

Rmqm + rbmqm = 0m rH

bmqm + ρbmqm = 1

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

8

Stationary Case Rc = d

  • When the process is stationary, R is Toeplitz
  • This permits more efficient algorithms to be used
  • In general the inversion of an M × M matrix requires O(M 3)
  • perations
  • When R is Toeplitz we can

– Invert the matrix in O(M 2) operations – Perform an LDLH decomposition in O(M 2) operations

  • See Section 7.7 for details
  • We will consider both the stationary FIR filter and nonstationary

linear combiner cases

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

6

slide-3
SLIDE 3

Matrix Inversion by Partitioning Lemma qm = 1 αbm qm = bm αbm Qm = R−1

m + bmbH m

αbm We can then express R−1

m+1 as

R−1

m+1 =

Qm qm qH

m

qm

  • =

⎡ ⎣

  • R−1

m + bmbH

m

αbm

  • bm

αbm bH

m

αbm 1 αbm

⎤ ⎦ = R−1

m

0m 0H

m

  • +

1 αbm bm 1 bH

m

1

  • This is a rank-one modification
  • Called the matrix inversion by partitioning lemma
  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

11

Inversion of Partitioned Hermitian Matrices Continued RmQm + rbmqH

m = Im

rH

bmQm + ρbmqH m = 0H m

Rmqm + rbmqm = 0m rH

bmqm + ρbmqm = 1

We only have three unknowns, so only three of these equations are needed to obtain the solutions. qm = −R−1

m rbmqm

qm = 1 ρbm − rH

bmR−1 m rbm

qm = −R−1

m rbm

ρbm − rH

bmR−1 m rbm

Qm = R−1

m − R−1 m rbmqH m = R−1 m + R−1 m rbm(−R−1 m rbm)H

ρbm − rH

bmR−1 m rbm

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

9

Matrix Inversion Counterpart Following similar steps, we can show R−1

m+1

  • ρfm

rH

fm

rfm Rfm −1 = 0H

m

0m R−1

fm

  • +

1 αfm

  • 1

am 1 aH

m

  • where

am −R−1

fmrfm

αfm ρfm + rH

fmR−1 fmrfm

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

12

Inversion of Partitioned Hermitian Matrices Continued qm = 1 ρbm − rH

bmR−1 m rbm

qm = −R−1

m rbm

ρbm − rH

bmR−1 m rbm

Qm = R−1

m + R−1 m rbm(−R−1 m rbm)H

ρbm − rH

bmR−1 m rbm

Now if we define bm −R−1

m rbm

αbm ρbm − rH

bmR−1 m rbm = ρbm + rH bmbm

The we can simplify the expressions for the components of Qm+1 as qm = 1 αbm qm = bm αbm Qm = R−1

m + bmbH m

αbm

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

10

slide-4
SLIDE 4

Order Updates of Parameter Vectors cm+1 = R−1

m+1dm+1

= R−1

m

0m 0H

m

dm dm+1

  • +

1 αbm

  • bm

1 bH

m

1 dm dm+1

  • =
  • R−1

m dm

  • +
  • bm

1 bH

mdm + dm+1

αbm =

  • cm
  • +
  • bm

1

  • kcm

where kcm βcm αbm βcm bH

mdm + dm+1

Here the c subscript is presumably an indicator that these coefficients are for updating the parameter vector c

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

15

Interpretations ebm = xm+1 − ˆ xm+1 = xm+1 + bH

m

  • x1

x2 . . . xm T efm = x1 − ˆ x1 = x1 + aH

m

x2 x3 . . . xm+1 T and Pbm = ρbm − rH

bmR−1 m rbm= αbm

Pfm = ρfm − rH

fmR−1 fmrfm = αfm

  • The vector bm is the MMSE estimator of the last element of

xm+1 from the first m elements

  • Similarly, am is the MMSE estimator of the first element of xm+1

from the last m elements

  • If xm+1 =

x(n) x(n − 1) . . . x(n − m) , then these are the forward (FLP) and backward (BLP) linear predictors

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

13

Order Updates of Parameter Vectors Discussion bm = −R−1

m rbm

αbm = ρbm + rH

bmbm

βcm bH

mdm + dm+1

kcm βcm αbm cm+1 = cm

  • +

bm 1

  • kcm

R−1

m+1 =

R−1

m

0m 0H

m

  • +

1 αbm bm 1 bH

m

1

  • The update equation for cm+1 is called a Levinson recursion
  • If we know bm, we can determine cm+1
  • However, solving for bm requires a matrix-vector product which

requires O(m2) operations for the update to cm+1

  • Contrary to the text, this approach is more efficient than solving

for cm+1 = R−1

m+1dm+1 directly, which would require O(m3)

  • perations
  • We can do even better if Rm+1 is Toeplitz
  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

16

Hermitian Inversion Summary bm = −R−1

m rbm

αbm = ρbm + rH

bmbm

R−1

m+1 =

R−1

m

0m 0H

m

  • +

1 αbm

  • bm

1 bH

m

1

  • We have shown that we can obtain the inverse of Rm+1 from the

inverse of Rm with a rank one update

  • Requires the R−1

m and αbm be invertible

  • Applies to any Hermitian matrix
  • The update requires O(m2) operations
  • Is more efficient than solving the normal equations repeatedly for

every m of interest

  • Is not more efficient if only cM is of interest
  • But how is cm+1 is related to cm?
  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

14

slide-5
SLIDE 5

Stationary Case: Parameter Updates cm+1 =

  • cm
  • +
  • bm

1

  • kcm

kcm βcm Pbm bm = −R−1

m Jrm

βcm bH

mdm + dm+1

= (−R−1

m Jrm)Hdm + dm+1

= −rH

mJR−1 m dm + dm+1

= −rH

mJcm + dm+1

  • The text has the wrong expression for this last equality
  • So far we haven’t gained anything
  • The most expensive operation computationally is still solving for

bm and requires O(m2) operations

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

19

Stationary Case: Notation When x(n) and y(n) are jointly stationary and xm+1(n) =

  • x(n)

x(n − 1) . . . x(n − m)

  • rm
  • r(1)

r(2) . . . r(m) T and Rm+1 is Toeplitz Rm+1 = ⎡ ⎢ ⎢ ⎢ ⎣ r(0) . . . r(m − 1) r(m) . . . ... . . . . . . r∗(m − 1) . . . r(0) r(1) r∗(m) . . . r∗(1) r(0) ⎤ ⎥ ⎥ ⎥ ⎦ =

  • r(0)

rT

m

r∗

m

Rm

  • =
  • ρfm

rH

fm

rfm Rm

  • =

Rm Jrm rH

mJ

r(0)

  • =

Rm rbm rH

bm

ρbm

  • dm+1 =
  • dm

dm+1

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

17

Stationary Case: The Trick Recall that in the stationary case the FLP and BLP are related bm = Ja∗

m

Pm Pbm = Pfm Hence, we can trivially determine bm from am and vice versa Suppose now that we solve for the FLP coefficients. In this case y(n) = x(n + 1) cm = am dm = r(1) . . . r(m − 1)T = rm The recursive expression for the optimal FLP coefficients is then am+1 =

  • am
  • +
  • bm

1

  • km =
  • am
  • +
  • Ja∗

m

1

  • km

where km − βm Pm βm bH

mr∗ m + r∗(m + 1) = aT mJr∗ m + r∗(m + 1)

Cool!

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

20

Stationary Case: Autocorrelation Matrix Inversion From the matrix inversion by partitioning lemma, we have R−1

m+1 =

R−1

m

0m 0H

m

  • +

1 Pbm

  • bm

1 bH

m

1

  • bm = −R−1

m rbm

= −R−1

m Jrm

Pbm = αbm = ρbm + rH

bmbm

= r(0) + rH

mJbm

= Pfm

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

18

slide-6
SLIDE 6

Levinson Algorithm Given {r(ℓ)}M

0 , {dm}M 0 , Py, initialize the algorithm with

P−1 = r(0) β−1 = 0 k∗

m−1 = 0

a0 = ∅ For m = 0, 1, . . . , M − 1 rm = r(1) r(2) . . . r(m)T βm = aT

mJr∗ m + r∗(m + 1)

Pm = Pm−1 + βm−1k∗

m−1

km = − βm Pm am+1 =

  • am
  • +
  • Ja∗

m

1

  • km

βcm = −cH

mJrm + dm+1

kcm = βcm Pm cm+1 =

  • cm
  • +
  • Ja∗

m

1

  • kcm

Pc(m+1) = Pcm + βcmk∗

cm

See Table 7.2 in text for more orderly presentation of the algorithm.

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

23

Stationary Case: BLP/FLP MMSE Pm = r(0) + rH

ma∗ m = r(0) + aT mrm

Pm+1 = r(0) + rH

m+1a∗ m+1

= r(0) +

  • rH

m

r∗(m + 1) am

  • +
  • Ja∗

m

1

  • km

∗ = r(0) + rH

ma∗ m +

  • rH

mb∗ m + r∗(m + 1)

  • k∗

m

= Pm + βmk∗

m

= Pm + β∗

mkm

= Pm + (−kmPm)∗km = Pm(1 − |km|2) The BLP MMSE therefore is only decreased as we increase m in the stationary case

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

21

Levinson and Levinson-Durbin Algorithms

  • The Levinson algorithm consists of two parts
  • 1. Recursions for the optimum FLP/BLP
  • 2. Recursions for the optimum filter
  • If the application is only FLP or BLP, the algorithm can be

simplified

  • This simplified algorithm is called the Levinson-Durbin algorithm

(see text for details)

  • There are other algorithms for calculating the optimal lattice filter

coefficients under the same conditions

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

24

Stationary Case: FIR Filter MMSE Similarly, kcm βcm Pm Pc(m+1) = Pcm − β∗

cmkcm = Pcm − |βcm|2

Pm ≤ Pcm Thus the MMSE of the FIR filter can also only decrease as m increases!

  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

22

slide-7
SLIDE 7

Remaining

  • Relationship to partial correlation
  • Example
  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

27

Partial Correlation pc(y, xm) = E

  • (x(n − m) − ˆ

x(n − m)) (y(n) − ˆ y(n))∗ = E [ebm(n)e∗

m(n)]

= bmdm + dm+1 = βcm

  • Recall that the partial correlation is defined as the correlation

between x(n − m) and y(n) after the “influence” of the intermediate variables {x(n), x(n − 1), . . . , x(n − m + 1)} has been removed

  • That is we have built estimators for x(n − m) and y(n) and

calculate the correlation of the residuals

  • The residual for x(n − m) is the backward prediction error
  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

25

Equivalent Solutions for Optimum Linear Prediction

  • There are three equivalent representations of optimum linear

predictors of a stationary process – Direct-form filter structure: {PM, a1, a2, . . . , aM} – Lattice filter structure: {PM, k0, k1, . . . , kM−1} – Autocorrelation sequence: {r(0), r(1), . . . , r(M)}

  • It is possible to convert from any set of these parameters to any
  • ther set
  • Covered in Section 7.6, but we won’t discuss in lecture
  • J. McNames

Portland State University ECE 539/639 Order Recursion

  • Ver. 1.01

26