Tracking Perform ance of the MMax Conjugate Gradient Algorithm Bei - - PowerPoint PPT Presentation

tracking perform ance of the mmax conjugate gradient
SMART_READER_LITE
LIVE PREVIEW

Tracking Perform ance of the MMax Conjugate Gradient Algorithm Bei - - PowerPoint PPT Presentation

Tracking Perform ance of the MMax Conjugate Gradient Algorithm Bei Xie and Tam al Bose Wireless@VT Bradley Dept. of Electrical and Computer Engineering Virginia Tech Outline Motivation Background Conjugate Gradient (CG) Algorithm


slide-1
SLIDE 1

Tracking Perform ance of the MMax Conjugate Gradient Algorithm

Bei Xie and Tam al Bose Wireless@VT Bradley Dept. of Electrical and Computer Engineering

Virginia Tech

slide-2
SLIDE 2

Outline

 Motivation  Background

  • Conjugate Gradient (CG) Algorithm
  • Partial Update Methods

 Partial Update CG Algorithm

  • Tracking Performance Analysis
  • Simulations

 Summary

slide-3
SLIDE 3

Motivation

 How to reduce the computational complexity of an adaptive filter? Solutions: Using partial update (PU) methods.  What are the tracking performance by using partial update methods to CG?

slide-4
SLIDE 4

Conjugate Gradient Algorithm

Solve system with form Equivalent to find a w to minimize the cost function The gradient of the cost function is: The residual vector is defined as:

b R  w

 

b w Rw w w ) ( ) ( ) ( 2 1 ) ( n n n n J

T T

 

 

b R w

w

   ) ( ) ( n n J w

 

) ( ) ( ) ( n n J n w R b w g

w

   

slide-5
SLIDE 5

A line search method for minimizing the cost function has the form: p(n) is the direction vector CG chooses the direction is conjugately

  • rthogonal to the previous directions

) ( ) 1 ( ) ( n n n p w w    

m n m n

T

  , ) ( ) ( Rp p

slide-6
SLIDE 6

Now we find to minimize The residual vector is also equal to The residual vector is orthogonal to the previous direction vectors,

 

) ( ) 1 ( n n J p w   

   

) ( ) ( ) ( ) 1 ( ) ( ) ( ) ( ) 1 (         b p p w n n n n n n n n J

T T

p p w R  

 

) ( ) ( ) 1 ( ) ( ) ( ) ( ) 1 ( ) ( ) ( n n n n n n n n n

T T T T

p p g p p p w b p R R R      

 

) ( ) ( ) 1 ( ) ( ) ( ) 1 ( w ) ( w ) ( n n n n n n n n Rp g p R b R b g           

m n m n

T

  , ) ( ) ( p g

slide-7
SLIDE 7

CG chooses the direction in the form of Use Polak-Ribière (PR) method,

) ( ) ( ) ( ) 1 ( n n β n n p g p   

 

) 1 ( ) 1 ( ) ( ) 1 ( ) ( ) (      n n n n n n β

T T

g g g g g

PR method is chosen because it is a non-reset method and performs better for non-constant matrix R CG with PR method usually converges faster than Fletcher- Reeves (FR) method

slide-8
SLIDE 8

CG Algorithm in adaptive filter system

A basic adaptive filter system model is d(n) is the desired signal x(n)=[x(n),x(n-1),…,x(n-N+1)]T is the input data vector of an unknown system w *=[w1

*,w2 *,…,wN *]T is the impulse response

vector of the unknown system w * is constant for time-invariant system w * changes for time-varying sytem v(n) is a white noise

) ( ) ( ) (

*

n v n n d

T

  w x

slide-9
SLIDE 9

To estimate the R and b in The exponentially decaying data window is used

b R  w

) ( ) ( ) 1 ( ) ( ) ( ) ( ) ( ) ( ) 1 ( ) ( ) ( ) ( n n d n i i d n n n n i i n

n i i n T n i T i n

x b x b x x R x x R        

 

   

   

λ is the forgetting factor

slide-10
SLIDE 10

 

 

) ( ) ( ) ( ) 1 ( ) 1 ( ) 1 ( ) ( ) 1 ( ) ( ) ( ) 1 ( ) ( ) ( ) ( ) ( ) ( ) ( ) 1 ( ) ( ) ( ) ( ) ( ) ( ) 1 ( ) ( 5 . , ) ( ) ( ) ( ) 1 ( ) ( ) ( ) ( ) ( ) 1 ( ) ( n n β n n n n n n n n β n n n d n n n n n n n n n n n n n n n n n n n n n n

T T T T T T

p g p g g g g g w x x p R g R b g p w w p p g p x x R R                                     w R

The CG algorithm in an adaptive filter system is summarized as: Initial conditions:

w (0)=0 , R(0)=0 , p(1)=g(0) η is used to guarantee convergence

slide-11
SLIDE 11

Partial Update (PU) Methods

 Update part of the weights to save the

computational complexity

 Each update step, update M<N coefficients  Basic PU methods include periodic,

sequential, stochastic, and MMax methods

  • The periodic method: update the

weights at every Sth iteration and copy the weights at the other iterations, where

       M N S

slide-12
SLIDE 12
  • The sequential method: choose the

subset of the weights in a round-robin fashion.

  • The stochastic method: is a randomized

version of the sequential method. Usually a uniformly distributed random process will be applied.

  • The MMax method: the elements of the

weight w are updated according to the position of the M largest elements of the input vector x(n).

slide-13
SLIDE 13

 

 

) ( ) ( ) ( ) 1 ( ) 1 ( ) 1 ( ) ( ) 1 ( ) ( ) ( ) 1 ( ) ( ) ( ) ( ˆ ) ( ) ( ˆ ) ( ) 1 ( ) ( ) ( ) 1 ( ) ( 5 . , ) ( ) ( ˆ ) ( ) 1 ( ) ( ) ( ) ( ˆ ) ( ˆ ) 1 ( ˆ ) ( ˆ n n β n n n n n n n n β n n n d n n n n n n n n n n n n n n n n n n n

T T T T T T

p g p g g g g g w x x p R g g p w w p R p g p x x R R                                  

Partial Update CG Algorithm

The partial update CG algorithm in an adaptive filter system is summarized as:

slide-14
SLIDE 14

  1 , ) ( , ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ˆ

1 2 1

               

n i M n i n i n i n i n n n n

k N k k N M M

       I x I x

The number of multiplications of CG is 3N2+10N+3 per sample The number of multiplications of PU CG is 2N2+M2+9N+M+3 per sample

slide-15
SLIDE 15

The MMax method: the elements of the input x are chosen according to the position of the M largest elements of the input vector x(n)

 

    

 

  • therwise

M n n if n i

l N l k k

, ) ( max ) ( 1 ) (

1

x x

SORTLINE method is used for comparison Mmax CG needs 2 + 2log2N comparisons

slide-16
SLIDE 16

Desired signal becomes: Time-varying system w *(n) uses a first-order Markov model is very close to unity is process noise

Tracking Performance Analysis of PU CG

) ( ) ( w ) ( ) (

*

n v n n n d

T

  x ) ( ) 1 ( w ) ( w

* *

n n n     

) (n 

slide-17
SLIDE 17

Assumptions:

  • The coefficient error w (n)-w *(n) is small

and independent of the input signal x(n) at steady state

  • White noise v(n) is independent of the

input signal x(n) and is independent of process noise η(n)

  • Input signal x(n) is independent of both

v(n) and η(n)

slide-18
SLIDE 18

At steady state, the MSE of PU CG with correlated input is

 

 

                       

  

      R R R R R w x

2 2 1 2 2 2 2

1 ~ ˆ ~ 1 1 ) ( ) ( ) ( ) (

T v v T

tr n n n d E n e E

     

) ( ˆ ) ( ˆ ˆ ) ( ) ( ˆ ) 1 ( ~ ) ( ~ ) ( ~ ) 1 ( ) ( ) ( ˆ ~ ) ( ) ( n n E n n n n n n n E n n E

T T T T

x x R x x R R R x x R x x R          

slide-19
SLIDE 19

 

 

) ( ) ~ ( ) ˆ ( ) ( ) ( 1 1 ) 1 ( ) (

2 2 2 ~ 2 ˆ 2 2 2 2 4 ~ 2 ˆ 2 2 2 2

n v E tr tr tr tr N n e E

v x x x x x x x v v

         

             

R R R R

At steady state, the MSE of PU CG with white input is

Variance of noise

slide-20
SLIDE 20

 

 

) ( 1 1 ) 1 ( ) (

2 2 2 2 2 2 

        R tr N n e E

x v v

     

For MMax method and white input, κ <1, κ is close to 1

2 2 ~ 2 2 ˆ

,

x x x x

     

For white process noise η(n)

 

 

2 2 2 2 2 2 2

1 1 ) 1 ( ) (

        

x v v

N n e E      

slide-21
SLIDE 21

Simulations

System identification model The initial impulse response of unknown system is 16-order (N=16) FIR filter w*(n)=[0.01,0.02,-0.04,-0.08,0.15,-0.3,0.45, 0.6, 0.6,0.45,-0.3,0.15,-0.08,-0.04,0.02,0.01]T

slide-22
SLIDE 22

The variance of the input noise σv

2=0.0001

Parameter λ=0.9 and η=0.6 White input, variance is 1 in Markov model is 0.9998

slide-23
SLIDE 23

Tracking Performance of MMax CG

Comparison of MSE of MMax CG for varying process noise η, M=8 Comparison of MSE of MMax CG for varying process noise η, M=4

500 1000 1500 2000 2500 3000 3500 4000

  • 50
  • 40
  • 30
  • 20
  • 10

10 20 30

samples MSE(dB)

MMax CG, M=8, =0.0001 MMax CG, M=8, =0.001 MMax CG, M=8, =0.01 500 1000 1500 2000 2500 3000 3500 4000

  • 50
  • 40
  • 30
  • 20
  • 10

10 20 30

samples MSE(dB)

MMax CG, M=4, =0.0001 MMax CG, M=4, =0.001 MMax CG, M=4, =0.01

slide-24
SLIDE 24
  • The MSE of MMax CG increases when the

process noise increases.

  • The variance of the MSE increases when

the process noise increases.

  • The partial update length does not have

much effect on the MSE results.

  • The partial update length only affects the

convergence rate. The convergence rate decreases as the partial update length decreases.

slide-25
SLIDE 25

Process noise ση Simulated MSE (dB) Theoretical MSE (dB)

0.0001

  • 39.2381
  • 39.3584

0.001

  • 32.9019
  • 32.9019

0.01

  • 13.4403
  • 11.0287

The theoretical results match the simulated results.

Table 1 . The simulated MSE and theoretical MSE of MMax CG for varying process noise η, M = 8. Table 2 . The simulated MSE and theoretical MSE of MMax CG for varying process noise η, M = 4.

Process noise ση Simulated MSE (dB) Theoretical MSE (dB)

0.0001

  • 38.9965
  • 39.0672

0.001

  • 31.2768
  • 30.4378

0.01

  • 11.6397
  • 11.0282
slide-26
SLIDE 26

Performance comparison between MMax CG and MMax RLS

After 2000 samples/iterations pass, the unknown system is changed by multiplying all coefficients by -1.

500 1000 1500 2000 2500 3000 3500 4000

  • 50
  • 40
  • 30
  • 20
  • 10

10 20 30

samples MSE(dB)

CG MMax CG M=4 RLS MMax RLS M=4

500 1000 1500 2000 2500 3000 3500 4000

  • 50
  • 40
  • 30
  • 20
  • 10

10 20 30

samples MSE(dB) CG MMax CG M=8 RLS MMax RLS M=8

Comparison of MSE of MMax CG with CG, RLS, MMax RLS for white input, N=16, M=8. Comparison of MSE of MMax CG with CG, RLS, MMax RLS for white input, N=16, M=4. The partial update length only affects the convergence rate at the beginning in this case.

slide-27
SLIDE 27

Algorithm s Num ber of m ultiplications per sym bol Num ber of com parisons per sym bol CG (N=16) 3003

  • MMax CG (N=8)

731 10 MMax CG (N=4) 679 10 RLS (N=16) 3721

  • MMax RLS (N=8)

825 10 MMax RLS (N=4) 693 10

Table 3 . The computational complexities of CG, MMax CG, RLS, and MMax RLS.

slide-28
SLIDE 28

Sum m ary

 The tracking performance of the MMax CG is analyzed  Theoretical mean-square performance is derived for white and correlated inputs  The tracking performance of MMax CG is compared with CG, RLS, MMax RLS by using computer simulations  The MMax CG algorithm can achieve similar performance to the full-update CG while reducing computational complexity significantly  The MMax CG algorithm can achieve similar performance to the MMax RLS while having lower computational complexity