parameter estimation
play

Parameter Estimation Saravanan Vijayakumaran sarva@ee.iitb.ac.in - PowerPoint PPT Presentation

Parameter Estimation Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay October 25, 2012 1 / 44 Motivation System Model used to Derive Optimal Receivers s ( t ) y ( t )


  1. Parameter Estimation Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay October 25, 2012 1 / 44

  2. Motivation

  3. System Model used to Derive Optimal Receivers s ( t ) y ( t ) Channel y ( t ) = s ( t ) + n ( t ) s ( t ) Transmitted Signal y ( t ) Received Signal n ( t ) Noise Simplified System Model. Does Not Account For • Propagation Delay • Carrier Frequency Mismatch Between Transmitter and Receiver • Clock Frequency Mismatch Between Transmitter and Receiver In short, Lies! Why? 3 / 44

  4. You want answers? 4 / 44

  5. I want the truth! 5 / 44

  6. You can't handle the truth! . . . right at the beginning of the course. Now you can. 6 / 44

  7. Why Study the Simplified System Model? s ( t ) y ( t ) Channel y ( t ) = s ( t ) + n ( t ) • Receivers estimate propagation delay, carrier frequency and clock frequency before demodulation • Once these unknown parameters are estimated, the simplified system model is valid • Then why not study parameter estimation first? • Hypothesis testing is easier to learn than parameter estimation • Historical reasons 7 / 44

  8. Unsimplifying the System Model Effect of Propagation Delay • Consider a complex baseband signal ∞ � b n p ( t − nT ) s ( t ) = n = −∞ and the corresponding passband signal � √ 2 s ( t ) e j 2 π f c t � s p ( t ) = Re . • After passing through a noisy channel which causes amplitude scaling and delay, we have y p ( t ) = As p ( t − τ ) + n p ( t ) where A is an unknown amplitude, τ is an unknown delay and n p ( t ) is passband noise 8 / 44

  9. Unsimplifying the System Model Effect of Propagation Delay • The delayed passband signal is � √ 2 s ( t − τ ) e j 2 π f c ( t − τ ) � s p ( t − τ ) = Re � √ 2 s ( t − τ ) e j θ e j 2 π f c t � = Re where θ = − 2 π f c τ mod 2 π . For large f c , θ is modeled as uniformly distributed over [ 0 , 2 π ] . • The complex baseband representation of the received signal is then y ( t ) = Ae j θ s ( t − τ ) + n ( t ) where n ( t ) is complex Gaussian noise. 9 / 44

  10. Unsimplifying the System Model Effect of Carrier Offset • Frequency of the local oscillator (LO) at the receiver differs from that of the transmitter • Suppose the LO frequency at the transmitter is f c � √ 2 s ( t ) e j 2 π f c t � s p ( t ) = Re . • Suppose that the LO frequency at the receiver is f c − ∆ f • The received passband signal is y p ( t ) = As p ( t − τ ) + n p ( t ) • The complex baseband representation of the received signal is then y ( t ) = Ae j ( 2 π ∆ ft + θ ) s ( t − τ ) + n ( t ) 10 / 44

  11. Unsimplifying the System Model Effect of Clock Offset • Frequency of the clock at the receiver differs from that of the transmitter • The clock frequency determines the sampling instants at the matched filter output • Suppose the symbol rate at the transmitter is 1 T symbols per second • Suppose the receiver sampling rate is 1 + δ symbols per T second where | δ | ≪ 1 and δ may be positive or negative • The actual sampling instants and ideal sampling instants will drift apart over time 11 / 44

  12. The Solution Estimate the unknown parameters τ , θ , ∆ f and δ Timing Synchronization Estimation of τ Carrier Synchronization Estimation of θ and ∆ f Clock Synchronization Estimation of δ Perform demodulation after synchronization 12 / 44

  13. Parameter Estimation

  14. Parameter Estimation • Hypothesis testing was about making a choice between discrete states of nature • Parameter or point estimation is about choosing from a continuum of possible states Example Consider the complex baseband signal below y ( t ) = Ae j θ s ( t − τ ) + n ( t ) • The phase θ can take any real value in the interval [ 0 , 2 π ) • The amplitude A can be any real number • The delay τ can be any real number 14 / 44

  15. System Model for Parameter Estimation • Consider a family of distributions Y ∼ P θ , θ ∈ Λ where the observation vector Y ∈ Γ ⊆ R n for n ∈ N and Λ ⊆ R m is the parameter space • Example: Y = A + N where A is an unknown parameter and N is a standard Gaussian RV • The goal of parameter estimation is to find θ given Y • An estimator is a function from the observation space to the parameter space ˆ θ : Γ → Λ 15 / 44

  16. Which is the Optimal Estimator? • Assume there is a cost function C which quantifies the estimation error C : Λ × Λ → R such that C [ a , θ ] is the cost of estimating the true value of θ as a • Examples of cost functions Squared Error C [ a , θ ] = ( a − θ ) 2 Absolute Error C [ a , θ ] = | a − θ | � 0 if | a − θ | ≤ ∆ Threshold Error C [ a , θ ] = 1 if | a − θ | > ∆ 16 / 44

  17. Which is the Optimal Estimator? • With an estimator ˆ θ we associate a conditional cost or risk conditioned on θ � � �� R θ (ˆ ˆ θ ) = E θ C θ ( Y ) , θ • Suppose that the parameter θ is the realization of a random variable Θ • The average risk or Bayes risk is given by � � r (ˆ R Θ (ˆ θ ) = E θ ) • The optimal estimator is the one which minimizes the Bayes risk 17 / 44

  18. Which is the Optimal Estimator? • Given that � � � � � � �� � R θ (ˆ ˆ ˆ � θ ) = E θ C θ ( Y ) , θ = E C θ ( Y ) , Θ � Θ = θ � the average risk or Bayes risk is given by � � �� r (ˆ ˆ θ ) = E C θ ( Y ) , Θ � � � � �� � ˆ � = E E C θ ( Y ) , Θ � Y � • The optimal estimate for θ can be found by minimizing for each Y = y the posterior cost � � � � � ˆ � E C θ ( y ) , Θ � Y = y � 18 / 44

  19. Minimum-Mean-Squared-Error (MMSE) Estimation • C [ a , θ ] = ( a − θ ) 2 • The posterior cost is given by � � � � 2 � (ˆ ˆ θ ( y ) − Θ) 2 � E � Y = y = θ ( y ) � � � � − 2 ˆ � θ ( y ) E Θ � Y = y � � � � Θ 2 � + E � Y = y � • The Bayes estimate is given by � � � ˆ � θ MMSE ( y ) = E Θ � Y = y � 19 / 44

  20. Example 1: MMSE Estimation • Suppose X and Y are jointly Gaussian random variables • Let the joint pdf be given by � � 1 − 1 2 ( s − µ ) T Σ − 1 ( s − µ ) p XY ( x , y ) = exp 1 2 π | Σ | 2 � σ 2 � x � � µ x � � ρσ x σ y x where s = , µ = and Σ = σ 2 y µ y ρσ x σ y y • Suppose Y is observed and we want to estimate X • The MMSE estimate of X is � � � ˆ � X MMSE ( y ) = E � Y = y X � 20 / 44

  21. Example 1: MMSE Estimation • The conditional distribution of X given Y = y is a Gaussian RV with mean µ X | y = µ x + σ x ρ ( y − µ y ) σ y and variance σ 2 X | y = ( 1 − ρ 2 ) σ 2 x • Thus the MMSE estimate of X given Y = y is X MMSE ( y ) = µ x + σ x ˆ ρ ( y − µ y ) σ y 21 / 44

  22. Example 2: MMSE Estimation • Suppose A is a Gaussian RV with mean µ and known variance v 2 • Suppose we observe Y i , i = 1 , 2 , . . . , M such that Y i = A + N i where N i ’s are independent Gaussian RVs with mean 0 and known variance σ 2 • Suppose A is independent of the N i ’s • The MMSE estimate is given by Mv 2 σ 2 ˆ A 1 ( y ) + µ ˆ A MMSE ( y ) = Mv 2 σ 2 + 1 where ˆ � M A 1 ( y ) = 1 i = 1 y i M 22 / 44

  23. Minimum-Mean-Absolute-Error (MMAE) Estimation • C [ a , θ ] = | a − θ | • The Bayes estimate ˆ θ ABS is given by the median of the posterior density p (Θ | Y = y ) � � � � � � , t < ˆ � � Pr Θ < t � Y = y ≤ Pr Θ > t � Y = y θ ABS ( y ) � � � � � � � � , t > ˆ � � Pr Θ < t � Y = y ≥ Pr Θ > t � Y = y θ ABS ( y ) � � p ( θ | Y = y ) � � � � Pr Θ < t � Y = y � � � � � Pr Θ > t � Y = y � t ˆ θ ABS ( y ) 23 / 44

  24. Minimum-Mean-Absolute-Error (MMAE) Estimation � ∞ • For Pr [ X ≥ 0 ] = 1, E [ X ] = 0 Pr [ X > x ] dx • Since | ˆ θ ( y ) − Θ | ≥ 0 � � � | ˆ � E θ ( y ) − Θ | � Y = y � � ∞ � � � | ˆ � = θ ( y ) − Θ | > x � Y = y Pr dx � 0 � ∞ � � � Θ > x + ˆ � = Pr θ ( y ) � Y = y dx � 0 � ∞ � � � Θ < − x + ˆ � + Pr θ ( y ) � Y = y dx � 0 � ∞ � � � � = Pr Θ > t � Y = y dt � ˆ θ ( y ) � ˆ θ ( y ) � � � � + Pr Θ < t � Y = y dt � −∞ 24 / 44

  25. Minimum-Mean-Absolute-Error (MMAE) Estimation � � � | ˆ wrt to ˆ � θ ( y ) − Θ | � Y = y θ ( y ) Differentiating E � � ∂ � � | ˆ � E θ ( y ) − Θ | � Y = y � ∂ ˆ θ ( y ) � ∞ � � � ∂ � = Pr Θ > t � Y = y dt � ∂ ˆ θ ( y ) ˆ θ ( y ) � ˆ θ ( y ) � � � ∂ � + Pr Θ < t � Y = y dt � ∂ ˆ θ ( y ) −∞ � � � � � � Θ < ˆ Θ > ˆ � � − Pr = Pr θ ( y ) � Y = y θ ( y ) � Y = y � � • The derivative is nondecreasing tending to − 1 as θ ( y ) → −∞ and + 1 as ˆ ˆ θ ( y ) → ∞ • The minimum risk is achieved at the point the derivative changes sign 25 / 44

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend