analysis of compressive sensing in radar
play

Analysis of Compressive Sensing in Radar Holger Rauhut Lehrstuhl C - PowerPoint PPT Presentation

Analysis of Compressive Sensing in Radar Holger Rauhut Lehrstuhl C f ur Mathematik (Analysis) RWTH Aachen Compressed Sensing and Its Applications TU Berlin December 8, 2015 1 / 44 Overview Several radar setups with compressive sensing


  1. Analysis of Compressive Sensing in Radar Holger Rauhut Lehrstuhl C f¨ ur Mathematik (Analysis) RWTH Aachen Compressed Sensing and Its Applications TU Berlin December 8, 2015 1 / 44

  2. Overview Several radar setups with compressive sensing approaches ◮ Range-Doppler resolution via compressive sensing ◮ Sparse MIMO Radar ◮ Antenna arrays with randomly positioned antennas 2 / 44

  3. Time-Frequency Structured Random Matrices Resolution of Range-Doppler in Radar 3 / 44

  4. Resolution of Range-Doppler Received signal is superposition of delayed and modulated (Doppler shifted) versions of sent signal. Task: Determine delays (corresponding to distances; range) and Doppler shifts (corresponding to radial speed) from subsampled receive signal! 4 / 44

  5. Gabor Systems in Finite Dimensions Translation and Modulation on C m ( T k g ) j = g ( j − k ) ( M ℓ g ) j = e 2 π i ℓ j / m g j . and mod m Time-frequency shifts π ( λ ) = M ℓ T k , λ = ( k , ℓ ) ∈ { 0 , . . . , m − 1 } 2 . For g ∈ C m define Gabor synthesis matrix ( ω = e 2 π i / m ) Ψ g = ( π ( λ ) g ) λ ∈{ 0 ,..., m − 1 } 2   g 0 g m − 1 g 1 g 0 g 1 g 1 · · · · · · · · · ω m − 1 g 2 g 1 g 0 g 2 ω g 1 ω g 2  · · · · · · · · ·  ω 2 g 2 ω 2 g 3 ω 2( m − 1) g 3   g 2 g 1 g 3 · · · · · · · · ·   ω 3 g 3 ω 3 g 4 ω 3( m − 1) g 4 = .   g 3 g 2 g 4 · · · · · · · · ·   . . . . . . ... ... . . . . . . . . . . . . ω ( m − 1) 2 g 0 ω m − 1 g m − 1 ω m − 1 g 0 g m − 1 g m − 2 g 0 · · · · · · · · · Use of Ψ g ∈ C m × m 2 as measurement matrix in compressive sensing

  6. Gabor Systems in Finite Dimensions Translation and Modulation on C m ( T k g ) j = g ( j − k ) ( M ℓ g ) j = e 2 π i ℓ j / m g j . and mod m Time-frequency shifts π ( λ ) = M ℓ T k , λ = ( k , ℓ ) ∈ { 0 , . . . , m − 1 } 2 . For g ∈ C m define Gabor synthesis matrix ( ω = e 2 π i / m ) Ψ g = ( π ( λ ) g ) λ ∈{ 0 ,..., m − 1 } 2   g 0 g m − 1 g 1 g 0 g 1 g 1 · · · · · · · · · ω m − 1 g 2 g 1 g 0 g 2 ω g 1 ω g 2  · · · · · · · · ·  ω 2 g 2 ω 2 g 3 ω 2( m − 1) g 3   g 2 g 1 g 3 · · · · · · · · ·   ω 3 g 3 ω 3 g 4 ω 3( m − 1) g 4 = .   g 3 g 2 g 4 · · · · · · · · ·   . . . . . . ... ... . . . . . . . . . . . . ω ( m − 1) 2 g 0 ω m − 1 g m − 1 ω m − 1 g 0 g m − 1 g m − 2 g 0 · · · · · · · · · Use of Ψ g ∈ C m × m 2 as measurement matrix in compressive sensing 5 / 44

  7. Radar model (Herman, Strohmer 2008) Emitted signal: g ∈ C m . Objects scatters g and radar device receives the contribution x λ π ( λ ) g = x k ,ℓ M ℓ T k g . T k corresponds to delay, i.e., distance of object M ℓ corresponds to Doppler shift, i.e., speed of the object x k ,ℓ reflectivity of object Received signal is superposition of contribution of all scatteres: � y = x λ π ( λ ) g = Ψ g x . λ ∈ Λ Usually few scatterers so that x ∈ C m 2 can be assumed sparse. 6 / 44

  8. Radar model (Herman, Strohmer 2008) Emitted signal: g ∈ C m . Objects scatters g and radar device receives the contribution x λ π ( λ ) g = x k ,ℓ M ℓ T k g . T k corresponds to delay, i.e., distance of object M ℓ corresponds to Doppler shift, i.e., speed of the object x k ,ℓ reflectivity of object Received signal is superposition of contribution of all scatteres: � y = x λ π ( λ ) g = Ψ g x . λ ∈ Λ Usually few scatterers so that x ∈ C m 2 can be assumed sparse. We will choose g as random vector below. 6 / 44

  9. Reconstruction via compressive sensing Reconstruction of x from y = Ax via ℓ 1 -minimization min � z � 1 subject to Az = y min � z � 1 subject to � Az − y � 2 ≤ η 7 / 44

  10. Reconstruction via compressive sensing Reconstruction of x from y = Ax via ℓ 1 -minimization min � z � 1 subject to Az = y min � z � 1 subject to � Az − y � 2 ≤ η Alternatives: Matching Pursuits Iterative hard thresholding (pursuit) Iteratively reweighted least squares ... 7 / 44

  11. Uniform vs. nonuniform recovery Often recovery results are for random matrices A ∈ R m × N ; choose generator g ∈ C m for Ψ g at random ◮ Uniform recovery With high probability on A every sparse vector is recovered; P ( ∀ s -sparse x , recovery of x is successful using A ) ≥ 1 − ε. 8 / 44

  12. Uniform vs. nonuniform recovery Often recovery results are for random matrices A ∈ R m × N ; choose generator g ∈ C m for Ψ g at random ◮ Uniform recovery With high probability on A every sparse vector is recovered; P ( ∀ s -sparse x , recovery of x is successful using A ) ≥ 1 − ε. Recovery conditions on A ◮ Null space property ◮ Restricted isometry property 8 / 44

  13. Uniform vs. nonuniform recovery Often recovery results are for random matrices A ∈ R m × N ; choose generator g ∈ C m for Ψ g at random ◮ Uniform recovery With high probability on A every sparse vector is recovered; P ( ∀ s -sparse x , recovery of x is successful using A ) ≥ 1 − ε. Recovery conditions on A ◮ Null space property ◮ Restricted isometry property ◮ Nonuniform recovery A fixed sparse vector is recovered with high probability using A ∈ R m × N ; ∀ s -sparse x : P (recovery of x is successful using A ) ≥ 1 − ε. 8 / 44

  14. Uniform vs. nonuniform recovery Often recovery results are for random matrices A ∈ R m × N ; choose generator g ∈ C m for Ψ g at random ◮ Uniform recovery With high probability on A every sparse vector is recovered; P ( ∀ s -sparse x , recovery of x is successful using A ) ≥ 1 − ε. Recovery conditions on A ◮ Null space property ◮ Restricted isometry property ◮ Nonuniform recovery A fixed sparse vector is recovered with high probability using A ∈ R m × N ; ∀ s -sparse x : P (recovery of x is successful using A ) ≥ 1 − ε. Recovery conditions on A ◮ Tangent cone (descent cone) of norm at x intersects ker A trivially. ◮ Dual certificates 8 / 44

  15. Restricted isometry property (RIP) Definition The restricted isometry constant δ s of a matrix A ∈ C m × N is defined as the smallest δ s such that (1 − δ s ) � x � 2 2 ≤ � A x � 2 2 ≤ (1 + δ s ) � x � 2 2 for all s -sparse x ∈ C N . 9 / 44

  16. Stable and robust recovery Theorem (Cand` es, Romberg, Tao ’04 – Cai, Zhang ’13) √ Let A ∈ C m × N with δ 2 s < 1 / 2 ≈ 0 . 7071 . Let x ∈ C N , and assume that noisy data are observed, y = A x + e with � e � 2 ≤ τ . Let x # by a solution of z � z � 1 � A z − y � 2 ≤ τ. min such that Then � x − x # � 2 ≤ C σ s ( x ) 1 √ s + D τ, � x − x # � 1 ≤ C σ s ( x ) 1 + D √ s τ for constants C , D > 0 , that depend only on δ 2 s . Here σ s ( x ) 1 = z : � z � 0 ≤ s � x − z � 1 . inf Implies exact recovery in the s -sparse and noiseless case. 10 / 44

  17. Dual certificate Theorem (Fuchs 2004, Tropp 2005) For A ∈ C m × N , x ∈ C N with support S is the unique solution of min � z � 1 subject to A z = A x if A S is injective and there exists a dual vector h ∈ C m such that ( A ∗ h ) j = sgn( x j ) , | ( A ∗ h ) ℓ | < 1 , j ∈ S , ℓ ∈ S . Corollary Let a 1 , . . . , a N be the columns of A ∈ C m × N . For x ∈ C N with support S, if the matrix A S is injective and if |� A † S a ℓ , sgn( x S ) �| < 1 for all ℓ ∈ S , then the vector x is the unique ℓ 1 -minimizer with y = A x . Here, A † S is Moore-Penrose pseudo inverse. 11 / 44

  18. Dual certificate Theorem (Fuchs 2004, Tropp 2005) For A ∈ C m × N , x ∈ C N with support S is the unique solution of min � z � 1 subject to A z = A x if A S is injective and there exists a dual vector h ∈ C m such that ( A ∗ h ) j = sgn( x j ) , | ( A ∗ h ) ℓ | < 1 , j ∈ S , ℓ ∈ S . Corollary Let a 1 , . . . , a N be the columns of A ∈ C m × N . For x ∈ C N with support S, if the matrix A S is injective and if |� A † S a ℓ , sgn( x S ) �| < 1 for all ℓ ∈ S , then the vector x is the unique ℓ 1 -minimizer with y = A x . Here, A † S is Moore-Penrose pseudo inverse. One ingredient: Check that � A ∗ S A S − I � 2 → 2 ≤ δ < 1. 11 / 44

  19. Stability and robustness via dual certificate Theorem Let x ∈ C N and A ∈ C m × N with ℓ 2 -normalized columns. Denote by S ⊂ [ N ] the indices of the s largest absolute entries of x . Assume that (i) there is a dual certificate u = A ∗ h ∈ C N with h ∈ C m s.t. � h � 2 ≤ 3 √ s . � u T c � ∞ ≤ 1 u T = sgn( x ) T , 2 , (ii) � A ∗ T A T − I � 2 → 2 ≤ 1 2 . Given noisy measurements y = A x + e ∈ C m with � e � 2 ≤ τ , the x ∈ C N of noise-constrained ℓ 1 -minimization satisfies solution ˆ x � 2 ≤ 52 √ s τ + 16 σ s ( x ) 1 . � x − ˆ 12 / 44

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend