adaptive signal recovery by convex optimization
play

Adaptive Signal Recovery by Convex Optimization Dmitrii Ostrovskii - PowerPoint PPT Presentation

Adaptive Signal Recovery by Convex Optimization Dmitrii Ostrovskii CWI, Amsterdam 19 April 2018 Signal denoising problem Recover complex signal x = ( x ) , = n , ..., n , from noisy observations y = x + , = n ,


  1. Adaptive Signal Recovery by Convex Optimization Dmitrii Ostrovskii CWI, Amsterdam 19 April 2018

  2. Signal denoising problem Recover complex signal x = ( x τ ) , τ = − n , ..., n , from noisy observations y τ = x τ + σξ τ , τ = − n , ..., n , where ξ τ are i.i.d. standard complex Gaussian random variables. 2 5 1.5 1 0.5 0 0 -0.5 -1 -1.5 -2 -5 0 20 40 60 80 100 0 20 40 60 80 100 Signal Observations • Assumption: signal has unknown shift-invariant structure . Dmitrii Ostrovskii Adaptive Signal Recovery by Convex Optimization 1 / 31

  3. Preliminaries • Finite-dimensional spaces and norms: C n ( Z ) = { x = ( x τ ) τ ∈ Z : x τ = 0 whenever | τ | > n } ; ℓ p -norms restricted to C n ( Z ): �� | τ |≤ n | x τ | p � 1 p ; � x � p = Scaled ℓ p -norms: 1 � x � n , p = (2 n + 1) 1 / p � x � p . • Loss: ℓ ( � x , x ) = | � x 0 − x 0 | – pointwise loss; ℓ ( � x , x ) = � � x − x � n , 2 – ℓ 2 -loss. • Risk: 1 x , x ) 2 ] 2 ; R ( � x , x ) = [ E ℓ ( � R δ ( � x , x ) = min { r ≥ 0 : ℓ ( � x , x ) ≤ r with probability ≥ 1 − δ } . Dmitrii Ostrovskii Adaptive Signal Recovery by Convex Optimization 2 / 31

  4. Adaptive estimation: disclaimer Classical approach Given a set X containing x , look for a near-minimax , over X , estimator x o is linear in y (e.g. for pointwise loss)*. x o . One can often assume that � � x o becomes an unavailable linear oracle . Mimic it! If X is unknown , � Oracle approach x o with small risk R ( � x o , x ), Knowing that there exists a linear oracle � construct an adaptive estimator � x = � x ( y ) satisfying an oracle inequality : x o , x ) + Rem , x o , x ) . R ( � x , x ) ≤ P · R ( � Rem ≪ R ( � x o can change but P and Rem must be uniformly bounded over ( � x o , x ). x , � • P = “price of adaptation”. Inequalities with P = 1 are called sharp *. *[Ibragimov and Khasminskii, 1984; Donoho et al., 1990], *[Tsybakov, 2008] Dmitrii Ostrovskii Adaptive Signal Recovery by Convex Optimization 3 / 31

  5. Classical example: unknown smoothness Let x be a regularly sampled function: x t = f ( t / N ) , t = − N , ..., N , where f : [ − 1 , 1] → R has weak derivative D s f of order s ≥ 1 on [ − 1 , 1], and belongs to a Sobolev ( q = 2) or H¨ older ( q = ∞ ) smoothness class:* F s , L = { f ( · ) : � D s f � L q ≤ L } . • Linear oracle: kernel estimator with properly chosen bandwidth h : � τ � � 1 � f ( t / N ) = K y t − τ , | t | ≤ N − hN . 2 hN + 1 hN | τ |≤ hN • Adaptive bandwidth selection*: Lepski’s method, Stein’s method, ... *[Adams and Fournier, 2003; Brown et al., 1996; Watson, 1964; Nadaraya, 1964; Tsybakov, 2008; Johnstone, 2011], *[Lepski, 1991; Lepski et al., 1997, 2015; Goldenshluger et al., 2011] Dmitrii Ostrovskii Adaptive Signal Recovery by Convex Optimization 4 / 31

  6. Recoverable signals • We consider convolution-type (or time-invariant) estimators � � x t = [ ϕ ∗ y ] t := ϕ τ y t − τ , τ ∈ Z where ∗ is discrete convolution , and ϕ ∈ C n ( Z ) is called a filter . Definition* A signal x is ( n , ρ )- recoverable if there exists φ o ∈ C n ( Z ) which satisfies � E | x t − [ φ o ∗ y ] t | 2 � 1 / 2 σρ √ 2 n + 1 , ≤ | t | ≤ 3 n . 3 n , 2 ] 1 / 2 ≤ • Consequence: small ℓ 2 -risk: [ E � x − φ o ∗ y � 2 σρ √ 2 n +1 . -4n -3n -2n -n 0 n 2n 3n 4n *[Juditsky and Nemirovski, 2009; Nemirovski, 1991; Goldenshluger and Nemirovski, 1997] Dmitrii Ostrovskii Adaptive Signal Recovery by Convex Optimization 5 / 31

  7. Adaptive signal recovery: main questions Goal Assuming that x is ( n , ρ )-recoverable, construct an adaptive filter σρ ϕ = � � ϕ ( y ) such that the pointwise or ℓ 2 -risk of � x = � ϕ ∗ y is close to √ 2 n +1 . Main questions: • Can we adapt to the oracle? Yes , but we must pay the price polynomial in ρ ; • Can � ϕ be efficiently computed? Yes , by solving a well-structured convex optimization problem. • Do recoverable signals with small ρ exist? Yes: when the signal belongs to shift-invariant subspace S ⊂ C ( Z ), dim( S ) = s , we have “nice” bounds on ρ = ρ ( s ). Dmitrii Ostrovskii Adaptive Signal Recovery by Convex Optimization 6 / 31

  8. Adaptive estimators and their analysis

  9. Main idea • “Bias-variance decomposition” x t − [ φ o ∗ y ] t = x t − [ φ o ∗ x ] t + σ [ φ o ∗ ξ ] t . � �� � � �� � � �� � total error bias stochastic error • ( n , ρ )-recoverability implies σρ ρ | x t − [ φ o ∗ x ] t | ≤ � φ o � 2 ≤ √ 2 n + 1 , | t | ≤ 3 n , and √ 2 n + 1 . • Unitary Discrete Fourier transform operator F n : C n ( Z ) → C n ( Z ). Look at the Fourier transforms Estimate x via � x = � ϕ ∗ y , where � ϕ = � ϕ ( y ) ∈ C 2 n ( Z ) minimizes the Fourier-domain residual �F 2 n [ y − ϕ ∗ y ] � p while keeping �F 2 n [ ϕ ] � 1 small. Dmitrii Ostrovskii Adaptive Signal Recovery by Convex Optimization 7 / 31

  10. Motivation: new oracle Oracle with small ℓ 1 -norm of DFT* If x is ( n , ρ )-recoverable, then there exists a ϕ o ∈ C 2 n ( Z ) s.t. for R = 2 ρ 2 , C σ R R | x t − [ ϕ o ∗ x ] t | ≤ �F 2 n [ ϕ o ] � 1 ≤ √ 4 n + 1 , | t | ≤ 2 n , √ 4 n + 1 . Proof. 1 o . Consider ϕ o = φ o ∗ φ o ∈ C 2 n ( Z ). On one hand, for | t | ≤ 2 n , | x t − [ ϕ o ∗ x ] t | = | x t − [ φ o ∗ x ] t | + | [ φ o ∗ ( x − φ o ∗ x )] t | | τ |≤ 3 n | x τ − [ φ o ∗ x ] τ | ≤ σρ (1 + ρ ) ≤ (1 + � φ o � 1 ) max √ 2 n + 1 . 2 o . On the other hand, we get √ 2 ρ 2 4 n + 1 �F 2 n [ ϕ o ] � 1 = √ 4 n + 1 �F 2 n [ φ o ] � 2 4 n + 1 �F n [ φ o ] � 2 2 = 2 ≤ √ 4 n + 1 . � *[Juditsky and Nemirovski, 2009] Dmitrii Ostrovskii Adaptive Signal Recovery by Convex Optimization 8 / 31

  11. Uniform-fit estimators • Constrained uniform-fit estimator *: � � R ϕ ∈ Argmin � �F n [ y − ϕ ∗ y ] � ∞ : �F n [ ϕ ] � 1 ≤ √ 2 n + 1 . (CUF) ϕ ∈ C n ( Z ) • Penalized estimator : for some λ ≥ 0, � � √ ϕ ∈ Argmin �F n [ y − ϕ ∗ y ] � ∞ + σλ 2 n + 1 �F n [ ϕ ] � 1 . (PUF) � ϕ ∈ C n ( Z ) Pointwise upper bound for uniform-fit estimators 2 ⌉ , ρ )-recoverable. Let R = 2 ρ 2 for the constrained estimator, Let x be ( ⌈ n � and λ = 2 log[(2 n + 1) /δ ] for the penalized one, then w.p. ≥ 1 − δ , ϕ ∗ y ] 0 | ≤ C σρ 4 � log[(2 n + 1) /δ ] √ 2 n + 1 | x 0 − [ � . High price of adaptation: O ( ρ 3 √ log n ) . *[Juditsky and Nemirovski, 2009] Dmitrii Ostrovskii Adaptive Signal Recovery by Convex Optimization 9 / 31

  12. Analysis of uniform-fit estimators Let � ϕ be an optimal solution to (CUF) with R = R , and let � Θ n ( ζ ) = �F n [ ζ ] � ∞ = O ( log n ) w.h.p. 1 o . Already in the first step, we see why the new oracle is useful: | [ x − � ϕ ∗ y ] 0 | ≤ σ | [ � ϕ ∗ ζ ] 0 | + | [ x − � ϕ ∗ x ] 0 | ≤ σ �F n [ � ϕ ] � 1 �F n [ ζ ] � ∞ + | [ x − � ϕ ∗ x ] 0 | [Young’s ineq.] ≤ σ Θ n ( ζ ) R √ 2 n + 1 + | [ x − � ϕ ∗ x ] 0 | . [Feasibility of � ϕ ] 2 o . To control | [ x − � ϕ ∗ x ] 0 | , we can add & subtract convolution with ϕ o : ϕ ∗ x ] 0 | ≤ | [ ϕ o ∗ ( x − � ϕ ) ∗ ( x − ϕ o ∗ x )] 0 | | x 0 − [ � ϕ ∗ x )] 0 | + | [(1 − � ϕ � 1 ) � [ x − ϕ o ∗ x ] � ∞ ≤ �F n [ ϕ o ] � 1 �F n [ x − � ϕ ∗ x ] � ∞ + (1 + � � R ϕ ∗ x ] � ∞ + CR (1 + R ) √ 2 n + 1 �F n [ x − � √ 2 n + 1 . ≤ Dmitrii Ostrovskii Adaptive Signal Recovery by Convex Optimization 10 / 31

  13. Analysis of uniform-fit estimators, cont. 3 o . It remains to control �F n [ x − � ϕ ∗ x ] � ∞ which can be done as follows: �F n [ x − � ϕ ∗ x ] � ∞ ≤ �F n [ y − � ϕ ∗ y ] � ∞ + σ �F n [ ζ − � ϕ ∗ ζ ] � ∞ ≤ �F n [ y − � ϕ ∗ y ] � ∞ + σ (1 + � � ϕ � 1 )Θ n ( ζ ) ≤ �F n [ y − ϕ o ∗ y ] � ∞ + σ (1 + � � ϕ � 1 )Θ n ( ζ ) [Feas. of ϕ o ] ≤ �F n [ x − ϕ o ∗ x ] � ∞ + 2 σ (1 + R )Θ n ( ζ ) . 4 o . Finally, note that �F n [ x − ϕ o ∗ x ] � ∞ ≤ �F n [ x − ϕ o ∗ x ] � 2 = � [ x − ϕ o ∗ x ] � 2 [Parseval’s identity] √ 2 n + 1 � x − ϕ o ∗ x � ∞ ≤ σ CR . ≤ Collecting the above, we obtain a bound dominated by σ CR (1+ R )Θ n ( ζ ) . � √ 2 n +1 Dmitrii Ostrovskii Adaptive Signal Recovery by Convex Optimization 11 / 31

  14. Limit of performance Proposition: pointwise lower bound For any integer n ≥ 2, α < 1 / 4, and ρ satisfying 1 ≤ ρ ≤ n α , one can point out a family of signals X n ,ρ ∈ C 2 n ( Z ) such that • any signal in X n ,ρ is ( n , ρ )-recoverable; • for any estimate � x 0 of x 0 from observations y ∈ C 2 n ( Z ), one can find x ∈ X n ,ρ satisfying � � x 0 | ≥ c σρ 2 � (1 − 4 α ) log n P | x 0 − � √ 2 n + 1 ≥ 1 / 8 . Conclusion: there is a gap ρ 2 between upper and lower bounds. • To bridge it (and encompass ℓ 2 -loss), we introduce new estimators. Dmitrii Ostrovskii Adaptive Signal Recovery by Convex Optimization 12 / 31

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend