M-estimation Our Paper Isometry Between (M)-estimation & Lasso
M-Estimation under High-Dimensional Asymptotics
DLD, Andrea Montanari 2014-05-01
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-Estimation under High-Dimensional Asymptotics DLD, Andrea - - PowerPoint PPT Presentation
M-estimation Our Paper Isometry Between (M)-estimation & Lasso M-Estimation under High-Dimensional Asymptotics DLD, Andrea Montanari 2014-05-01 DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics M-estimation
M-estimation Our Paper Isometry Between (M)-estimation & Lasso
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation
Location model Yi = θ + Zi , i = 1, . . . , n Errors: Zi ∼ F, not necessarily Gaussian. “Loss” Function ρ(t) eg t2, |t|, − log(f (t)),. . . (M) min
θ n
ρ(Yi − θ) Asymptotic Distribution √n(ˆ θn − θ) ⇒D N(0, V ), n → ∞. Asymptotic Variance: ψ = ρ′: V (ψ, F) =
(
Information Bound V (ψ, F) ≥ 1 I(F) DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation
P. J. BICKEL*
Simple "one-step" versions
the linear model are introduced. Some relevant Monte Carlo results
in the Princeton project [1] are singled out and discussed. The large sample behavior
these procedures is examined under very mild regularity conditions.
In 1964 Huber [7] introduced a class
estimates (referred to as (M)) in the location problem, studied their asymptotic behavior and identified robust members
the group. These procedures are the solutions 8 of equa- tions
the form,
n
E +(X}-')
(1.1) where Xi = 0 + El, * ?Xn
= + E. and
El, ** En are unknown independent, identically distributed errors which have a distribution F which is symmetric about 0. If F has a density f which is smooth and if f is known, then maximum likelihood estimates if they exist satisfy (1.1) with F6
= -f'/f.
Under successively milder regularity conditions
and F, Huber showed in [7] and [8] that such ' were consistent and asymptotically normal with mean and variance K(#k, F)/n where K(VI, F) =1 2 /(t)f(t)
dt f(t)do(t)I
. (1.2)
If F is unknown but close to a normal distribution with mean 0 and known variance in a suitable sense, Huber in [7] further showed that (M) estimates based
#K(t) = t
if Itl < K = Ksgnt if Itl >K (1.3) have a desirable minimax robustness property. If K is finite these estimates can
be calculated iteratively. It has, however, been
by Fisher, Neyman and
that if F is known and ' = ((- f'/f), the estimate
by starting with a Vn consistent estimate 6 and performing
Gauss-Newton iteration
(1.1) is asymptotically efficient even when the MLE is not and is equivalent to it when it is (cf. [13]). One purpose
this note is to show that under mild conditions this
* P.J. Bickel is professor, Department
University
Berkeley,
was performed with partial support
O.N.R. under Contract N00014-67-A-D151-0017 with Princeton University, and N00014-67-A0114-0004 with the University
California at Berkeley, as well as that
the John Simon Guggenheim Foundation. The author would like to thank P.J. Huber,
and C. Van Eeden and D. Relles for providing him with reprints
their work
subject;
III for programming the Monte Carlo computations
Section 3, which appeared in the Princeton project; and a referee who made Tables 1 and 2 reflect numerical realities.
equivalence holds in the more general context
the linear model for general 46. Typically the estimates
from (1.1) are not scale equivariant.1 To obtain acceptable procedures a scale equivariant and location invariant estimate
scale 6 must be calculated from the data and 6 be
as the solution
n E, Off (Xi - 0)
= O(1.4)
j-1
where (x) = (x/a) . (1.5) The resulting 6 is then both location and scale equi- variant. The estimate 6 can be
simultaneously with ' by solving a system
equations such as those
Huber's Proposal 2 [8, p. 96] or the "likelihood equations"
n Xj - ;\
E 4 ) 0X a-1 O' (1.6)
n Xi
where x (t) = t4* (t)
Or, we may choose 6 indepen- dently. For instance, in this article, the normalized inter- quartile range, l = (X(n-[n/4]+1)
and the symmetrized interquartile range,
62 = median
{ i - m I/(D-1(3),
(1.8) are used where X(l) < ... < X(n) are the
statistics, 4' is the standard normal cdf and m is the sample median. If 6
as hy- pothesized, then the asymptotic theory for the location model continues to be valid with K (t, F) replaced by K(#6( (F)), F). (E.g., cf. [7].) We shall show (in the con- text
the linear model) under mild conditions that the
"Gauss-Newton" approximation to (1.4)-O being the
unknown-behaves asymptotically like the root. The estimates corresponding to
OK have
a rather ap- pealing form and,
course, all of these Gauss-Newton
I In this article location (scale) invariance refers to procedures which remain unchanged when the data are shifted (rescaled). The term "equivariant" is in ac- cord with its usage in [2]. Thus, ; location and scale equivariant means that ;(aX1 + b, * ,
*, aX,) + b and a scale equivariant means that 3(aXi, * *, aX.) = Ia I(Xi, * *, Xn). a Journal
the American Statistical Association June 1975, Volume 70, Number 350 Theory and Methods Section
428
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation
Regression model Yi = X ′
i θ + Zi ,
Zi ∼iid F, i = 1, . . . , n Objective function of (M): R(ϑ) =
n
ρ(Yi − X ′
i ϑ)
(M) min
ϑ R(ϑ)
One-step estimate: ˜ θn any √n-consistent estimate of θ: ˆ θ1 = ˜ θn − [Hess R| ˜
θn ]−1∇R| ˜ θn .
Effectiveness: ˆ θ true solution of M-equation: E(ˆ θ1 − ˆ θ)(ˆ θ1 − ˆ θ)′ = o(n−1) DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation
PJ Huber, Annals of Statistics 1973 DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation
Noureddine El Karouia,1, Derek Beana, Peter J. Bickela,1, Chinghway Limb, and Bin Yua
aDepartment of Statistics, University of California, Berkeley, CA 94720; and bDepartment of Statistics and Applied Probability, Faculty of Science,
National University of Singapore, 119077 Contributed by Peter J. Bickel, April 25, 2013 (sent for review March 1, 2012)
We study regression M-estimates in the setting where p, the num- ber of covariates, and n, the number of observations, are both large, but p ≤ n. We find an exact stochastic representation for the distribution of ^ β = argminβ∈Rp∑n
i=1 ρðYi − Xi′βÞ at fixed p and n
under various assumptions on the objective function ρ and our statistical model. A scalar random variable whose deterministic limit rρðκÞ can be studied when p=n → κ > 0 plays a central role in this representation. We discover a nonlinear system of two deter- ministic equations that characterizes rρðκÞ. Interestingly, the sys- tem shows that rρðκÞ depends on ρ through proximal mappings of ρ as well as various aspects of the statistical model underlying our
that, when p=n is large enough, least squares becomes preferable to least absolute deviations for double-exponential errors.
prox function | high-dimensional statistics | concentration of measure
aged to obtain rigorous proofs for many of our assertions. They will be presented elsewhere because they are very long and tech- nical.) We give several results for covariates that are Gaussian or derived from Gaussian but present grounds that the behavior holds much more generally—the key being concentration of certain quadratic forms involving the vectors of covariates. We also investigate the sensitivity of our results to the geometry of the design matrix. [Further results with different designs can be found in our work (5).] We find that (i) estimates of coordinates and contrasts that have coefficients independent of the observed covariates con- tinue to be unbiased and asymptotically normal; and (ii) as in the fixed p case, this happens at scale n−1=2, at least when the min- imal and maximal eigenvalues of the covariance of the predictors stay bounded away from 0 and ∞, respectively.* These findings are obtained by (i) using leave-one-out per- turbation arguments both for the data units and predictors; (ii)
STATISTICS
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation
Classical setting - random design, p fixed, n → ∞. var(ˆ θi ) → V (ψ, F), n → ∞.
HDA setting - for n/pn → δ ∈ (1, ∞) ◮ Effective Score ˜ Ψ = ˜ Ψδ,ψ,F (to be described... ) ◮ Effective Error Distribution ˜ F = F ⋆ N(0, τ2
∞)
Extra Gaussian noise: τ∞ = τ∞(δ, ψ, F). ◮ Asymptotic Variance under HDA var(ˆ θi ) → V ( ˜ ψ, ˜ F), n, pn → ∞. ◮ Classical Correspondence ˜ Ψδ,ψ,F → ψ, ˜ Fδ,ψ,F → F, δ → ∞. DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso Classical M-estimation Big Data M-estimation
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
1 2 3 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Determination of b3 b Average Slope 5 10 15 20 0.27 0.28 0.29 0.3 0.31 0.32 0.33 History of bt AMP iteration t bt
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
5 10 15 20 1.5 2 2.5 3 3.5 4 4.5 5 5.5 6 AMP iteration t R M SE( ˆ θ t, θ 0) C onve rge nc e of A M P ˆ θ t to θ 0 AMP M 5 10 15 20 1 2 3 4 5 6 C onve rge nc e of A M P ˆ θ t to ˆ θ AMP iteration t R M SE( ˆ θ t, ˆ θ)
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
0 =
2/n = MSE(
i ) = Var(Z) + Var(X(θ0 −
0 = MSE(
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
t }t≥0, starting at τ 2 0 ∈ R≥0 by
t+1 = V(τ 2 t , b(τt)) = V(τt, b(τt; δ, F); δ, F).
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
0.5 1 1.5 2 2.5 3 3.5 4 0.5 1 1.5 2 2.5 3 3.5 4 Input Variance 0
2
Output Variance 1
2
Variance map V(2), Contaminated Normal (0.472,0.472) V(tau2) y=x
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
0.5 1 1.5 2 2.5 3 3.5 4 0.5 1 1.5 2 2.5 3 3.5 4 Input Variance 0
2
Output 1
2
Dynamics of t
2
2.056 0.819 0.545 0.486 V(tau2) y=x
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
20 40 60 10 20 30 40 MSE State Evolution Iteration t 20 40 60 1 2 3 4 5 MAE State Evolution Iteration t 20 40 60 0.5 1 1.5 2 2.5 3 t State Evolution Iteration t 10 20 30 40 50 0.26 0.28 0.3 0.32 0.34 0.36 bt
CN(0.05, 10), with Huber ψ, λ = 3. Upper Left: ˆ τt = θt − θ02/√n. Upper Right: ˆ
Mean Squared Error. Lower Right: MAE, Mean Absolute Error. Blue ‘+’ symbols: Empirical means of AMP
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
τ2
t ≥
1 δI(F) .
I(F ⋆ N(0, τ2)) ≤ I(F) 1 + τ2I(F) .
τ2
t ≥
1 + 1
δ + 1 δ2 + · · · + 1 δk
δI(F) .
τ2
∗ ≥
1 δ − 1 · 1 I(F) .
lim
n→∞ Var(ˆ
θi ) ≥ 1 1 − 1/δ · 1 I(F) DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
Basic Assumptions: A1 Discrepancy function ρ is convex and smooth; A2 Matrices {X(n)}n are ∼iid N(0, 1
n )
A3 θ0, θ0 = 0 are deterministic sequences such that AMSE(θ0, θ0) = δτ2
0 .
A4 F has finite second moment. Terminology:
Let {τ2
t }t≥0 denote the state evolution sequence with initial condition τ2 0 .
Let { θt, Rt}t≥0 be the AMP trajectory with parameters bt . Definition: A function ξ : Rk → R is pseudo-Lipschitz if there exists L < ∞ such that, for all x, y ∈ Rk , |ξ(x) − ξ(y)| ≤ L(1 + x2 + y2) x − y2. DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
Simply apply existing paper: Bayati, Mohsen, and Andrea Montanari. ”The dynamics of message passing on dense graphs, with applications to compressed sensing.” IEEE IT 57.2 (2011): 764-785.
Bayati-Montanari 2011 designed for a seemingly different problem: Asymptotics of Lasso in p > n case. min ˜ Y − ˜ Xβ2
2/2 + λβ1
Setting was compressed sensing, where ˜ Xi,j ∼iid N(0, 1
n ), and
pn/n → δ ∈ (0, 1).
Formalism of AMP and State Evolution was introduced, and developed in DLD Arian Maleki, and Andrea Montanari. ”Message-passing algorithms for compressed sensing.” PNAS 106.45 (2009): 18914-18919. DLD, Arian Maleki, and Andrea Montanari. ”The noise-sensitivity phase transition in compressed sensing.” IEEE IT 57.10 (2011): 6920-6941.
These papers systematically understood and used the ‘Extra Gaussian Noise’ property of High Dimensional Asymptotics.
Generality of Bayati-Montanari treatment, easily accommodated M-estimation. DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ
Central Recursion in Bayati-Montanari 2011
ht, qt ∈ RN
zt, mt ∈ Rn
Initial Condition q0; m−1 = 0. ht+1 = A∗mt − ξtqt, mt = gt(bt, w) bt = Aqt − λtmt−1, qt = ft(ht, x0)
Reaction Coefficients: ξt = g′
t (bt, w); λt = 1 δ f ′ t (ht, x0)
State Evolution τ2
t = E{gt(σtZ, W )2};
σ2
t =
1 δ E{ft(τt−1Z, X0)} where W ∼ FW , X0 ∼ FX0 DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso AMP Algorithm State Evolution Correctness of State Evolution Convergence of AMP to ˆ θ ϑt+1 = δXTΨ(W + St; bt) + qtϑt ϑt+1 XT δΨ(W + St; bt) qt ϑt ht+1 = A∗mt − ξtqt ht+1 A∗ mt ξt −qt St = −Xϑt + Ψ(W + St−1; bt−1) St X −ϑt 1 Ψ(W + St−1; bt−1) bt = Aqt − λtmt−1 bt A qt −λt mt−1
We get exact correspondence between the two systems, provided we identify δΨ(W + St; bt) with mt = gt(bt; w) and −δht with ft(ht). One has, in particular, that λt = 1
δ f ′ t (ht) = −1, and that
ξt = g′
t (bt, w) = δΨ′(W + St; bt) = qt.
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso (Lassoλ) min
β∈R˜ p
1 2
Xβ2
2 + λ ˜ p
|βi | , (14) (Huberλ) min
ϑ∈Rp n
ρH(Yi − Xi , ϑ; λ) (15) Let X be a matrix with orthonormal rows such that XX = 0, i.e. null( X) = image(X) , (16) finally, set Y = XY . Proposition. With problem instances (Y , X) and ( Y , X) related as above, the optimal values of the Lasso problem (Lassoλ) and the Huber problem (Huberλ) are identical. The solutions of the two problems are in one-one-relation. In particular, we have
β) . (17) (numerous references: e.g. Art Owen/IPOD & earlier) DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso
◮ Lasso in ε-sparse regression problem, p > n ◮ Huber (M)-estimation in ε-contaminated data, p < n DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics
M-estimation Our Paper Isometry Between (M)-estimation & Lasso
DLD, Andrea Montanari M-Estimation under High-Dimensional Asymptotics