High-dimensional classification by sparse logistic regression Felix - - PowerPoint PPT Presentation

high dimensional classification by sparse logistic
SMART_READER_LITE
LIVE PREVIEW

High-dimensional classification by sparse logistic regression Felix - - PowerPoint PPT Presentation

High-dimensional classification by sparse logistic regression Felix Abramovich Tel Aviv University (based on joint work with Vadim Grinshtein, The Open University of Israel and Tomer Levy, Tel Aviv University) Felix Abramovich (Tel Aviv


slide-1
SLIDE 1

High-dimensional classification by sparse logistic regression

Felix Abramovich

Tel Aviv University

(based on joint work with Vadim Grinshtein, The Open University of Israel and Tomer Levy, Tel Aviv University)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 1 / 29

slide-2
SLIDE 2

Outline

1 Review on (binary) classification 2 High-dimensional (binary) classification by sparse logistic regression ◮ model, feature selection by penalized maximum likelihood ◮ theory: misclassification excess bounds, adaptive minimax classifiers ◮ computational issues: logistic Lasso and Slope 3 Multiclass extensions ◮ sparse multinomial logistic regression ◮ theory ◮ multinomial logistic group Lasso and Slope Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 2 / 29

slide-3
SLIDE 3

Binary Classification

(X, Y ) ∼ F : Y |X = x ∼ B(1, p(x)), X ∈ Rd ∼ f (x) Classifier η : Rd → {0, 1} Missclassification error R(η) = P(Y = η(x)) Bayes classifier η∗(x) = arg minη R(η) η∗(x) = I{p(x) ≥ 1/2}, R(η∗) = EX (min(p(X), 1 − p(X)))

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 3 / 29

slide-4
SLIDE 4

Binary Classification

(X, Y ) ∼ F : Y |X = x ∼ B(1, p(x)), X ∈ Rd ∼ f (x) Classifier η : Rd → {0, 1} Missclassification error R(η) = P(Y = η(x)) Bayes classifier η∗(x) = arg minη R(η) η∗(x) = I{p(x) ≥ 1/2}, R(η∗) = EX (min(p(X), 1 − p(X))) Data D = (X1, Y1), . . . , (Xn, Yn)) ∼ F (conditional) Missclassification error R(ˆ η) = P(Y = ˆ η(x)|D) Misclassification excess risk E(ˆ η, η∗) = ER(ˆ η) − R(η∗)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 3 / 29

slide-5
SLIDE 5

Vapnik-Chervonenkis (VC) dimension

Definition

Let C be a set of classifiers. VC(C) is the maximal number of points in X that can be arbitrarily classified by classifiers in C.

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 4 / 29

slide-6
SLIDE 6

Vapnik-Chervonenkis (VC) dimension

Definition

Let C be a set of classifiers. VC(C) is the maximal number of points in X that can be arbitrarily classified by classifiers in C. Example: VC of linear classifiers C = {η(x) = I{βtx ≥ 0}, β ∈ Rd} X = R2, C = {η(x) = I{β0 + β1x1 + β2x2 ≥ 0} VC(C) = 3 (= d) X = Rd−1, β ∈ Rd (x0 = 1) VC(C) = d

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 4 / 29

slide-7
SLIDE 7

Example: VC of sine classifiers: X = R, C = {η(x) = I{x ≥ sin(θx), θ > 0} Can classify any finite subset of points, VC(C) = ∞

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 5 / 29

slide-8
SLIDE 8

Minimax lower bound

Minimax lower bound. Let 2 ≤ VC(C) < ∞, n ≥ VC(C) and R(η∗) > 0. Then, inf

˜ η

sup

η∗∈C,f (x)

E(˜ η, η∗) ≥ C

  • VC(C)

n (e.g., Devroye, Gy¨

  • rfi and Lugosi, ’96).

In particular, for linear classifiers inf

˜ η

sup

η∗∈C,f (x)

E(˜ η, η∗) ≥ C

  • d

n

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 6 / 29

slide-9
SLIDE 9

Two main approaches

  • 1. Empirical Risk Minimization (ERM)

ˆ η = arg min

η∈C

ˆ R(η) = arg min

η∈C

1 n

n

  • i=1

I(Yi = η(xi)) well-developed theory (Devroye, Gy¨

  • rfi and Lugosi ’96; Vapnik ’00; see also Boucheron,

Bousquet and Lugosi ’05 for review) sup

η∗∈C

E(ˆ η, η∗) ≤ C

  • VC(C)

n (optimal order) computationally infeasible, various convex surrogates (e.g., SVM)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 7 / 29

slide-10
SLIDE 10
  • 2. Plug-in Classifiers

estimate p(x) from the data (e.g, (parametric) logistic regression: ln

p(x) 1−p(x) = βtx or

nonparametic: Yang ’99, Koltchinskii and Beznosova ’05, Audibert and Tsybakov ’07) plug-in ˆ η(x) = I(ˆ p(x) ≥ 1/2)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 8 / 29

slide-11
SLIDE 11
  • 2. Plug-in Classifiers

estimate p(x) from the data (e.g, (parametric) logistic regression: ln

p(x) 1−p(x) = βtx or

nonparametic: Yang ’99, Koltchinskii and Beznosova ’05, Audibert and Tsybakov ’07) plug-in ˆ η(x) = I(ˆ p(x) ≥ 1/2) Logistic regression classifier

1 ln

p(x) 1−p(x) = βtx

2 estimate β by MLE 3 plug-in ˆ

η = I(ˆ p(x) ≥ 1/2) = I(ˆ β

tx ≥ 0) – linear classifier

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 8 / 29

slide-12
SLIDE 12

Big Data era – curse of dimensionality

For large d classification without feature (model) selection is as bad as just pure random guessing (e.g., Bickel and Levina ’04; Fan and Fan ’08)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 9 / 29

slide-13
SLIDE 13

Big Data era – curse of dimensionality

For large d classification without feature (model) selection is as bad as just pure random guessing (e.g., Bickel and Levina ’04; Fan and Fan ’08) Sparse logistic regression classifier

1 model/feature selection –

M

2 plug-in ˆ

η

M = I(ˆ

β

t

  • Mx ≥ 0)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 9 / 29

slide-14
SLIDE 14

Sparse logistic regression

(X, Y ) ∼ F : Y |X = x ∼ B(1, p(x)), X ∈ Rd ∼ f (x) logit(p(x)) = ln

p(x) 1−p(x) = βtx

sparsity assumption: ||β||0 ≤ d0

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 10 / 29

slide-15
SLIDE 15

Sparse logistic regression

(X, Y ) ∼ F : Y |X = x ∼ B(1, p(x)), X ∈ Rd ∼ f (x) logit(p(x)) = ln

p(x) 1−p(x) = βtx

sparsity assumption: ||β||0 ≤ d0

Lemma (thanks to Noga Alon)

Let C(d0) = {η(x) = I{βtx ≥ 0} : β ∈ Rd, ||β||0 ≤ d0}. d0 log2 2d d0

  • ≤ VC(C(d0)) ≤ 2d0 log2

de d0

  • , i.e.

VC(C(d0)) ∼ d0 ln de d0

  • Felix Abramovich (Tel Aviv University)

High-dimensional classification by sparse logistic regression 10 / 29

slide-16
SLIDE 16

Model/feature selection by penalized MLE

For a given model M ⊆ {1, . . . , d}, MLE:

  • βM = arg max
  • β∈BM

n

  • i=1
  • β

t MxiYi − ln

  • 1 + exp(

βM)txi

  • ,

where BM = {β ∈ Rd : βj = 0 iff j / ∈ M}

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 11 / 29

slide-17
SLIDE 17

Model/feature selection by penalized MLE

For a given model M ⊆ {1, . . . , d}, MLE:

  • βM = arg max
  • β∈BM

n

  • i=1
  • β

t MxiYi − ln

  • 1 + exp(

βM)txi

  • ,

where BM = {β ∈ Rd : βj = 0 iff j / ∈ M}

  • M = arg minM

n

i=1

  • ln
  • 1 + exp(

β

t Mxi)

β

t MxiYi

  • + Pen(|M|)
  • p

M(x) =

exp( β

t

  • Mx)

1 + exp( β

t

  • Mx)
  • η

M(x) = I(

p

M(x) ≥ 1/2) = I(

β

t

  • Mx ≥ 0)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 11 / 29

slide-18
SLIDE 18

Complexity Penalties

linear-type penalties Pen(|M|) = λ|M| λ = 1 AIC (Akaike, ’73) λ = ln(n)/2 BIC (Schwarz, ’78) λ = ln d RIC (Foster and George, ’94)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 12 / 29

slide-19
SLIDE 19

Complexity Penalties

linear-type penalties Pen(|M|) = λ|M| λ = 1 AIC (Akaike, ’73) λ = ln(n)/2 BIC (Schwarz, ’78) λ = ln d RIC (Foster and George, ’94) k ln(d/k)-type nonlinear penalties Pen(|M|) ∼ C|M| ln(de/|M|)

(Birg´ e and Massart, ’01, ’07; Bunea et al. ’07; AG ’10 for Gaussian regression; AG ’16 for GLM)

k ln(d/k) ∼ ln d k

  • − log(number of models of size k)

In addition, for classification, k ln(d/k) ∼ VC(C(k)) (recall Lemma)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 12 / 29

slide-20
SLIDE 20

Various complexity penalties

k Pen(k)

AIC RIC 2kln(de/k)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 13 / 29

slide-21
SLIDE 21

Let supp(f (x)) be bounded, w.l.o.g. ||x||2 ≤ 1 for all x ∈ X

Assumption (boundedness)

There exists 0 < δ < 1/2 such that δ < p(x) < 1 − δ or, equivalently, there exists C0 > 0 such that |βtx| < C0 for all x ∈ X. The assumption prevents the variance Var(Y ) = p(x)(1 − p(x)) to be infinitely close to zero.

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 14 / 29

slide-22
SLIDE 22

Excess risk bounds

Theorem (upper bound)

Under the boundedness assumption, for Pen(|M|) = C|M| ln

  • de

|M|

  • ,

sup

η∈C(d0)

E(ˆ η

M, η∗) ≤ C(δ)

  • d0 ln de

d0

n The idea of the proof:

1 E(ˆ

η

M, η∗) ≤

  • 2 EKL(p∗, ˆ

p

M) (Zhang ’04; Bartlett et al. ’06)

2 supβ∈B(d0) EKL(p∗,

p

M) = O

  • d0 ln de

d0

n

  • (AG ’16)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 15 / 29

slide-23
SLIDE 23

Excess risk bounds

Theorem (upper bound)

Under the boundedness assumption, for Pen(|M|) = C|M| ln

  • de

|M|

  • ,

sup

η∈C(d0)

E(ˆ η

M, η∗) ≤ C(δ)

  • d0 ln de

d0

n The idea of the proof:

1 E(ˆ

η

M, η∗) ≤

  • 2 EKL(p∗, ˆ

p

M) (Zhang ’04; Bartlett et al. ’06)

2 supβ∈B(d0) EKL(p∗,

p

M) = O

  • d0 ln de

d0

n

  • (AG ’16)

Recall the lower bound for 2 ≤ d0 ln

  • de

d0

  • ≤ n :

inf

˜ η

sup

η∗∈C(d0),f (x)

E(˜ η, η∗) ≥ C

  • VC(C(d0))

n ≥ C

  • d0 ln de

d0

n

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 15 / 29

slide-24
SLIDE 24

Tighter bounds under the additional low-noise condition

The main challenges are near the hyperplane βtx = 0, where p(x) = 1/2.

Assumption (low-noise condition)

P(|p(X) − 1/2| ≤ h) ≤ Chα, α ≥ 0 (Tsybakov ′04) α = 0 – no assumptions on the noise (as previously) α = ∞ – there exists a “corridor” of width 2 ln 1+2h

1−2h that separates

the sets {x : βtx > 0} and {x : βtx < 0}

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 16 / 29

slide-25
SLIDE 25

Tighter bounds under the additional low-noise condition

The main challenges are near the hyperplane βtx = 0, where p(x) = 1/2.

Assumption (low-noise condition)

P(|p(X) − 1/2| ≤ h) ≤ Chα, α ≥ 0 (Tsybakov ′04) α = 0 – no assumptions on the noise (as previously) α = ∞ – there exists a “corridor” of width 2 ln 1+2h

1−2h that separates

the sets {x : βtx > 0} and {x : βtx < 0} Under the low-noise assumption, for all 1 ≤ d0 ≤ min(d, n) and all α ≥ 0, sup

η∈C(d0)

E(ˆ η

M, η∗) ≤

  • C(δ)

d0 ln de

d0

n α+1

α+2

  • η

M is rate-optimal and adaptive to both d0 and α.

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 16 / 29

slide-26
SLIDE 26

Computational aspects

  • M = arg min

M {−ℓ(M) + Pen(|M|)}

combinatorial search over 2d models (NP problem)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 17 / 29

slide-27
SLIDE 27

Computational aspects

  • M = arg min

M {−ℓ(M) + Pen(|M|)}

combinatorial search over 2d models (NP problem) Greedy algorithms (e.g., forward selection) – approximate the global solution by a stepwise sequence of local ones (require strong constraints on design)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 17 / 29

slide-28
SLIDE 28

Computational aspects

  • M = arg min

M {−ℓ(M) + Pen(|M|)}

combinatorial search over 2d models (NP problem) Greedy algorithms (e.g., forward selection) – approximate the global solution by a stepwise sequence of local ones (require strong constraints on design) Convex relaxation methods – replace the original combinatorial problem by some convex surrogate

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 17 / 29

slide-29
SLIDE 29

Convex relaxation methods

Recall that ||x||2 ≤ 1. logistic Lasso (for linear penalties): ||ˆ β||0 → ||ˆ β||1 ˆ βLasso = arg minβ

  • − 1

nℓ(β) + λ||β||1

  • ◮ fixed λ ∝
  • ln d

n

: rate-suboptimal (up to an extra log-factor: O(

  • d0 ln d

n

)) (van de Geer ’08, Bellec et al. ‘16)

◮ adaptively chosen λ : rate-optimal (O(

  • d0 ln(de/d0)

n

) (Bellec et al. ’16 for Gaussian regression; conjecture for classification)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 18 / 29

slide-30
SLIDE 30

Convex relaxation methods

Recall that ||x||2 ≤ 1. logistic Lasso (for linear penalties): ||ˆ β||0 → ||ˆ β||1 ˆ βLasso = arg minβ

  • − 1

nℓ(β) + λ||β||1

  • ◮ fixed λ ∝
  • ln d

n

: rate-suboptimal (up to an extra log-factor: O(

  • d0 ln d

n

)) (van de Geer ’08, Bellec et al. ‘16)

◮ adaptively chosen λ : rate-optimal (O(

  • d0 ln(de/d0)

n

) (Bellec et al. ’16 for Gaussian regression; conjecture for classification)

logistic Slope: k ln(2d/k) ∼ k

j=1 ln(2d/j)

ˆ βSlope = arg minβ

  • − 1

nℓ(β) + d j=1 λj|β|(j)

  • ,

λ1 ≥ . . . ≥ λd > 0 λj ∝

  • ln(2d/j)

n

: rate-optimal (O(

  • d0 ln(de/d0)

n

) (AG ’19)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 18 / 29

slide-31
SLIDE 31

Multiclass classification

appears in a variety applications, a lot of methods much less theory behind

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 19 / 29

slide-32
SLIDE 32

Multiclass classification

appears in a variety applications, a lot of methods much less theory behind Main approaches :

1 reduction to a series of binary classifications ◮ One-vs-All – each class is compared against all others ◮ One-vs-One – all pairs of classes are compared to each other 2 extensions of binary classification approaches Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 19 / 29

slide-33
SLIDE 33

Multiclass classification

(X, Y ) ∼ F : Y |X = x ∼ Mult(p1(x), . . . , pL(x)), X ∈ Rd ∼ f (x) Classifier η : Rd → {1, . . . , L} Missclassification error R(η) = P(Y = η(x)) Bayes classifier η∗(x) = arg max1≤j≤L pj(x), R(η∗) = 1 − EX (max1≤j≤L pj(X)) Misclassification excess risk E(ˆ η, η∗) = ER(ˆ η) − R(η∗)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 20 / 29

slide-34
SLIDE 34

Multinomial logistic regression

Y ∼ Mult(p1(x), . . . , pL(x)), x ∈ Rd,

L

  • j=1

pj(x) = 1 θj = ln pj(x) pL(x) = βt

j x,

pj(x) = exp

  • βt

j x

  • L

k=1 exp

  • βt

kx

, j = 1, . . . , L; βL = 0 (the choice of the reference class is arbitrary) To each Y assign the corresponding indicator vector ξ ∈ {0, 1}L MLE: B ∈ Rd×L – matrix of regression coefficients (B·L = 0) ℓ(B) =

n

  • i=1
  • xt

i Bξi − ln L

  • l=1

exp(βt

l xi)

  • → max

B

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 21 / 29

slide-35
SLIDE 35

Sparse multinomial logistic regression

for multiclass setup there are various ways to define sparsity global sparsity: part of features do not have any impact on classification at all, i.e. Bj· = 0 for a given model M ⊆ {1, . . . , d}

◮ |M| = #{non − zero rows of B} = rB ◮

  • BM = arg max
  • B∈BM

n

  • i=1
  • Xt

i

Bξi − ln

L

  • l=1

exp( β

t l Xi)

  • ,

where BM = {B ∈ Rd×L : B·L = 0, and Bj· = 0 iff j ∈ M}

  • M = arg minM{−ℓ(

BM) + Pen(|M|)}

  • η

M = arg max1≤l≤L

β

t

  • Mlx

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 22 / 29

slide-36
SLIDE 36

CL(d0) = {η(x) = arg max1≤l≤L βt

l x : B ∈ Rd×L, B·L = 0 and rB ≤ d0}

Assumption (boundedness)

There exists 0 < δ < 1/2 such that δ ≤ pl(x) ≤ 1 − δ or, equivalently, |βt

l x| < C0 with C0 = ln 1−δ δ

for all l = 1, . . . , L and x ∈ X. Consider the complexity penalty Pen(|M|) = C1 |M|(L − 1)

  • # parameters,AIC

+ C2 |M| ln de |M|

  • log(# models of size |M|)

Theorem (upper bound)

Assume d0-sparse multinomial logistic regression model. Under the boundedness assumption, sup

η∗∈CL(d0)

E( η

M, η∗) ≤ C(δ)

  • d0(L − 1) + d0 ln
  • de

d0

  • n

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 23 / 29

slide-37
SLIDE 37

Excess risk bounds

Theorem (lower bound)

Let 2 ≤ d0 ln

  • de

d0

  • ≤ n, d0(L − 1) ≤ n and R(η∗) > 0. Then,

inf

  • η

sup

η∗∈CL(d0),f (x)

E( η, η∗) ≥ C

  • d0(L − 1) + d0 ln
  • de

d0

  • n

The idea of the proof:

1 the error cannot be smaller than that for binary classification :

Error ≥ C

  • d0 ln
  • de

d0

  • n

(see above)

2 for a given true (oracle) model with |M0| = d0 :

Error ≥ C

  • d0(L−1)

n

– via multiclass extension of VC (Daniely et al., ’12, ’15)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 24 / 29

slide-38
SLIDE 38

Two regimes

1 Small number of classes:L ≤ 2 + ln d

d0

◮ Pen(|M|) ∼ |M| ln de

|M|

◮ the error is of the order

  • d0 ln
  • de

d0

  • n

(does not depend on L, binary case)

2 Large number of classes: 2 + ln d

d0 < L < n d0

◮ Pen(|M|) ∼ |M|(L − 1) (AIC) ◮ the error is of the order

  • d0(L−1)

n

(regardless of d)

3 L > n

d0 – consistent classification is impossible

As before, the rates can be improved under the additional low-noise condition P

  • p(1)(X) − p(2)(X) ≤ h
  • ≤ Chα

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 25 / 29

slide-39
SLIDE 39

Multinomial logistic group Lasso

B has a a row-wise sparsity. Let |B|j = ||Bj·||2 , ||x||2 ≤ 1

  • BgL = arg min
  • B

   1 n

n

  • i=1
  • ln

L

  • l=1

exp( β

t l xi)

  • − xt

i

Bξi

  • + λ

d

  • j=1

| B|j    with λ ∼

  • L+ln d

n

Under the boundedness assumption, sup

η∗∈CL(d0)

E( ηgL, η∗) ≤ C(δ)

  • d0(L − 1) + d0 ln d

n (sub-optimal)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 26 / 29

slide-40
SLIDE 40

Multinomial logistic group Slope

  • BgS = arg min
  • B

   1 n

n

  • i=1
  • ln

L

  • l=1

exp( β

t l xi)

  • − xt

i

Bξi

  • +

d

  • j=1

λj| B|(j)    with λj ∼

  • L+ln(d/j)

n

Under the boundedness assumption, sup

η∗∈CL(d0)

E( ηgS, η∗) ≤ C(δ)

  • d0(L − 1) + d0 ln de

d0

n (optimal)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 27 / 29

slide-41
SLIDE 41

Future work/extensions

different types of sparsity (e.g., double sparsity: nonzero rows are also sparse – multinomial logistic sparse group Slope

  • BsgS = arg min
  • B
  • 1

n

n

  • i=1
  • ln

L

  • l=1

exp( β

t l xi)

  • − xt

i

Bξi

  • +

d

  • j=1

λj| B|(j) +

d

  • j=1

L

  • l=1

αl| Bj(l)|   

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 28 / 29

slide-42
SLIDE 42

Future work/extensions

different types of sparsity (e.g., double sparsity: nonzero rows are also sparse – multinomial logistic sparse group Slope

  • BsgS = arg min
  • B
  • 1

n

n

  • i=1
  • ln

L

  • l=1

exp( β

t l xi)

  • − xt

i

Bξi

  • +

d

  • j=1

λj| B|(j) +

d

  • j=1

L

  • l=1

αl| Bj(l)|    different types of design (e.g., Gaussian, sub-Gaussian)

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 28 / 29

slide-43
SLIDE 43

Future work/extensions

different types of sparsity (e.g., double sparsity: nonzero rows are also sparse – multinomial logistic sparse group Slope

  • BsgS = arg min
  • B
  • 1

n

n

  • i=1
  • ln

L

  • l=1

exp( β

t l xi)

  • − xt

i

Bξi

  • +

d

  • j=1

λj| B|(j) +

d

  • j=1

L

  • l=1

αl| Bj(l)|    different types of design (e.g., Gaussian, sub-Gaussian) cost-sensitive classification

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 28 / 29

slide-44
SLIDE 44

Thank You!

Felix Abramovich (Tel Aviv University) High-dimensional classification by sparse logistic regression 29 / 29