Nonlinear Eigenproblems in Data Analysis and Graph Partitioning - - PowerPoint PPT Presentation

nonlinear eigenproblems in data analysis and graph
SMART_READER_LITE
LIVE PREVIEW

Nonlinear Eigenproblems in Data Analysis and Graph Partitioning - - PowerPoint PPT Presentation

Nonlinear Eigenproblems in Data Analysis and Graph Partitioning Matthias Hein Department of Mathematics and Computer Science Saarland University, Saarbr ucken, Germany Minisymposium: Modern matrix methods for large scale data and networks


slide-1
SLIDE 1

Nonlinear Eigenproblems in Data Analysis and Graph Partitioning

Matthias Hein

Department of Mathematics and Computer Science Saarland University, Saarbr¨ ucken, Germany Minisymposium: Modern matrix methods for large scale data and networks SIAM Conference on Applied Linear Algebra, Valencia 19.06.2012

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 1 / 27

slide-2
SLIDE 2

Linear Eigenproblems in Machine Learning

Motivation: Eigenvalue problems are abundant in data analysis Principal Component Analysis: Largest eigenvectors of covariance matrix of the data Usage: Denoising by projection onto largest eigenvectors. Spectral Clustering: Second smallest eigenvector of the graph Laplacian Usage: Graph partitioning using thresholded eigenvector. Latent Semantic Analysis: Singular value decomposition of term-document matrix Usage: Recover underlying latent semantic structure. Many more ... !

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 2 / 27

slide-3
SLIDE 3

The Symmetric Linear Eigenproblem

Generalized Symmetric Linear Eigenproblem: Let A, B ∈ Rn×n be symmetric and B positive definite. Then Ax = x, Ax x, Bx Bx ⇐ ⇒ x critical point of x, Ax x, Bx. Variational Principle: Courant-Fischer min-max theorem yields n eigenvalues: λm = minUm∈Um maxx∈Um x, Ax x, Bx, m = 1, . . . , n, where Um is the class of all m-dimensional subspaces of Rn. Critical point theory for ratios of quadratic functions

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 3 / 27

slide-4
SLIDE 4

Robust PCA

−8 −7 −6 −5 −4 −3 −2 −1 1 2 3 −4 −3 −2 −1 1 2 3 4

Principal Component Analysis (PCA)

first PCA component

PCA Type of eigenproblem Linear Ratio

n

  • i=1w,Xi− 1

n

n

j=1 Xj 2

w2

2 Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 4 / 27

slide-5
SLIDE 5

Robust PCA

−8 −7 −6 −5 −4 −3 −2 −1 1 2 3 −4 −3 −2 −1 1 2 3 4

Principal Component Analysis (PCA)

first PCA component (original) first PCA component (perturbed)

Source of outliers noisy data adversarial manipulation PCA Type of eigenproblem Linear Ratio

n

  • i=1w,Xi− 1

n

n

j=1 Xj 2

w2

2

Robustness no

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 4 / 27

slide-6
SLIDE 6

Robust PCA

−8 −7 −6 −5 −4 −3 −2 −1 1 2 3 −4 −3 −2 −1 1 2 3 4

Principal Component Analysis (PCA)

first PCA component (original) first PCA component (perturbed)

Source of outliers noisy data adversarial manipulation PCA Type of eigenproblem Linear Nonlinear Ratio

n

  • i=1w,Xi− 1

n

n

j=1 Xj 2

w2

2

V

  • w,X1,...,w,Xn
  • w2

Robustness no yes

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 4 / 27

slide-7
SLIDE 7

The Symmetric Linear Eigenproblem

Pros: Fast solvers available Cons: Restriction to ratio of quadratic functionals = ⇒ limited modeling abilities Quadratic functionals are non-robust against outliers (PCA). Quadratic functionals cannot induce eigenvectors which are sparse. Idea: Replace quadratic functionals by convex p-homogeneous functions !

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 5 / 27

slide-8
SLIDE 8

The Nonlinear Eigenproblem

(Homogeneous) Nonlinear Eigenproblem: Let R, S : Rn → R be convex, even and p-homogeneous (R(γx) = |γ|pR(x)) and S(x) = 0 ⇔ x = 0. Then 0 ∈ ∂R(x) − R(x) S(x) ∂S(x) ⇐ = x critical point of R(x) S(x) . Variational Principle: Lusternik-Schnirelmann min-max theorem yields n nonlinear eigenvalues: λm = minK∈Km maxx∈K R(x) S(x) , m = 1, . . . , n, where Km is the class of all compact symmetric subsets of {x ∈ Rn | S(x) > 0} with Krasnoselskii genus greater or equal to m. New: general more than n eigenvectors exist.

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 6 / 27

slide-9
SLIDE 9

The Nonlinear Eigenproblem II

Pros: Stronger modeling power using non-quadratic functions R and S Specific properties of eigenvectors like robustness against outliers or sparsity can be induced by nonsmooth choices of S respectively R. Challenges: Optimization problems for eigenproblems are typically nonconvex and nonsmooth. Need for new efficient algorithms !

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 7 / 27

slide-10
SLIDE 10

(Inverse) Power Method for Nonlinear Eigenproblems

Inverse Power Method for Linear Eigenproblems Afk+1 = B fk ⇐ ⇒ fk+1 = arg min

u∈Rn

1 2 u, Au −

  • u, Bf k

Sequence fk converges to smallest eigenvector of generalized eigenproblem. Inverse Power Method for Nonlinear Eigenproblems (H.,B.(2010))

p > 1 p = 1 gk+1= arg min

u∈Rn

{R(u) − u, s(fk)} fk+1= arg min

u2≤1

{R(u) − λk u, s(fk)} fk+1= gk+1/S(gk+1)1/p s(fk+1)∈ ∂S(fk+1) s(fk+1)∈ ∂S(fk+1) λk+1= R(fk+1)

S(fk+1)

λk+1= R(fk+1)

S(fk+1)

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 8 / 27

slide-11
SLIDE 11

Properties of Nonlinear Inverse Power Method

Theorem (Hein, B¨ uhler (2010)): It holds either λk+1 < λk

  • r λk+1 = λk and the sequence terminates. Moreover, for every cluster

point f ∗ of the sequence fk one has 0 ∈ ∂R(f ∗) − λ∗ ∂S(f ∗), where λ∗ = R(f ∗) S(f ∗) . Guarantees: monotonic descent method convergence guaranteed to some nonlinear eigenvector but not necessarily the one associated with the smallest eigenvalue.

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 9 / 27

slide-12
SLIDE 12

Benefits of Nonlinear Eigenproblems

Linear EP Nonlinear EP Modeling power low high Relaxation of loose tight combinatorial problems

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 10 / 27

slide-13
SLIDE 13

The Cheeger Cut Problem

Cheeger cut: (C, C) is a partition of the weighted, undirected graph φ(C) = cut(C, C) min{|C| ,

  • C
  • } ,

where cut(A, B) =

  • i∈A,j∈B

wij Optimal Cheeger cut, φ∗ = min

C φ(C) , is NP-hard

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 11 / 27

slide-14
SLIDE 14

Balanced Graph Cuts - Applications

Clustering/Community detection Image Segmentation

  • Parallel Computing (Matrix Reordering)
  • Hein (Saarland University)

Nonlinear Eigenproblems in Data Analysis 12 / 27

slide-15
SLIDE 15

Relaxation of Cheeger Cut Problem

Relaxation into semi-definite program with |V |3 constraints: Best known (worst case) approximation guarantee: O(

  • log |V |).

Spectral Relaxation based on graph Laplacian L L = D − W , Isoperimetric inequality (Alon, Milman (1984)) (φ∗)2 2 maxi di ≤ λ2(L) ≤ 2φ∗. there are graphs known which realize lower bound bipartitioning obtained by optimal thresholding of second eigenvector

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 13 / 27

slide-16
SLIDE 16

1-Spectral Clustering

1-graph Laplacian: The nonlinear graph 1-Laplacian ∆1 induces the functional F1(f ), F1(f ) := f , ∆1f f 1 =

1 2

n

i,j=1 wij|fi − fj|

f 1 . Theorem (Hein,B¨ uhler (2010)): Let G be connected, then min

C

cut(C, C) min{|C| ,

  • C
  • } =

min

f nonconstant

median(f)=0

F1(f ) = λ2(∆1), where λ2(∆1) is the second smallest eigenvalue of ∆1. The second eigenvector of ∆1 is the indicator vector of the optimal partition. Tight relaxation of the optimal Cheeger cut !

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 14 / 27

slide-17
SLIDE 17

Quality Guarantee

Tight relaxation of Cheeger cut: Minimization of continuous relaxation is as hard as original Cheeger cut problem = ⇒ non-convex and non-smooth No guarantee that one obtains optimal solution by NIPM !

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 15 / 27

slide-18
SLIDE 18

Quality Guarantee

Tight relaxation of Cheeger cut: Minimization of continuous relaxation is as hard as original Cheeger cut problem = ⇒ non-convex and non-smooth No guarantee that one obtains optimal solution by NIPM ! but Quality guarantee:

Theorem

Let (A, A) be a given partition of V . If one uses as initialization of NIPM, f 0 = 1A, then either NIPM terminates after one step or it yields an f 1 which after optimal thresholding gives a partition (B, B) which satisifies cut(B, B) min{|B|, |B|} < cut(A, A) min{|A|, |A|}. Next Goal: Global approximation guarantees.

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 15 / 27

slide-19
SLIDE 19

Cheeger Cut: 1-Laplacian (NLEP) vs. 2-Laplacian (LEP)

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 16 / 27

slide-20
SLIDE 20

Linear Nonlinear Ratio

n

i,j=1 wij(xi−xj)2

x2

2

n

i,j=1 wij|xi−xj|

x1

Approximation Guarantee loose tight ! Hein, B¨

uhler(2010)

Convergence globally optimal locally optimal Scalability

  • Quality

+ +++ 1-Spectral Clustering beats state of the art methods on graph partitioning benchmark

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 17 / 27

slide-21
SLIDE 21

Balanced Graph Cuts and Nonlinear Eigenproblems

Balanced graph cut problem: min

A⊂V

cut(A, A) ˆ S(A) . Balancing set function ˆ S: Name ˆ S(A) Cheeger cut min{|A|, |A|} Ratio cut |A||A| Hard balanced cut

  • 1,

if min{|A|, |A|} ≥ K 0, else. Modeling of different bias towards balanced partitions via choice of ˆ S. Do there exist tight relaxations for all balancing set functions ?

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 18 / 27

slide-22
SLIDE 22

Definition

Let f ∈ RV be ordered in increasing order f1 ≤ f2 ≤ . . . ≤ fn and define Ci = {j ∈ V | fj > fi}.Then S : RV → R given by S(f ) =

n

  • i=1

fi

  • ˆ

S(Ci−1) − ˆ S(Ci)

  • =

n−1

  • i=1

ˆ S(Ci)(fi+1 − fi) + f1 ˆ S(V ) is the Lovasz extension of ˆ S : 2V → R. One has S(1A) = ˆ S(A), ∀A ⊂ V .

Definition

A set function ˆ F : 2V → R is submodular if for all A, B ⊂ V , ˆ F(A ∪ B) + ˆ F(A ∩ B) ≤ ˆ F(A) + ˆ F(B).

Proposition

Every set function can be written as difference of two submodular functions.

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 19 / 27

slide-23
SLIDE 23

Balanced Graph Cuts as Nonlinear Eigenproblems

Theorem (Hein, Setzer (2011)) It holds minf ∈RV

1 2

n

i,j=1 wij|fi − fj|

S(f ) = minA⊂V cut(A, A) ˆ S(A) , if either one of the following two conditions holds

1 S is positively one-homogeneous, even, convex and S(f + α1) = S(f )

for all f ∈ RV , α ∈ R and ˆ S is defined as ˆ S(A) = S(1A), ∀ A ⊂ V .

2 S is the Lovasz extension of the non-negative, symmetric set function

ˆ S with ˆ S(∅) = 0. Let f ∈ RV and Ct := {i ∈ V | fi > t}, then it holds in both cases, mint∈R cut(Ct, Ct) ˆ S(Ct) ≤

1 2

n

i,j=1 wij|fi − fj|

S(f ) .

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 20 / 27

slide-24
SLIDE 24

Ratio DCA (Hein, Setzer 2011) Minimization of a non-negative ratio of 1-homogeneous d.c. functions minf ∈Rn R1(f ) − R2(f ) S1(f ) − S2(f ) . Note that for a 1-homogeneous convex function S(f ) ≥ u, f , ∀f ∈ Rn, g ∈ Rn, u ∈ ∂S(g) Minimize at each step the convex-concave ratio R1(f ) − r2, f f , s1 − S2(f ), where r2 ∈ ∂R2(f k), s1 ∈ ∂S1(f k) via Dinkelbach’s method. This yields the convex opt. problem: minf ∈D R1(f ) − r2, f + λk S2(f ) − s1, f

  • Monotonic descent and convergence to critical point as for NIPM

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 21 / 27

slide-25
SLIDE 25

Combinatorial Fractional Problems

Latest result: minC⊂V ˆ R(C) ˆ S(C)

  • subj. to :
  • Mi(C) ≥ ki,

i = 1, . . . , K has a tight relaxation into a nonlinear eigenproblem if ˆ R, ˆ S are non-negative set functions ˆ R(∅) = ˆ S(∅) = 0 The constraint functions ˆ Mi underlie no restrictions Integration of prior knowledge in clustering/community detection problems via constraints !

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 22 / 27

slide-26
SLIDE 26

Constrained Normalized Cut

Clustering with prior knowledge (Rangapuram, Hein (2012)) must-link and cannot-link constraints a partition is called consistent if all constraints are satisfied Constrained ratio cut problem: min

(C,C) consistent

cut(C, C) vol(C) vol(C) previous methods can not guarantee that constraints are satisfied

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 23 / 27

slide-27
SLIDE 27

Constrained Normalized Cut - II

Must-link and cannot-link constraints Result of unconstrained 1-Spectral Clustering (left) and constrained normalized cut (right)

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 24 / 27

slide-28
SLIDE 28

Constrained Normalized Cut - Results II

Our NLEP formulation: COSC Binary-partitioning problem (Spam dataset |V | = 4207): Multi-partitioning problem (extended MNIST dataset, |V | = 630000):

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 25 / 27

slide-29
SLIDE 29

Conclusion and Outlook

Benefits of Nonlinear Eigenproblems better integration of modeling goals using additional degrees of freedom generalized inverse power method makes computation of nonlinear eigenvectors feasible Tight relaxation of combinatorial problems as nonlinear eigenproblems Open Problems in Nonlinear Eigenproblems: What is a suitable min-max principle for nonlinear eigenvectors ? Computation of higher-order eigenvectors Theory of modeling properties of eigenvectors via choice of R and S Approximation guarantees for tight relaxations of combinatorial problems . . .

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 26 / 27

slide-30
SLIDE 30

Job Advertisement

Ph.D. and Postdoc positions ERC Starting Grant starting in Autumn Nonlinear Eigenproblems for Data Analysis Desired background in one or more of the following areas

1

convex (and non-convex)

  • ptimization

2

machine learning/statistics

3

functional analysis, variational problems

Hein (Saarland University) Nonlinear Eigenproblems in Data Analysis 27 / 27