Low rank SDP extreme points and Applications
Mohit Singh Georgia Tech
Low rank SDP extreme points and Applications Mohit Singh Georgia - - PowerPoint PPT Presentation
Low rank SDP extreme points and Applications Mohit Singh Georgia Tech SDP extreme points Pataki, Gbor. "On the rank of extreme matrices in semidefinite programs and the multiplicity of optimal eigenvalues." Math of OR 98 .
Mohit Singh Georgia Tech
the multiplicity of optimal eigenvalues." Math of OR ‘98.
quadratic maps." Discrete & Computational Geometry ‘95.
Singh, Vempala ‘18,’19]
×,
Theorem: Every extreme point has at most non-zero variables. Therefore, there exists an optimal solution that has at most non-zero variables. Numerous generalization, applications. [Neil’s Talk tomorrow]
matrix is positive semi-definite (definite), i.e. if
× with orthogonal columns and × symmetric,
diagonal with positive entries. (r=n).
. (
rank at most where
be rank where We find such that and are
× with orthonormal columns, × diagonal
with positive entries.
×, s.t. , are feasible.
Want:
Proof:
are same as eigenvalues of
and therefore, if will ensure .
But the above are
If , then there is always a non-trivial solution.
rank at most where
Farkas Lemma: Suppose
S-Lemma (Yakubovich ‘79). Suppose
× symmetric matrices such that
for some Proof: Consider the SDP. Observe that primal optimal is rank 1. Thus
. But then objective is at least 0. This
implies dual is at least 0.
Dual : Max
the multiplicity of optimal eigenvalues." Math of OR ‘98.
quadratic maps." Discrete & Computational Geometry ‘95.
Singh, Vempala ‘18,’19]
directions.
× of
data points. Want to reduce from to dimensions.
:
× of
data points. Want to reduce from to dimensions.
:
form where is projection matrix on top singular vectors of .
min
: 𝐸 − 𝑉 = min ∈𝒬 𝐸 − 𝐸𝑄 = 𝑈𝑠(𝐸𝐸) − m𝑏𝑦 ∈𝒬 𝐸𝐸 ⋅ 𝑄
𝒬 = {𝑄 ∈ 𝑆×: 𝑄 𝑡𝑧𝑛𝑛𝑓𝑢𝑠𝑗𝑑, 𝑠𝑏𝑜𝑙 𝑄 = 𝑒, 𝑄 = 𝑄}
[Sirovich, Kirby’87, Turk, Pentland’91]
[Gill et al, Nature 2014]
Standard PCA on face data LFW of male and female. Equalizing male and female weight before PCA
Data belongs to users of different groups, say images of men and women. PCA ensures average error in projection is small. Errors for two different groups are different.
and a projection matrix
×.
.
maximum error.
Fair PCA:=
∈𝒬 ∈[]
Fair PCA as rank constrained SDP.
𝑨 ≥ 𝑈𝑠 𝐸
𝐸 − 𝐸 𝐸 ⋅ 𝑄 ∀𝑗 = 1, … , 𝑙
𝑠𝑏𝑜𝑙 𝑄 = 𝑒 0 ≼ 𝑄 ≼ 𝐽
utility of each group.
Fair DR
max
∈𝒬 𝑣 𝑄 , 𝑣 𝑄 , … , 𝑣 𝑄
𝒬 = {𝑄 ∈ 𝑆×: 𝑄 𝑡𝑧𝑛𝑛𝑓𝑢𝑠𝑗𝑑, 𝑠𝑏𝑜𝑙 𝑄 = 𝑒, 𝑄 = 𝑄} Fair PCA: Special case with 𝑣 𝑄 = −𝐹𝑠𝑠 𝐸, 𝑄 and . = min 𝑀𝑝𝑡𝑡 𝑄 = 𝐸 − 𝐸𝑄
− 𝐸 − 𝐸𝑄 ∗
∗ is the best rank 𝑒 projection for group 𝑗.
[Singer’08]
is concave.
Fair PCA:= min
∈𝒬 max ∈[] 𝐹𝑠𝑠 𝐸, 𝑄 ≔ 𝐸 − 𝐸𝑄
Fair DR
max
∈𝒬 𝑣 𝑄 , 𝑣 𝑄 , … , 𝑣 𝑄
𝒬 = {𝑄 ∈ 𝑆×: 𝑄 𝑡𝑧𝑛𝑛𝑓𝑢𝑠𝑗𝑑, 𝑠𝑏𝑜𝑙 𝑄 = 𝑒, 𝑄 = 𝑄}
at most
at most whose objective is at most
, where Δ ≔ max
⊆[] ∑
𝜏
𝐸
Fair PCA:= min
∈𝒬 max ∈[] 𝐹𝑠𝑠 𝐸, 𝑄 ≔ 𝐸 − 𝐸𝑄
Theorem 1: Every extreme point of the SDP-Relaxation has rank at most Thus the objective of the two programs are identical. Corollary: Fair PCA is polynomial time solvable for 2 groups. Related: Barvinok’95, Pataki’98. S-Lemma [Yakubovich’71].
Rank-Constrained SDP min 𝐷 ⋅ 𝑌 𝐵 ⋅ 𝑌 ≤ 𝑐 𝑠𝑏𝑜𝑙 𝑌 ≤ 𝑒 0 ≼ 𝑌 ≼ 𝐽 SDP-Relaxation min 𝐷 ⋅ 𝑌 𝐵 ⋅ 𝑌 ≤ 𝑐 𝑈𝑠𝑏𝑑𝑓 𝑌 ≤ 𝑒 0 ≼ 𝑌 ≼ 𝐽
Theorem 2: Every extreme point of the SDP-Relaxation has rank at most .
returns a rank at most
Rank-Constrained SDP min 𝐷 ⋅ 𝑌 𝐵 ⋅ 𝑌 ≤ 𝑐 ∀𝑗 = 1, … , 𝑛 𝑠𝑏𝑜𝑙 𝑌 ≤ 𝑒 0 ≼ 𝑌 ≼ 𝐽 SDP-Relaxation min 𝐷 ⋅ 𝑌 𝐵 ⋅ 𝑌 ≤ 𝑐 ∀𝑗 = 1, … , 𝑛 𝑢𝑠𝑏𝑑𝑓 𝑌 ≤ 𝑒 0 ≼ 𝑌 ≼ 𝐽
Theorem: Every extreme point of the SDP-Relaxation has rank at most Thus the
Corollary: Fair PCA is polynomial time solvable for 2 groups. Related: Barvinok’95, Pataki’98. S-Lemma [Yakubovich’71].
Rank-Constrained SDP min 𝐷 ⋅ 𝑌 𝐵 ⋅ 𝑌 ≤ 𝑐 𝑠𝑏𝑜𝑙 𝑌 ≤ 𝑒 0 ≼ 𝑌 ≼ 𝐽 SDP-Relaxation min 𝐷 ⋅ 𝑌 𝐵 ⋅ 𝑌 ≤ 𝑐 𝑈𝑠𝑏𝑑𝑓 𝑌 ≤ 𝑒 0 ≼ 𝑌 ≼ 𝐽
be an extreme point with r fractional eigenvalues.
diagonal matrix with
Claim: If
symmetric matrix such that
Assuming the claim, we obtain a contradiction to extreme point. Fact: Eigenvalues of are same as eigenvalues of and eigenvalues of are same as eigenvalues of .
min 𝐷 ⋅ 𝑌 𝐵 ⋅ 𝑌 ≤ 𝑐 ∀𝑗 = 1, … , 𝑛. 𝑈𝑠 𝑌 ≤ 𝑒 0 ≼ 𝑌 ≼ 𝐽
symmetric matrix such that
If
such that all satisfy the above constraints. Consider F= for small enough .
+ U(𝐸 ± F)U satisfies the linear constraints.
= are bounded away from 0 and 1.
min 𝐷 ⋅ 𝑌 𝐵 ⋅ 𝑌 ≤ 𝑐 ∀𝑗 = 1, … , 𝑛. 𝑈𝑠 𝑌 ≤ 𝑒 0 ≼ 𝑌 ≼ 𝐽
Theorem: Every extreme point of the SDP
Generalizes Barvinok’95, Pataki’98.
Theorem: There is an iterative rounding algorithm that given min 𝐷 ⋅ 𝑌 𝐵 ⋅ 𝑌 ≤ 𝑐 ∀𝑗 = 1, … , 𝑛 𝑈𝑠 𝑌 ≤ 𝑒 0 ≼ 𝑌 ≼ 𝐽 with optimal solution 𝑌∗ returns a feasible solution 𝑍 s.t. 1. rank 𝑍 ≤ 𝑒. 2. C ⋅ 𝑍 ≤ 𝐷 ⋅ 𝑌∗ . 3. 𝐵 ⋅ 𝑍 ≥ 𝐵 ⋅ 𝑌∗ − Δ Where Δ = max
⊆[] ∑
𝜏
𝐵
∈
Idea: Fix eigenvalues to 0 and 1. Maintain two subspaces and for corresponding eigenfaces. Update SDP to work only in the orthogonal space F. Show a constraint can be removed or one of the eigenvalues is 0 or 1.
.
Thanks!