Multiresolution Matrix Factorization
Risi Kondor, The University of Chicago
Nedelina Teneva UChicago Pramod Mudrakarta UChicago
Multiresolution Matrix Factorization Risi Kondor, The University of - - PowerPoint PPT Presentation
Multiresolution Matrix Factorization Risi Kondor, The University of Chicago Nedelina Teneva Pramod Mudrakarta UChicago UChicago . Wavelets on graphs Learning on graphs Semi-supervised learning [Shuman et al., 2013]] 2 / 32 2/32 .
Risi Kondor, The University of Chicago
Nedelina Teneva UChicago Pramod Mudrakarta UChicago
.
[Shuman et al., 2013]]
2/32 .
2/32
.
3/32 .
3/32
.
Wavelets on graphs
Fast multilevel matrix algorithms
4/32 .
4/32
.
translation operator / Laplacian).
wavelet by translations and dilations.
discontinuities.
6/32 .
6/32
.
m(x) = 2−ℓ/2 ψ(2−ℓx − m)
ℓ,m αℓ
mψℓ m(x) + ∑
m βmφm(x)
7/32 .
7/32
.
Repeatedly split the space of functions on X into the direct sum of a
m
)
m
). The key to fast wavelet transforms is the that each orthogonal map
is a very sparse.
8/32 .
8/32
.
Mallat [1989] showed (roughly) that if
j Vℓ = {0},
ℓ Vℓ is dense in L2(R),
then there is a mother wavelet ψ and a father wavelet φ s. t.
m = 2−ℓ/2 ψ(2−ℓx − m)
and
m = 2−ℓ/2 φ(2−ℓx − m).
9/32 .
9/32
.
Which of the ideas from classical multiresolution still make sense?
Qℓ
m is derived by translating ψℓ → MAYBE
10/32 .
10/32
.
f∈Vℓ\{0} ⟨f, Tf⟩ / ⟨f, f⟩
increases at a given rate.
x∈X sup y∈X
m(y)
increases no faster than a certain rate.
m = ∑dim(Vℓ−1) i=1
i
m = ∑dim(Vℓ−1) i=1
i
each Qℓ orthogonal transform is sparse, guaranteeing the existence of a fast wavelet transform (Φ0 is taken to be the standard basis, φ0
m = em).
11/32 .
11/32
.
Classical approach: Define wavelets
Derive FWT MMF approach: Prescribe form of FWT
Wavelets fall out
13/32 .
13/32
.
.
1
L
Here A can be the Laplacian of a graph or any symmetric matrix.
14/32 .
14/32
.
.
1
L
The columns of Q⊤
1 Q⊤ 2 . . . Q⊤ L are a
MMF structure is a generalization of the notion of rank.
15/32 .
15/32
.
MMF reduces find the wavelet basis to an optimization problem minimize
[n] ⊇ S1 ⊇ . . . ⊇ SL H∈Hn
SL; Q1, . . . , QL∈ Q
1 . . . Q⊤ L H QL . . . Q1 ∥2
Frob.
for a given class Q of local rotations and dimensions δ1 ≥ δ2 ≥ . . . δL. Natural greedy optimization approach:
1 → Q2Q1AQ⊤ 1 Q⊤ 2 → . . . .
In practice combined with randomization and othe tricks to make it fast.
16/32 .
16/32
.
The sequence in which MMF (with k ≥ 3) eliminates dimensions induces a (soft) hierarchical clustering amongst the dimensions (mixture of trees).
17/32 .
17/32
.
18/32 .
18/32
.
heuristic approach.
different scales.
communicate with each other.
19/32 .
19/32
.
Highly optimized open source parallel C++ library:
21/32 .
21/32
.
Rows/columns are clustered, matrix is correspondingly blocked, and rotations are found within clusters. A run of rotations conforming to the same clustering structure is called a stage. .
.
1
.
.
1
Different columns of blocks (“towers”) can be sent to different processors.
22/32 .
22/32
.
After the stage is complete, rows/columns are reclustered. It is critical that reblocking also be efficient. .
. . . . .
. . . . .
.
.
.
23/32 .
23/32
.
When applying an MMF factorization to a vector, the vector must go through the same reblocking process.
.
.
.
.
24/32 .
24/32
.
25/32 .
25/32
.
26/32 .
26/32
.
27/32 .
27/32
.
28/32 .
28/32
.
[Meneveau] [Lieberman-Aiden et al., 2009]
29/32 .
29/32
.
it to less structured setting.
computational and statistical ends.
31/32 .
31/32
.
Co-authors:
Thanks:
32/32 .
32/32