Discovery of Latent Factors in High-dimensional Data via Spectral Methods
Furong Huang University of Maryland
Workshop on Quantum Machine Learning
1 / 39
Discovery of Latent Factors in High-dimensional Data via Spectral - - PowerPoint PPT Presentation
Discovery of Latent Factors in High-dimensional Data via Spectral Methods Furong Huang University of Maryland Workshop on Quantum Machine Learning 1 / 39 Machine Learning - Excitements Success of Supervised Learning Image classification
Workshop on Quantum Machine Learning
1 / 39
2 / 39
2 / 39
2 / 39
◮ State-of-the-art: Humans are better than machines ◮ Goal: Intelligent machines that summarize key features in data
◮ Theoretically guaranteed learning ◮ Extracted features are interpretable 2 / 39
3 / 39
Cell T ypes T
Communities
3 / 39
Cell T ypes T
Communities
3 / 39
Cell T ypes T
Communities
3 / 39
Topics Education Crime Sports
4 / 39
Topics Education Crime Sports
4 / 39
5 / 39
5 / 39
103 104 105
Perplexity Tensor Variational
2 4 6 8 10 ×104
Running Time (s) 6 / 39
103 104 105
Perplexity Tensor Variational
2 4 6 8 10 ×104
Running Time (s)
Facebook: n ∼ 20k Yelp: n ∼ 40k DBLPsub: n ∼ 0.1m DBLP: n ∼ 1m
10-2 10-1 100 101
Error /group FB YP DBLPsub DBLP
102 103 104 105 106
Running Times (s) FB YP DBLPsub DBLP 6 / 39
103 104 105
Perplexity Tensor Variational
2 4 6 8 10 ×104
Running Time (s)
Facebook: n ∼ 20k Yelp: n ∼ 40k DBLPsub: n ∼ 0.1m DBLP: n ∼ 1m
10-2 10-1 100 101
Error /group FB YP DBLPsub DBLP
102 103 104 105 106
Running Times (s) FB YP DBLPsub DBLP
“Online Tensor Methods for Learning Latent Variable Models”, F. Huang, U. Niranjan, M. Hakeem, A. Anandkumar, JMLR14. “Tensor Methods on Apache Spark”, F. Huang, A. Anandkumar, Oct. 2015. 6 / 39
7 / 39
0.5 1.0 1.5 2.0 2.5 k Spatial point process (ours) Average expression level ( ) previous
“Discovering Neuronal Cell Types and Their Gene Expression Profiles Using a Spatial Point Process Mixture Model”, F. Huang,
8 / 39
football soccer tree
The weather is good. Her life spanned years of incredible change for women. Mary lived through an era of liberating reform for women.
“Convolutional Dictionary Learning through Tensor Factorization”, by F. Huang, A. Anandkumar, In Proceedings of JMLR 2015. 9 / 39
” Scalable Latent TreeModel and its Application to Health Analytics ” by F. Huang, N. U.Niranjan, I. Perros, R. Chen, J. Sun,
10 / 39
11 / 39
12 / 39
Words Topics Choice Variable life gene data DNA RNA k1 k2 k3 k4 k5 h A A A A A
Unlabeled data Latent variable model Learning Algorithm Inference 13 / 39
Words Topics Choice Variable life gene data DNA RNA k1 k2 k3 k4 k5 h A A A A A
Unlabeled data Latent variable model Learning Algorithm Inference
13 / 39
Words Topics Choice Variable life gene data DNA RNA k1 k2 k3 k4 k5 h A A A A A
Unlabeled data Latent Variable model MCMC Inference
◮ Exponential mixing time 13 / 39
Words Topics Choice Variable life gene data DNA RNA k1 k2 k3 k4 k5 h A A A A A
Unlabeled data Latent variable model
▲ ✁el ✐✂ ✂ ✄ ☎ ✆ t ✐✂ ✄✝Inference
◮ Exponential mixing time
◮ Exponential critical points 13 / 39
Words Topics Choice Variable life gene data DNA RNA k1 k2 k3 k4 k5 h A A A A A
Unlabeled data Latent variable model
✞ ✟ ✠el ✟ ✡☛ ☛ ☞ ✌ ✍ ✎ ✡☛ ☞✏Inference
◮ Exponential mixing time
◮ Exponential critical points
13 / 39
Words Topics Choice Variable life gene data DNA RNA k1 k2 k3 k4 k5 h A A A A A
Unlabeled data Latent variable model
❚ensor Decomposition Inference
= + +
◮ Exponential mixing time
◮ Exponential critical points
13 / 39
Words Topics Choice Variable life gene data DNA RNA k1 k2 k3 k4 k5 h A A A A A
Unlabeled data Latent variable model T ensor Decomposition Inference
= + +
14 / 39
Words Topics Choice Variable life gene data DNA RNA k1 k2 k3 k4 k5 h A A A A A
Unlabeled data Latent variable model T ensor Decomposition Inference
= + +
14 / 39
1
2
3
4
5
15 / 39
1
◮ Moments
Eθ[f(X)]
2
i=1):
◮ Empirical Moments
3
◮ Moment matching
Eθ[f(X)] n→∞ =
i=1 ∼ N(µ, Σ2)?
16 / 39
17 / 39
18 / 39
19 / 39
=
20 / 39
R
21 / 39
1
1 + e2e⊤ 2 = u1u⊤ 1 + u2u⊤ 2
√ 2 2 , − √ 2 2 ]
√ 2 2 , √ 2 2 ]
22 / 39
1
1 + e2e⊤ 2 = u1u⊤ 1 + u2u⊤ 2
Unique with eigenvalue gap
√ 2 2 , − √ 2 2 ]
√ 2 2 , √ 2 2 ]
22 / 39
1
1 + e2e⊤ 2 = u1u⊤ 1 + u2u⊤ 2
Unique with eigenvalue gap
√ 2 2 , − √ 2 2 ]
√ 2 2 , √ 2 2 ]
22 / 39
1
1 + e2e⊤ 2 = u1u⊤ 1 + u2u⊤ 2
Unique with eigenvalue gap
√ 2 2 , − √ 2 2 ]
√ 2 2 , √ 2 2 ]
22 / 39
1
1 + e2e⊤ 2 = u1u⊤ 1 + u2u⊤ 2
Unique with eigenvalue gap
√ 2 2 , − √ 2 2 ]
√ 2 2 , √ 2 2 ]
22 / 39
1
2
3
4
5
23 / 39
1
2
3
4
5
24 / 39
Infer topics of documents Learn hidden process drives the
Topic proportion ∼ Dir(α) for a doc Draw a topic, then a word for a token
Topics Topic Proportion
police witness campus police witness campus police witness campus 25 / 39
Infer topics of documents Learn hidden process drives the
Topic proportion ∼ Dir(α) for a doc Draw a topic, then a word for a token
Topics Topic Proportion
police witness campus police witness campus police witness campus police witness
crime S p
t s Educa
✑campus 25 / 39
Infer topics of documents Learn hidden process drives the
Topic proportion ∼ Dir(α) for a doc Draw a topic, then a word for a token
Topics Topic Proportion
police witness campus police witness campus police witness campus police witness
crime S p
t s Educa
✒campus
campus police witness
25 / 39
campus police witness E[word|topic = j] =
P[word = ei|topic = j]ei =column j
26 / 39
campus police witness E[word|topic = j] =
P[word = ei|topic = j]ei =column j
E[word] =
E[word|topic = j]P[topic = j]
campus police witness
crime Sports Educa
✓campus
♣ ✔ ✕ ✖ ✗e
✇ ✖ ✘ ✙ ✚ ✛ ✛campus po
❧ice
✜itness 26 / 39
campus police witness E[word|topic = j] =
P[word = ei|topic = j]ei =column j
E[word] =
E[word|topic = j]P[topic = j]
campus police witness
crime Sports Educa
✢campus
✣ ✤ ✥ ✦ ✧e
★ ✦ ✩ ✪ ✫ ✬ ✬campus po
✭ice
✮itness
26 / 39
campus police witness E[word|topic = j] =
P[word = ei|topic = j]ei =column j
E[word1 ⊗ word2] =
E[word1|topic1 = j] ⊗ E[word2|topic2 = j′]P[topic1 = j, topic2 = j′]
campus police witness
crime Sports Educa
✯campus
✰ ✱ ✲ ✳ ✴e
✵ ✳ ✶ ✷ ✸ ✹ ✹campus po
✺ice
✻itness 26 / 39
campus police witness E[word|topic = j] =
P[word = ei|topic = j]ei =column j
E[word1 ⊗ word2] =
E[word1|topic1 = j] ⊗ E[word2|topic2 = j′]P[topic1 = j, topic2 = j′]
campus police witness
crime Sports Educa
✼campus
✽ ✾ ✿ ❀ ❁e
❂ ❀ ❃ ❄ ❅ ❆ ❆campus po
❇ice
❈itness
26 / 39
W W W
E[word1 ⊗ word2] =
E[word1|topic1 = j] ⊗ E[word2|topic2 = j′]P[topic1 = j, topic2 = j′]
campus police witness
crime Sports Educa
❉campus
❊ ❋e
❏ ❍ ❑ ▼ ◆ ❖ ❖campus po
Pice
◗itness
26 / 39
W W W
26 / 39
W W W
26 / 39
27 / 39
cians
❢ ❣ ❤e t a r i a n s
❥ ❦ ♠ ♥ ♦ ♥ q r ♠ s ✉ ✈ ① ② ③ ④ ④ ⑤ ⑥ ⑦ ⑧ ⑨ ⑩ ❶ ❷ ❸ace
❹ ❺ ❻ ❼ ❽a
❾ ❿ ➀27 / 39
cians
➔ → ➣e t a r i a n s
↔ ↕ ➙ ➛ ➜ ➛ ➝ ➞ ➙ ➟ ➠ ➡ ➢ ➤ ➥ ➦ ➦ ➧ ➨ ➩ ➫ ➭ ➯ ➲ ➳ ➵ace
➸ ➺ ➻ ➼ ➽a
➾ ➚ ➪27 / 39
1
2
3
4
5
28 / 39
29 / 39
i
i aj = 0
29 / 39
i
i aj = 0
29 / 39
i
i aj = 0
∀i,ui2=1
29 / 39
i
i aj = 0
∀i,ui2=1
29 / 39
i
i aj = 0
∀i,ui2=1
29 / 39
i
i aj = 0
∀i,ui2=1
29 / 39
“Escaping From Saddle Points — Online Stochastic Gradient for Tensor Decomposition”,by R. Ge, F. Huang, C. Jin, Y. Yuan, COLT 2015. 30 / 39
escape stuck
“Escaping From Saddle Points — Online Stochastic Gradient for Tensor Decomposition”,by R. Ge, F. Huang, C. Jin, Y. Yuan, COLT 2015. 30 / 39
escape stuck
“Escaping From Saddle Points — Online Stochastic Gradient for Tensor Decomposition”,by R. Ge, F. Huang, C. Jin, Y. Yuan, COLT 2015. 30 / 39
escape stuck
“Escaping From Saddle Points — Online Stochastic Gradient for Tensor Decomposition”,by R. Ge, F. Huang, C. Jin, Y. Yuan, COLT 2015. 30 / 39
1
2
3
4
5
31 / 39
i=1, where xi ∈ Rd
u∈Rd,u2=1 u⊤Au,
m m
32 / 39
◮ ordered eigenvalues 1 ≥ λ1 ≥ . . . λd ≥ 0 ◮ and corresponding eigenvectors u1, . . . , ud.
◮ Power method
Akv0 Akv0 takes O( sd ∆ log( 1 φǫ))
◮ Lanczos method or accelerated power method takes O( sd
√ ∆ log( 1 φǫ))
⋆ Replacing the monomial Ak by its Chebyshev polynomial approximation
33 / 39
◮ exponential speed-up for factoring integers
◮ quadratic speed-up for searching in unstructured database
◮ Ω(d) → poly(log d) for solving d-dimensional linear equation systems. ◮ weaker output requirement ⋆ a quantum state whose vector representation is roughly the solution to
the linear equation system.
34 / 39
φǫ) + log3.5( s φǫ))/φ
35 / 39
◮ Quantum-walk ⋆ effectively constructs a degree-m Chebyshev polynomial of A/s. ◮ Quantum primitive: the linear combination of unitaries (LCU) ⋆ effectively linearly combines these Chebyshev polynomials to derive the
desired approximation polynomial.
36 / 39
1
2
3
4
5
37 / 39
38 / 39
= + + = + + = + + = + + M3 f1 sf1 f2 sf2
= +...+ + +...+
escape stuck
38 / 39
39 / 39