Understanding parallel analysis methods for rank selection in PCA - - PowerPoint PPT Presentation

understanding parallel analysis methods for rank
SMART_READER_LITE
LIVE PREVIEW

Understanding parallel analysis methods for rank selection in PCA - - PowerPoint PPT Presentation

Understanding parallel analysis methods for rank selection in PCA David Hong Yue Sheng Edgar Dobriban Wharton Statistics, University of Pennsylvania Random Matrices and Complex Data Analysis Workshop 10 December 2019 This work was supported


slide-1
SLIDE 1

Understanding parallel analysis methods for rank selection in PCA

David Hong Yue Sheng Edgar Dobriban Wharton Statistics, University of Pennsylvania Random Matrices and Complex Data Analysis Workshop 10 December 2019

This work was supported in part by NSF BIGDATA grant IIS 1837992 and NSF TRIPODS award 1934960.

slide-2
SLIDE 2

An illustrative example: principal components for genetics

1000G genetics data: n = 2318 individuals, p = 115019 SNPs Rounak Dey Xihong Lin

Parallel analysis for rank selection in PCA 1/22

slide-3
SLIDE 3

An illustrative example: principal components for genetics

1000G genetics data: n = 2318 individuals, p = 115019 SNPs Rounak Dey Xihong Lin PC’s can reveal population (and sub-population) structure, but how many are meaningful?

Parallel analysis for rank selection in PCA 1/22

slide-4
SLIDE 4

An illustrative example: principal components for genetics

Often, we look at the scree plot and the spectrum: Question: how can we make principled selections and reason about them?

Parallel analysis for rank selection in PCA 2/22

slide-5
SLIDE 5

An illustrative example: principal components for genetics

Often, we look at the scree plot and the spectrum: Question: how can we make principled selections and reason about them? The spectrum looks like a spiked covariance model...

Parallel analysis for rank selection in PCA 2/22

slide-6
SLIDE 6

Rank selection for PCA

Rank selection is important – it affects every downstream step! ◮ too many: add noise to downstream analyses ◮ too few: lose signals that were in the data Many excellent and practical methods: ◮ Likelihood ratio test (Bartlett 1950) ◮ Fixed threshold (Kaiser 1960) ◮ Scree plot (Cattell 1966) ◮ 4/ √ 3 (Gavish & Donoho 2014) ◮ bi-cross-validation (Owen & Wang 2016) ◮ ... Today’s talk: parallel analysis (Horn, 1965; Buja & Eyuboglu 1992)

Parallel analysis for rank selection in PCA 3/22

slide-7
SLIDE 7

Rank selection for PCA

Rank selection is important – it affects every downstream step! ◮ too many: add noise to downstream analyses ◮ too few: lose signals that were in the data Many excellent and practical methods: ◮ Likelihood ratio test (Bartlett 1950) ◮ Fixed threshold (Kaiser 1960) ◮ Scree plot (Cattell 1966) ◮ 4/ √ 3 (Gavish & Donoho 2014) ◮ bi-cross-validation (Owen & Wang 2016) ◮ ... Today’s talk: parallel analysis (Horn, 1965; Buja & Eyuboglu 1992) PA is a popular method with extensive empirical evidence, but limited theoretical understanding – exciting area for work!

Parallel analysis for rank selection in PCA 3/22

slide-8
SLIDE 8

Parallel analysis for rank selection

Parallel analysis is suggested in many reviews: ◮ Brown (2014): PA “is accurate in the vast majority of cases” ◮ Hayton et al. (2004): PA is “one of the most accurate factor retention methods” used in social science and management ◮ Costello and Osborne (2005): PA is “accurate and easy to use” ◮ Friedman et al. (2009): defaults to PA for rank selection

Parallel analysis for rank selection in PCA 4/22

slide-9
SLIDE 9

Parallel analysis for rank selection

Parallel analysis is suggested in many reviews: ◮ Brown (2014): PA “is accurate in the vast majority of cases” ◮ Hayton et al. (2004): PA is “one of the most accurate factor retention methods” used in social science and management ◮ Costello and Osborne (2005): PA is “accurate and easy to use” ◮ Friedman et al. (2009): defaults to PA for rank selection Also gaining popularity in applied statistics (esp. biological sciences): ◮ Leek and Storey (2007) ◮ Leek and Storey (2008) ◮ Lin et al. (2016) ◮ Gerard and Stephens (2017) ◮ Zhou et al. (2017) ◮ ...

Parallel analysis for rank selection in PCA 4/22

slide-10
SLIDE 10

Parallel analysis for rank selection

Parallel analysis is suggested in many reviews: ◮ Brown (2014): PA “is accurate in the vast majority of cases” ◮ Hayton et al. (2004): PA is “one of the most accurate factor retention methods” used in social science and management ◮ Costello and Osborne (2005): PA is “accurate and easy to use” ◮ Friedman et al. (2009): defaults to PA for rank selection Also gaining popularity in applied statistics (esp. biological sciences): ◮ Leek and Storey (2007) ◮ Leek and Storey (2008) ◮ Lin et al. (2016) ◮ Gerard and Stephens (2017) ◮ Zhou et al. (2017) ◮ ... But there remains limited theoretical understanding: PA is “at best a heuristic approach rather than a mathematically rigorous one” – Green et al. (2012)

Parallel analysis for rank selection in PCA 4/22

slide-11
SLIDE 11

Parallel analysis for rank selection

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly permuting the entries in each column

X

Parallel analysis for rank selection in PCA 5/22

slide-12
SLIDE 12

Parallel analysis for rank selection

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly permuting the entries in each column

X Xπ

Parallel analysis for rank selection in PCA 5/22

slide-13
SLIDE 13

Parallel analysis for rank selection

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly permuting the entries in each column

X Xπ

Parallel analysis for rank selection in PCA 5/22

slide-14
SLIDE 14

Parallel analysis for rank selection

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly permuting the entries in each column

X Xπ

Parallel analysis for rank selection in PCA 5/22

slide-15
SLIDE 15

Parallel analysis for rank selection

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly permuting the entries in each column

X Xπ

Parallel analysis for rank selection in PCA 5/22

slide-16
SLIDE 16

Parallel analysis for rank selection

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly permuting the entries in each column

X Xπ

Parallel analysis for rank selection in PCA 5/22

slide-17
SLIDE 17

Parallel analysis for rank selection

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly permuting the entries in each column

X Xπ

Parallel analysis for rank selection in PCA 5/22

slide-18
SLIDE 18

Parallel analysis for rank selection

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly permuting the entries in each column

X Xπ

Parallel analysis for rank selection in PCA 5/22

slide-19
SLIDE 19

Parallel analysis for rank selection

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly permuting the entries in each column
  • 2. Repeat several times
  • 3. Select the kth component if the kth singular value of X exceeds

the α-percentile of the kth singular value of Xπ

Parallel analysis for rank selection in PCA 5/22

slide-20
SLIDE 20

Parallel analysis for rank selection

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly permuting the entries in each column
  • 2. Repeat several times
  • 3. Select the kth component if the kth singular value of X exceeds

the α-percentile of the kth singular value of Xπ One component rises above the permuted version.

Parallel analysis for rank selection in PCA 5/22

slide-21
SLIDE 21

Parallel analysis for rank selection

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly permuting the entries in each column
  • 2. Repeat several times
  • 3. Select the kth component if the kth singular value of X exceeds

the α-percentile of the kth singular value of Xπ One component rises above the permuted version. Idea: recover “null” by destroying correlations between features.

Parallel analysis for rank selection in PCA 5/22

slide-22
SLIDE 22

A quick sneak peak...

For a larger version of the same problem, i.e., bigger n, p:

Parallel analysis for rank selection in PCA 6/22

slide-23
SLIDE 23

A quick sneak peak...

For a larger version of the same problem, i.e., bigger n, p: Permutation provides a good estimate of the noise spectrum.

Parallel analysis for rank selection in PCA 6/22

slide-24
SLIDE 24

A quick sneak peak...

For a larger version of the same problem, i.e., bigger n, p: Permutation provides a good estimate of the noise spectrum. ...let’s begin characterizing this a bit!

Parallel analysis for rank selection in PCA 6/22

slide-25
SLIDE 25

Parallel analysis under factor models

Model: data is a linear combination of factors λjk with noise εij Xij =

r

  • k=1

ηikλjk + εij,

Parallel analysis for rank selection in PCA 7/22

slide-26
SLIDE 26

Parallel analysis under factor models

Model: data is a linear combination of factors λjk with noise εij Xij =

r

  • k=1

ηikλjk + εij, i.e., low-rank signal + noise X = ηΛ⊤ + E = S + E. S

= +

Parallel analysis for rank selection in PCA 7/22

slide-27
SLIDE 27

Parallel analysis under factor models

Key idea: permutation “destroys” the signal S but not the noise E

Parallel analysis for rank selection in PCA 8/22

slide-28
SLIDE 28

Parallel analysis under factor models

Key idea: permutation “destroys” the signal S but not the noise E S Sπ ≪ S

Parallel analysis for rank selection in PCA 8/22

slide-29
SLIDE 29

Parallel analysis under factor models

Key idea: permutation “destroys” the signal S but not the noise E S Sπ ≪ S E Eπ =d E

Parallel analysis for rank selection in PCA 8/22

slide-30
SLIDE 30

Parallel analysis under factor models

Key idea: permutation “destroys” the signal S but not the noise E S Sπ ≪ S E Eπ =d E Consequence: PA estimates noise spectrum (i.e., noise floor) σk(Xπ) = σk(Sπ + Eπ) ≈ σk(Eπ) =d σk(Eπ).

Parallel analysis for rank selection in PCA 8/22

slide-31
SLIDE 31

Parallel analysis under factor models

Key idea: permutation “destroys” the signal S but not the noise E S Sπ ≪ S E Eπ =d E Consequence: PA estimates noise spectrum (i.e., noise floor) σk(Xπ) = σk(Sπ + Eπ) ≈ σk(Eπ) =d σk(Eπ). When does permutation successfully do this?

Parallel analysis for rank selection in PCA 8/22

slide-32
SLIDE 32

Important aside: small factors can fall below the noise

Example: Three factors, but only two rise above the phase transition.

Parallel analysis for rank selection in PCA 9/22

slide-33
SLIDE 33

Important aside: small factors can fall below the noise

Example: Three factors, but only two rise above the phase transition. asymptotic bulk spectrum (E → b > 0)

Parallel analysis for rank selection in PCA 9/22

slide-34
SLIDE 34

Important aside: small factors can fall below the noise

Example: Three factors, but only two rise above the phase transition. asymptotic bulk spectrum (E → b > 0) above noise factors

Parallel analysis for rank selection in PCA 9/22

slide-35
SLIDE 35

Important aside: small factors can fall below the noise

Example: Three factors, but only two rise above the phase transition. asymptotic bulk spectrum (E → b > 0) above noise factors below noise factor

Parallel analysis for rank selection in PCA 9/22

slide-36
SLIDE 36

Important aside: small factors can fall below the noise

Example: Three factors, but only two rise above the phase transition. asymptotic bulk spectrum (E → b > 0) above noise factors below noise factor Perceptible factor: singular value σk > b + δ a.s. for some δ > 0 Imperceptible factors: singular value σk < b − δ a.s. for some δ > 0

Parallel analysis for rank selection in PCA 9/22

slide-37
SLIDE 37

Important aside: small factors can fall below the noise

Example: Three factors, but only two rise above the phase transition. asymptotic bulk spectrum (E → b > 0) above noise factors below noise factor Perceptible factor: singular value σk > b + δ a.s. for some δ > 0 Imperceptible factors: singular value σk < b − δ a.s. for some δ > 0 Question: when does parallel analysis identify perceptible factors?

Parallel analysis for rank selection in PCA 9/22

slide-38
SLIDE 38

Formalizing the intuition

  • Theorem. Suppose X = S + E with signal S = ηΛ⊤ where

◮ η = UΨ1/2 for some Ψ where U ∈ Rn×r has ind. stand. entries; ◮ ΛΨ1/2 = (f1, . . . , fr) has bounded and delocalized columns, i.e., fk2 ≤ Cn1/4−δ/2 and fk4/fk2 → 0; and with noise E = ZΦ1/2 where Φ = diag(φ) is diagonal, ◮ Z ∈ Rn×p has ind. stand. entries with bounded fourth moment; ◮ entries of Z have bounded (6 + ∆)th moments; ◮ p−1

j δφj ⇒ H and maxj φj → U(H) as n, p → ∞ with p/n → γ > 0.

Then PA selects all perceptible and no imperceptible factors with prob → 1.

Parallel analysis for rank selection in PCA 10/22

slide-39
SLIDE 39

Formalizing the intuition

  • Theorem. Suppose X = S + E with signal S = ηΛ⊤ where

◮ η = UΨ1/2 for some Ψ where U ∈ Rn×r has ind. stand. entries; ◮ ΛΨ1/2 = (f1, . . . , fr) has bounded and delocalized columns, i.e., fk2 ≤ Cn1/4−δ/2 and fk4/fk2 → 0; and with noise E = ZΦ1/2 where Φ = diag(φ) is diagonal, ◮ Z ∈ Rn×p has ind. stand. entries with bounded fourth moment; ◮ entries of Z have bounded (6 + ∆)th moments; ◮ p−1

j δφj ⇒ H and maxj φj → U(H) as n, p → ∞ with p/n → γ > 0.

Then PA selects all perceptible and no imperceptible factors with prob → 1. Key: Provide conditions so that a) N → b > 0, b) Nπ =d N, c) Sπ → 0.

Parallel analysis for rank selection in PCA 10/22

slide-40
SLIDE 40

Formalizing the intuition

  • Theorem. Suppose X = S + E with signal S = ηΛ⊤ where

◮ η = UΨ1/2 for some Ψ where U ∈ Rn×r has ind. stand. entries; ◮ ΛΨ1/2 = (f1, . . . , fr) has bounded and delocalized columns, i.e., fk2 ≤ Cn1/4−δ/2 and fk4/fk2 → 0; and with noise E = ZΦ1/2 where Φ = diag(φ) is diagonal, ◮ Z ∈ Rn×p has ind. stand. entries with bounded fourth moment; ◮ entries of Z have bounded (6 + ∆)th moments; ◮ p−1

j δφj ⇒ H and maxj φj → U(H) as n, p → ∞ with p/n → γ > 0.

Then PA selects all perceptible and no imperceptible factors with prob → 1. Key: Provide conditions so that a) N → b > 0, b) Nπ =d N, c) Sπ → 0. Involved deriving new moment bounds

Parallel analysis for rank selection in PCA 10/22

slide-41
SLIDE 41

Numerical experiment

Setup: n = 500 samples with p = 300 features, r = 1 latent factor. X = θ√γηΛ⊤ + E, where η ∼ Unif(Sn−1), Λ ∼ Unif(Sp−1), and εij

iid

∼ N(0, 1/n).

Parallel analysis for rank selection in PCA 11/22

slide-42
SLIDE 42

Numerical experiment

Setup: n = 500 samples with p = 300 features, r = 1 latent factor. X = θ√γηΛ⊤ + E, where η ∼ Unif(Sn−1), Λ ∼ Unif(Sp−1), and εij

iid

∼ N(0, 1/n). Comparing against σ1(Xπ) can help combat overselection.

Parallel analysis for rank selection in PCA 11/22

slide-43
SLIDE 43

Numerical experiment

Setup: n = 500 samples with p = 300 features, r = 1 latent factor. X = θ√γηΛ⊤ + E, where η ∼ Unif(Sn−1), Λ ∼ Unif(Sp−1), and εij

iid

∼ N(0, 1/n). Comparing against σ1(Xπ) can help combat overselection.

Parallel analysis for rank selection in PCA 11/22

slide-44
SLIDE 44

What if the noise is not invariant under permutation?

Example: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

Parallel analysis for rank selection in PCA 12/22

slide-45
SLIDE 45

What if the noise is not invariant under permutation?

Example: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

Parallel analysis for rank selection in PCA 12/22

slide-46
SLIDE 46

What if the noise is not invariant under permutation?

Example: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

This heterogeneous data is less noisy, should be easier!

Parallel analysis for rank selection in PCA 12/22

slide-47
SLIDE 47

What if the noise is not invariant under permutation?

Example: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

Parallel analysis for rank selection in PCA 13/22

slide-48
SLIDE 48

What if the noise is not invariant under permutation?

Example: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

Parallel analysis for rank selection in PCA 13/22

slide-49
SLIDE 49

What if the noise is not invariant under permutation?

Example: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

But it performs much worse...

Parallel analysis for rank selection in PCA 13/22

slide-50
SLIDE 50

What if the noise is not invariant under permutation?

Example: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

But it performs much worse...what is happening?

Parallel analysis for rank selection in PCA 13/22

slide-51
SLIDE 51

What if the noise is not invariant under permutation?

Example: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

Parallel analysis for rank selection in PCA 14/22

slide-52
SLIDE 52

What if the noise is not invariant under permutation?

Example: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

Permutation shrinks the noise spectrum, leading to overselection.

Parallel analysis for rank selection in PCA 14/22

slide-53
SLIDE 53

Idea: Replace permutation with signflips → Signflip PA

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly sign-flipping all entries

X R ◦ X

Parallel analysis for rank selection in PCA 15/22

slide-54
SLIDE 54

Idea: Replace permutation with signflips → Signflip PA

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly sign-flipping all entries
  • 2. Repeat several times
  • 3. Select the kth component if the kth singular value of X exceeds

the α-percentile of the kth singular value of Xπ One component rises above the signflipped version.

Parallel analysis for rank selection in PCA 15/22

slide-55
SLIDE 55

Idea: Replace permutation with signflips → Signflip PA

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly sign-flipping all entries
  • 2. Repeat several times
  • 3. Select the kth component if the kth singular value of X exceeds

the α-percentile of the kth singular value of Xπ

Parallel analysis for rank selection in PCA 15/22

slide-56
SLIDE 56

Idea: Replace permutation with signflips → Signflip PA

Given: data matrix X ∈ Rn×p and percentile α ∈ [0, 1]

  • 1. Generate Xπ by randomly sign-flipping all entries
  • 2. Repeat several times
  • 3. Select the kth component if the kth singular value of X exceeds

the α-percentile of the kth singular value of Xπ Sign-flipping also recovers the “null” by destroying correlations.

Parallel analysis for rank selection in PCA 15/22

slide-57
SLIDE 57

Idea: Replace permutation with signflips → Signflip PA

For a larger version of the same problem, i.e., bigger n, p: Signflip PA also provides a good estimate of the noise spectrum.

Parallel analysis for rank selection in PCA 16/22

slide-58
SLIDE 58

Revisit: PA for the heterogeneous example

Recall: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

Permutation shrinks the noise spectrum, leading to overselection.

Parallel analysis for rank selection in PCA 17/22

slide-59
SLIDE 59

Revisit: Signflip PA for the heterogeneous example

Recall: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

Parallel analysis for rank selection in PCA 18/22

slide-60
SLIDE 60

Revisit: Signflip PA for the heterogeneous example

Recall: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

Signflips preserve the noise spectrum (in distribution).

Parallel analysis for rank selection in PCA 18/22

slide-61
SLIDE 61

Revisit: Signflip PA for the heterogeneous example

Recall: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

Parallel analysis for rank selection in PCA 19/22

slide-62
SLIDE 62

Revisit: Signflip PA for the heterogeneous example

Recall: εij

ind

∼ N(0, ω2

i /n), 90% have ω2 i = 0.4, 10% have ω2 i = 1.

Preserving the noise distribution with signflips addresses the

  • verselection of permutation.

Parallel analysis for rank selection in PCA 19/22

slide-63
SLIDE 63

Application to single cell RNA sequencing

Work with: Thomas Zhang, George Linderman, Yuval Kluger (Yale) Question: how to select rank for single-cell RNA sequencing data? Challenge: data does not (readily) fit our signal + noise setups.

Parallel analysis for rank selection in PCA 20/22

slide-64
SLIDE 64

Application to single cell RNA sequencing

Work with: Thomas Zhang, George Linderman, Yuval Kluger (Yale) Question: how to select rank for single-cell RNA sequencing data? Challenge: data does not (readily) fit our signal + noise setups. Model: n samples are drawn independently from a multinomial xi

ind

∼ Multinomial(si, ki), where S = (s1, . . . , sn)⊤ is row-stochastic and low-rank.

Parallel analysis for rank selection in PCA 20/22

slide-65
SLIDE 65

Application to single cell RNA sequencing

Work with: Thomas Zhang, George Linderman, Yuval Kluger (Yale) Question: how to select rank for single-cell RNA sequencing data? Challenge: data does not (readily) fit our signal + noise setups. Model: n samples are drawn independently from a multinomial xi

ind

∼ Multinomial(si, ki), where S = (s1, . . . , sn)⊤ is row-stochastic and low-rank. Writing it in a signal + noise form X = S + (X − S) = S + N, where N = X − S is centered (since EX = S), but has dep. entries.

Parallel analysis for rank selection in PCA 20/22

slide-66
SLIDE 66

Application to single cell RNA sequencing

Work with: Thomas Zhang, George Linderman, Yuval Kluger (Yale) Question: how to select rank for single-cell RNA sequencing data? Challenge: data does not (readily) fit our signal + noise setups. Model: n samples are drawn independently from a multinomial xi

ind

∼ Multinomial(si, ki), where S = (s1, . . . , sn)⊤ is row-stochastic and low-rank. Writing it in a signal + noise form X = S + (X − S) = S + N, where N = X − S is centered (since EX = S), but has dep. entries. Ongoing work: how do our insights about PA apply here?

Parallel analysis for rank selection in PCA 20/22

slide-67
SLIDE 67

Application to single cell RNA sequencing

Prelim experiment: rank-10 S matrix, diverse total count rates, ...

Parallel analysis for rank selection in PCA 21/22

slide-68
SLIDE 68

Application to single cell RNA sequencing

Prelim experiment: rank-10 S matrix, diverse total count rates, ...

Parallel analysis for rank selection in PCA 21/22

slide-69
SLIDE 69

Application to single cell RNA sequencing

Prelim experiment: rank-10 S matrix, diverse total count rates, ... Permutations seem to shrink the noise spectrum sometimes and signflips seem to preserve them...

Parallel analysis for rank selection in PCA 21/22

slide-70
SLIDE 70

Application to single cell RNA sequencing

Prelim experiment: rank-10 S matrix, diverse total count rates, ... Permutations seem to shrink the noise spectrum sometimes and signflips seem to preserve them... Ongoing: theoretical analysis/characterization – how to deal with the dependence among noise entries?

Parallel analysis for rank selection in PCA 21/22

slide-71
SLIDE 71

Conclusions

Today: ◮ explaination for how parallel analysis works using insights/tools from random matrix theory ◮ some theoretical guarantees/characterization for parallel analysis ◮ signflip variant to handle alternative noise models ◮ preliminary work on applications to scRNAseq Ongoing: ◮ characterization/analysis of signflip parallel analysis ◮ characterization of behavior under multinomial models ◮ application of similar ideas to other models? ◮ more evaluation in real data

Parallel analysis for rank selection in PCA 22/22

slide-72
SLIDE 72

Conclusions

Today: ◮ explaination for how parallel analysis works using insights/tools from random matrix theory ◮ some theoretical guarantees/characterization for parallel analysis ◮ signflip variant to handle alternative noise models ◮ preliminary work on applications to scRNAseq Ongoing: ◮ characterization/analysis of signflip parallel analysis ◮ characterization of behavior under multinomial models ◮ application of similar ideas to other models? ◮ more evaluation in real data Thanks!

Parallel analysis for rank selection in PCA 22/22