statistical estimation in the presence of group actions
play

Statistical Estimation in the Presence of Group Actions Alex Wein - PowerPoint PPT Presentation

Statistical Estimation in the Presence of Group Actions Alex Wein MIT Mathematics 1 / 39 In memoriam Amelia Perry 1991 2018 2 / 39 My research interests Statistical and computational limits of average-case inference problems


  1. A simple model: Gaussian Z / 2 synchronization ◮ G = Z / 2 = {± 1 } ◮ True signal x ∈ {± 1 } n (vector of group elements) ◮ For each i , j observe x i x j + N (0 , σ 2 ) ◮ Specifically, observe n × n matrix Y = λ + 1 n xx ⊤ √ nW � �� � � �� � signal ◮ λ ≥ 0 – signal-to-noise parameter noise ◮ W – random noise matrix: symmetric with entries N (0 , 1) ◮ Y ij is a noisy measurement of x i x j (same/diff) ◮ Normalization: MMSE is a constant (depending on λ ) This is a spiked Wigner model: in general x i ∼ P (some prior) Statistical physics makes extremely precise (non-rigorous) predictions about this type of problem ◮ Often later proved correct 8 / 39

  2. A simple model: Gaussian Z / 2 synchronization ◮ G = Z / 2 = {± 1 } ◮ True signal x ∈ {± 1 } n (vector of group elements) ◮ Observe n × n matrix Y = λ + 1 n xx ⊤ √ nW � �� � � �� � signal noise Image credit: [Deshpande, Abbe, Montanari ’15] 9 / 39

  3. Statistical physics and inference What does statistical physics have to do with Bayesian inference? 10 / 39

  4. Statistical physics and inference What does statistical physics have to do with Bayesian inference? n xx ⊤ + In inference, observe Y = λ 1 √ n W and want to infer x 10 / 39

  5. Statistical physics and inference What does statistical physics have to do with Bayesian inference? n xx ⊤ + In inference, observe Y = λ 1 √ n W and want to infer x Posterior distribution: Pr [ x | Y ] ∝ exp( λ x ⊤ Yx ) 10 / 39

  6. Statistical physics and inference What does statistical physics have to do with Bayesian inference? n xx ⊤ + In inference, observe Y = λ 1 √ n W and want to infer x Posterior distribution: Pr [ x | Y ] ∝ exp( λ x ⊤ Yx ) In physics, this is called a Boltzmann/Gibbs distribution: Pr [ x ] ∝ exp( − β H ( x )) ◮ Energy (“Hamiltonian”) H ( x ) = − x ⊤ Yx ◮ Temperature β = λ 10 / 39

  7. Statistical physics and inference What does statistical physics have to do with Bayesian inference? n xx ⊤ + In inference, observe Y = λ 1 √ n W and want to infer x Posterior distribution: Pr [ x | Y ] ∝ exp( λ x ⊤ Yx ) In physics, this is called a Boltzmann/Gibbs distribution: Pr [ x ] ∝ exp( − β H ( x )) ◮ Energy (“Hamiltonian”) H ( x ) = − x ⊤ Yx ◮ Temperature β = λ So posterior distribution of Bayesian inference obeys the same equations as a disordered physical system (e.g. magnet, spin glass) 10 / 39

  8. BP and AMP “Axiom” from statistical physics: the best algorithm for every* problem is BP (belief propagation) [1] [1] Pearl ’82 [2] Donoho, Maleki, Montanari ’09 11 / 39

  9. BP and AMP “Axiom” from statistical physics: the best algorithm for every* problem is BP (belief propagation) [1] ◮ Each unknown x i is a “node” [1] Pearl ’82 [2] Donoho, Maleki, Montanari ’09 11 / 39

  10. BP and AMP “Axiom” from statistical physics: the best algorithm for every* problem is BP (belief propagation) [1] ◮ Each unknown x i is a “node” ◮ Each observation (“interaction”) Y ij is an “edge” ◮ In our case, a complete graph [1] Pearl ’82 [2] Donoho, Maleki, Montanari ’09 11 / 39

  11. BP and AMP “Axiom” from statistical physics: the best algorithm for every* problem is BP (belief propagation) [1] ◮ Each unknown x i is a “node” ◮ Each observation (“interaction”) Y ij is an “edge” ◮ In our case, a complete graph ◮ Nodes iteratively pass “messages” or “beliefs” to each other along edges, and then update their own beliefs [1] Pearl ’82 [2] Donoho, Maleki, Montanari ’09 11 / 39

  12. BP and AMP “Axiom” from statistical physics: the best algorithm for every* problem is BP (belief propagation) [1] ◮ Each unknown x i is a “node” ◮ Each observation (“interaction”) Y ij is an “edge” ◮ In our case, a complete graph ◮ Nodes iteratively pass “messages” or “beliefs” to each other along edges, and then update their own beliefs ◮ Hard to analyze [1] Pearl ’82 [2] Donoho, Maleki, Montanari ’09 11 / 39

  13. BP and AMP “Axiom” from statistical physics: the best algorithm for every* problem is BP (belief propagation) [1] ◮ Each unknown x i is a “node” ◮ Each observation (“interaction”) Y ij is an “edge” ◮ In our case, a complete graph ◮ Nodes iteratively pass “messages” or “beliefs” to each other along edges, and then update their own beliefs ◮ Hard to analyze In our case (since interactions are “dense”), we can use a simplification of BP called AMP (approximate message passing) [2] [1] Pearl ’82 [2] Donoho, Maleki, Montanari ’09 11 / 39

  14. BP and AMP “Axiom” from statistical physics: the best algorithm for every* problem is BP (belief propagation) [1] ◮ Each unknown x i is a “node” ◮ Each observation (“interaction”) Y ij is an “edge” ◮ In our case, a complete graph ◮ Nodes iteratively pass “messages” or “beliefs” to each other along edges, and then update their own beliefs ◮ Hard to analyze In our case (since interactions are “dense”), we can use a simplification of BP called AMP (approximate message passing) [2] ◮ Easy/possible to analyze [1] Pearl ’82 [2] Donoho, Maleki, Montanari ’09 11 / 39

  15. BP and AMP “Axiom” from statistical physics: the best algorithm for every* problem is BP (belief propagation) [1] ◮ Each unknown x i is a “node” ◮ Each observation (“interaction”) Y ij is an “edge” ◮ In our case, a complete graph ◮ Nodes iteratively pass “messages” or “beliefs” to each other along edges, and then update their own beliefs ◮ Hard to analyze In our case (since interactions are “dense”), we can use a simplification of BP called AMP (approximate message passing) [2] ◮ Easy/possible to analyze ◮ Provably optimal mean squared error for many problems [1] Pearl ’82 [2] Donoho, Maleki, Montanari ’09 11 / 39

  16. AMP for Z / 2 synchronization Y = λ n xx ⊤ + 1 x ∈ {± 1 } n √ nW , 12 / 39

  17. AMP for Z / 2 synchronization Y = λ n xx ⊤ + 1 x ∈ {± 1 } n √ nW , AMP algorithm: ◮ State v ∈ R n – estimate for x 12 / 39

  18. AMP for Z / 2 synchronization Y = λ n xx ⊤ + 1 x ∈ {± 1 } n √ nW , AMP algorithm: ◮ State v ∈ R n – estimate for x ◮ Initialize v to small random vector 12 / 39

  19. AMP for Z / 2 synchronization Y = λ n xx ⊤ + 1 x ∈ {± 1 } n √ nW , AMP algorithm: ◮ State v ∈ R n – estimate for x ◮ Initialize v to small random vector ◮ Repeat: 1. Power iteration: v ← Yv (power iteration) 12 / 39

  20. AMP for Z / 2 synchronization Y = λ n xx ⊤ + 1 x ∈ {± 1 } n √ nW , AMP algorithm: ◮ State v ∈ R n – estimate for x ◮ Initialize v to small random vector ◮ Repeat: 1. Power iteration: v ← Yv (power iteration) 2. Onsager: v ← v + [Onsager term] 12 / 39

  21. AMP for Z / 2 synchronization Y = λ n xx ⊤ + 1 x ∈ {± 1 } n √ nW , AMP algorithm: ◮ State v ∈ R n – estimate for x ◮ Initialize v to small random vector ◮ Repeat: 1. Power iteration: v ← Yv (power iteration) 2. Onsager: v ← v + [Onsager term] 3. Entrywise soft projection: v i ← tanh( λ v i ) (for all i ) ◮ Resulting values in [ − 1 , 1] 12 / 39

  22. AMP is optimal Y = λ n xx ⊤ + 1 x ∈ {± 1 } n √ nW , For Z / 2 synchronization, AMP is provably optimal. Deshpande, Abbe, Montanari, ’15 13 / 39

  23. Free energy landscapes What do physics predictions look like? Lesieur, Krzakala, Zdeborov´ a ’15 14 / 39

  24. Free energy landscapes What do physics predictions look like? � γ λ 2 γ 2 � � � � 1 1 � log(2 cosh( γ + √ γ z )) f ( γ ) = λ 4 + 1 + γ λ 2 + 1 E − − 4 2 λ z ∼N (0 , 1) Lesieur, Krzakala, Zdeborov´ a ’15 14 / 39

  25. Free energy landscapes What do physics predictions look like? � γ λ 2 γ 2 � � � � 1 1 � log(2 cosh( γ + √ γ z )) f ( γ ) = λ 4 + 1 + γ λ 2 + 1 − − E λ 4 2 z ∼N (0 , 1) x-axis γ : correlation with true signal (related to MSE) y-axis f : free energy – AMP’s “objective function” (minimize) Lesieur, Krzakala, Zdeborov´ a ’15 14 / 39

  26. Free energy landscapes What do physics predictions look like? � γ λ 2 γ 2 � � � � 1 1 � log(2 cosh( γ + √ γ z )) f ( γ ) = λ 4 + 1 + γ λ 2 + 1 − − E λ 4 2 z ∼N (0 , 1) x-axis γ : correlation with true signal (related to MSE) y-axis f : free energy – AMP’s “objective function” (minimize) AMP – gradient descent starting from γ = 0 (left side) STAT (statistical) – global minimum Lesieur, Krzakala, Zdeborov´ a ’15 14 / 39

  27. Free energy landscapes What do physics predictions look like? � γ λ 2 γ 2 � � � � 1 1 � log(2 cosh( γ + √ γ z )) f ( γ ) = λ 4 + 1 + γ λ 2 + 1 − − E λ 4 2 z ∼N (0 , 1) x-axis γ : correlation with true signal (related to MSE) y-axis f : free energy – AMP’s “objective function” (minimize) AMP – gradient descent starting from γ = 0 (left side) STAT (statistical) – global minimum So yields computational and statistical MSE for each λ Lesieur, Krzakala, Zdeborov´ a ’15 14 / 39

  28. Our contributions Joint work with Amelia Perry, Afonso Bandeira, Ankur Moitra Perry, W., Bandeira, Moitra, Message-passing algorithms for synchronization problems over compact groups , to appear in CPAM Perry, W., Bandeira, Moitra, Optimality and Sub-optimality of PCA for Spiked Random Matrices and Synchronization , part I to appear in Ann. Stat 15 / 39

  29. Our contributions Joint work with Amelia Perry, Afonso Bandeira, Ankur Moitra ◮ Using representation theory we define a very general Gaussian observation model for synchronization over any compact group Perry, W., Bandeira, Moitra, Message-passing algorithms for synchronization problems over compact groups , to appear in CPAM Perry, W., Bandeira, Moitra, Optimality and Sub-optimality of PCA for Spiked Random Matrices and Synchronization , part I to appear in Ann. Stat 15 / 39

  30. Our contributions Joint work with Amelia Perry, Afonso Bandeira, Ankur Moitra ◮ Using representation theory we define a very general Gaussian observation model for synchronization over any compact group ◮ Significantly generalizes Z / 2 case Perry, W., Bandeira, Moitra, Message-passing algorithms for synchronization problems over compact groups , to appear in CPAM Perry, W., Bandeira, Moitra, Optimality and Sub-optimality of PCA for Spiked Random Matrices and Synchronization , part I to appear in Ann. Stat 15 / 39

  31. Our contributions Joint work with Amelia Perry, Afonso Bandeira, Ankur Moitra ◮ Using representation theory we define a very general Gaussian observation model for synchronization over any compact group ◮ Significantly generalizes Z / 2 case ◮ We give a precise analysis of the statistical and computational limits of this model Perry, W., Bandeira, Moitra, Message-passing algorithms for synchronization problems over compact groups , to appear in CPAM Perry, W., Bandeira, Moitra, Optimality and Sub-optimality of PCA for Spiked Random Matrices and Synchronization , part I to appear in Ann. Stat 15 / 39

  32. Our contributions Joint work with Amelia Perry, Afonso Bandeira, Ankur Moitra ◮ Using representation theory we define a very general Gaussian observation model for synchronization over any compact group ◮ Significantly generalizes Z / 2 case ◮ We give a precise analysis of the statistical and computational limits of this model ◮ Uses non-rigorous (but well-established) ideas from statistical physics ◮ Methods proven correct in related settings Perry, W., Bandeira, Moitra, Message-passing algorithms for synchronization problems over compact groups , to appear in CPAM Perry, W., Bandeira, Moitra, Optimality and Sub-optimality of PCA for Spiked Random Matrices and Synchronization , part I to appear in Ann. Stat 15 / 39

  33. Our contributions Joint work with Amelia Perry, Afonso Bandeira, Ankur Moitra ◮ Using representation theory we define a very general Gaussian observation model for synchronization over any compact group ◮ Significantly generalizes Z / 2 case ◮ We give a precise analysis of the statistical and computational limits of this model ◮ Uses non-rigorous (but well-established) ideas from statistical physics ◮ Methods proven correct in related settings ◮ Includes an AMP algorithm which we believe is optimal among all polynomial-time algorithms Perry, W., Bandeira, Moitra, Message-passing algorithms for synchronization problems over compact groups , to appear in CPAM Perry, W., Bandeira, Moitra, Optimality and Sub-optimality of PCA for Spiked Random Matrices and Synchronization , part I to appear in Ann. Stat 15 / 39

  34. Our contributions Joint work with Amelia Perry, Afonso Bandeira, Ankur Moitra ◮ Using representation theory we define a very general Gaussian observation model for synchronization over any compact group ◮ Significantly generalizes Z / 2 case ◮ We give a precise analysis of the statistical and computational limits of this model ◮ Uses non-rigorous (but well-established) ideas from statistical physics ◮ Methods proven correct in related settings ◮ Includes an AMP algorithm which we believe is optimal among all polynomial-time algorithms ◮ Also some rigorous statistical lower and upper bounds Perry, W., Bandeira, Moitra, Message-passing algorithms for synchronization problems over compact groups , to appear in CPAM Perry, W., Bandeira, Moitra, Optimality and Sub-optimality of PCA for Spiked Random Matrices and Synchronization , part I to appear in Ann. Stat 15 / 39

  35. Multi-frequency U (1) synchronization ◮ G = U (1) = { z ∈ C : | z | = 1 } (angles) 16 / 39

  36. Multi-frequency U (1) synchronization ◮ G = U (1) = { z ∈ C : | z | = 1 } (angles) ◮ True signal x ∈ U (1) n 16 / 39

  37. Multi-frequency U (1) synchronization ◮ G = U (1) = { z ∈ C : | z | = 1 } (angles) ◮ True signal x ∈ U (1) n ◮ W – complex Gaussian noise (GUE) 16 / 39

  38. Multi-frequency U (1) synchronization ◮ G = U (1) = { z ∈ C : | z | = 1 } (angles) ◮ True signal x ∈ U (1) n ◮ W – complex Gaussian noise (GUE) ◮ Observe 16 / 39

  39. Multi-frequency U (1) synchronization ◮ G = U (1) = { z ∈ C : | z | = 1 } (angles) ◮ True signal x ∈ U (1) n ◮ W – complex Gaussian noise (GUE) ◮ Observe Y (1) = λ 1 n xx ∗ + 1 √ nW (1) 16 / 39

  40. Multi-frequency U (1) synchronization ◮ G = U (1) = { z ∈ C : | z | = 1 } (angles) ◮ True signal x ∈ U (1) n ◮ W – complex Gaussian noise (GUE) ◮ Observe Y (1) = λ 1 n xx ∗ + 1 √ nW (1) Y (2) = λ 2 n x 2 x ∗ 2 + 1 √ nW (2) · · · Y ( K ) = λ K n x K x ∗ K + 1 √ nW ( K ) where x k means entry-wise k th power. 16 / 39

  41. Multi-frequency U (1) synchronization ◮ G = U (1) = { z ∈ C : | z | = 1 } (angles) ◮ True signal x ∈ U (1) n ◮ W – complex Gaussian noise (GUE) ◮ Observe Y (1) = λ 1 n xx ∗ + 1 √ nW (1) Y (2) = λ 2 n x 2 x ∗ 2 + 1 √ nW (2) · · · Y ( K ) = λ K n x K x ∗ K + 1 √ nW ( K ) where x k means entry-wise k th power. ◮ This model has information on different frequencies 16 / 39

  42. Multi-frequency U (1) synchronization ◮ G = U (1) = { z ∈ C : | z | = 1 } (angles) ◮ True signal x ∈ U (1) n ◮ W – complex Gaussian noise (GUE) ◮ Observe Y (1) = λ 1 n xx ∗ + 1 √ nW (1) Y (2) = λ 2 n x 2 x ∗ 2 + 1 √ nW (2) · · · Y ( K ) = λ K n x K x ∗ K + 1 √ nW ( K ) where x k means entry-wise k th power. ◮ This model has information on different frequencies ◮ Challenge: how to synthesize information across frequencies? 16 / 39

  43. AMP for U (1) synchronization Y ( k ) = λ k n x k x ∗ k + 1 √ nW ( k ) for k = 1 , . . . , K 17 / 39

  44. AMP for U (1) synchronization Y ( k ) = λ k n x k x ∗ k + 1 √ nW ( k ) for k = 1 , . . . , K Algorithm’s state: v ( k ) ∈ C n for each frequency k ◮ v ( k ) is an estimate of ( x k 1 , . . . , x k n ) 17 / 39

  45. AMP for U (1) synchronization Y ( k ) = λ k n x k x ∗ k + 1 √ nW ( k ) for k = 1 , . . . , K Algorithm’s state: v ( k ) ∈ C n for each frequency k ◮ v ( k ) is an estimate of ( x k 1 , . . . , x k n ) AMP algorithm: ◮ Power iteration (separately on each frequency): v ( k ) ← Y ( k ) v ( k ) 17 / 39

  46. AMP for U (1) synchronization Y ( k ) = λ k n x k x ∗ k + 1 √ nW ( k ) for k = 1 , . . . , K Algorithm’s state: v ( k ) ∈ C n for each frequency k ◮ v ( k ) is an estimate of ( x k 1 , . . . , x k n ) AMP algorithm: ◮ Power iteration (separately on each frequency): v ( k ) ← Y ( k ) v ( k ) ◮ “Soft projection” (separately on each index i ): v ( · ) ← F ( v ( · ) i ) i ◮ This synthesizes the frequencies in a non-trivial way 17 / 39

  47. AMP for U (1) synchronization Y ( k ) = λ k n x k x ∗ k + 1 √ nW ( k ) for k = 1 , . . . , K Algorithm’s state: v ( k ) ∈ C n for each frequency k ◮ v ( k ) is an estimate of ( x k 1 , . . . , x k n ) AMP algorithm: ◮ Power iteration (separately on each frequency): v ( k ) ← Y ( k ) v ( k ) ◮ “Soft projection” (separately on each index i ): v ( · ) ← F ( v ( · ) i ) i ◮ This synthesizes the frequencies in a non-trivial way ◮ Onsager correction term 17 / 39

  48. AMP for U (1) synchronization Y ( k ) = λ k n x k x ∗ k + 1 √ nW ( k ) for k = 1 , . . . , K Algorithm’s state: v ( k ) ∈ C n for each frequency k ◮ v ( k ) is an estimate of ( x k 1 , . . . , x k n ) AMP algorithm: ◮ Power iteration (separately on each frequency): v ( k ) ← Y ( k ) v ( k ) ◮ “Soft projection” (separately on each index i ): v ( · ) ← F ( v ( · ) i ) i ◮ This synthesizes the frequencies in a non-trivial way ◮ Onsager correction term Analysis of AMP: ◮ Exact expression for AMP’s MSE (as n → ∞ ) as a function of λ 1 , . . . , λ K 17 / 39

  49. AMP for U (1) synchronization Y ( k ) = λ k n x k x ∗ k + 1 √ nW ( k ) for k = 1 , . . . , K Algorithm’s state: v ( k ) ∈ C n for each frequency k ◮ v ( k ) is an estimate of ( x k 1 , . . . , x k n ) AMP algorithm: ◮ Power iteration (separately on each frequency): v ( k ) ← Y ( k ) v ( k ) ◮ “Soft projection” (separately on each index i ): v ( · ) ← F ( v ( · ) i ) i ◮ This synthesizes the frequencies in a non-trivial way ◮ Onsager correction term Analysis of AMP: ◮ Exact expression for AMP’s MSE (as n → ∞ ) as a function of λ 1 , . . . , λ K ◮ Also, exact expression for the statistically optimal MSE 17 / 39

  50. Results for U (1) synchronization Y ( k ) = λ k n x k x ∗ k + 1 √ nW ( k ) for k = 1 , . . . , K 18 / 39

  51. Results for U (1) synchronization Y ( k ) = λ k n x k x ∗ k + 1 √ nW ( k ) for k = 1 , . . . , K ◮ Single frequency: given Y ( k ) , can non-trivially estimate x k iff λ k > 1 18 / 39

  52. Results for U (1) synchronization Y ( k ) = λ k n x k x ∗ k + 1 √ nW ( k ) for k = 1 , . . . , K ◮ Single frequency: given Y ( k ) , can non-trivially estimate x k iff λ k > 1 ◮ Information-theoretically, with λ 1 = · · · = λ K = λ , need � λ ∼ log K / K (for large K ) 18 / 39

  53. Results for U (1) synchronization Y ( k ) = λ k n x k x ∗ k + 1 √ nW ( k ) for k = 1 , . . . , K ◮ Single frequency: given Y ( k ) , can non-trivially estimate x k iff λ k > 1 ◮ Information-theoretically, with λ 1 = · · · = λ K = λ , need � λ ∼ log K / K (for large K ) ◮ But AMP (and conjecturally, any poly-time algorithm) requires λ k > 1 for some k 18 / 39

  54. Results for U (1) synchronization Y ( k ) = λ k n x k x ∗ k + 1 √ nW ( k ) for k = 1 , . . . , K ◮ Single frequency: given Y ( k ) , can non-trivially estimate x k iff λ k > 1 ◮ Information-theoretically, with λ 1 = · · · = λ K = λ , need � λ ∼ log K / K (for large K ) ◮ But AMP (and conjecturally, any poly-time algorithm) requires λ k > 1 for some k ◮ Computationally hard to synthesize sub-critical ( λ ≤ 1) frequencies 18 / 39

  55. Results for U (1) synchronization Y ( k ) = λ k n x k x ∗ k + 1 √ nW ( k ) for k = 1 , . . . , K ◮ Single frequency: given Y ( k ) , can non-trivially estimate x k iff λ k > 1 ◮ Information-theoretically, with λ 1 = · · · = λ K = λ , need � λ ∼ log K / K (for large K ) ◮ But AMP (and conjecturally, any poly-time algorithm) requires λ k > 1 for some k ◮ Computationally hard to synthesize sub-critical ( λ ≤ 1) frequencies ◮ But once above the λ = 1 threshold, adding frequencies helps reduce MSE of AMP 18 / 39

  56. Results for U (1) synchronization Solid: AMP ( n = 100) ( K = num freq) Dotted: theoretical ( n → ∞ ) Same λ on each frequency Image credit: Perry, W., Bandeira, Moitra, Message-passing algorithms for synchronization problems over compact groups , to appear in CPAM 19 / 39

  57. General groups All of the above extends to any compact group ◮ E.g. Any finite group; SO (3) 20 / 39

  58. General groups All of the above extends to any compact group ◮ E.g. Any finite group; SO (3) How to even define the model? ◮ Need to add “noise” to a group element g i g − 1 j 20 / 39

  59. General groups All of the above extends to any compact group ◮ E.g. Any finite group; SO (3) How to even define the model? ◮ Need to add “noise” to a group element g i g − 1 j Answer: Use representation theory to represent a group element as a matrix (and then add Gaussian noise) 20 / 39

  60. General groups All of the above extends to any compact group ◮ E.g. Any finite group; SO (3) How to even define the model? ◮ Need to add “noise” to a group element g i g − 1 j Answer: Use representation theory to represent a group element as a matrix (and then add Gaussian noise) ◮ A representation ρ of G is a way to assign a matrix ρ ( g ) to each g ∈ G ◮ Formally, a homomorphism ρ : G → GL ( C d ) = { d × d invertible matrices } 20 / 39

  61. General groups All of the above extends to any compact group ◮ E.g. Any finite group; SO (3) How to even define the model? ◮ Need to add “noise” to a group element g i g − 1 j Answer: Use representation theory to represent a group element as a matrix (and then add Gaussian noise) ◮ A representation ρ of G is a way to assign a matrix ρ ( g ) to each g ∈ G ◮ Formally, a homomorphism ρ : G → GL ( C d ) = { d × d invertible matrices } Frequencies are replaced by irreducible representations of G ◮ Fourier theory for functions G → C 20 / 39

  62. General groups All of the above extends to any compact group ◮ E.g. Any finite group; SO (3) How to even define the model? ◮ Need to add “noise” to a group element g i g − 1 j Answer: Use representation theory to represent a group element as a matrix (and then add Gaussian noise) ◮ A representation ρ of G is a way to assign a matrix ρ ( g ) to each g ∈ G ◮ Formally, a homomorphism ρ : G → GL ( C d ) = { d × d invertible matrices } Frequencies are replaced by irreducible representations of G ◮ Fourier theory for functions G → C For U (1), 1D irreducible representation for each k : ρ k ( g ) = g k 20 / 39

  63. Part II: Orbit Recovery 21 / 39

  64. Back to cryo-EM Image credit: [Singer, Shkolnisky ’11] 22 / 39

  65. Back to cryo-EM Image credit: [Singer, Shkolnisky ’11] Synchronization is not the ideal model for cryo-EM 22 / 39

  66. Back to cryo-EM Image credit: [Singer, Shkolnisky ’11] Synchronization is not the ideal model for cryo-EM ◮ The synchronization approach disregards the underlying signal (the molecule) 22 / 39

  67. Back to cryo-EM Image credit: [Singer, Shkolnisky ’11] Synchronization is not the ideal model for cryo-EM ◮ The synchronization approach disregards the underlying signal (the molecule) ◮ Our Gaussian synchronization model assumes independent noise on each pair i , j of images, whereas actually there is independent noise on each image 22 / 39

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend