anchored correlation explanation topic modeling with
play

Anchored Correlation Explanation: Topic Modeling with Minimal - PowerPoint PPT Presentation

Anchored Correlation Explanation: Topic Modeling with Minimal Domain Knowledge Ryan J. Gallagher @ryanjgallag github.com/gregversteeg/corex_topic Anchored Correlation Explanation: How to Topic Model with Literally Thousands of Information


  1. Anchored Correlation Explanation: Topic Modeling with Minimal Domain Knowledge Ryan J. Gallagher @ryanjgallag github.com/gregversteeg/corex_topic

  2. Anchored Correlation Explanation: How to Topic Model with Literally Thousands of Information Bottlenecks 🍿 Topic Modeling with Minimal Domain Knowledge Ryan J. Gallagher @ryanjgallag github.com/gregversteeg/corex_topic

  3. LDA is a generative topic model @ryanjgallag NAACL 2018, New Orleans, LA 3

  4. LDA is a generative topic model The Good: Priors explicitly encode your beliefs about what topics can be, and easily allow for iterative development of new topic models @ryanjgallag NAACL 2018, New Orleans, LA 4

  5. Domain Knowledge via Dirichlet Forest Priors “Incorporating Domain Knowledge into Topic Modeling via Dirichlet Forest Priors.” Andrzejewski et al. ICML (2009) @ryanjgallag NAACL 2018, New Orleans, LA 5

  6. Domain Knowledge via First-Order Logic “A Framework for Incorporating General Domain Knowledge into Latent Dirichlet Allocation Using First-Order Logic.” Andrzejewski et al. IJCAI (2011). @ryanjgallag NAACL 2018, New Orleans, LA 6

  7. SeededLDA “Incorporating Lexical Priors into Topic Models.” Jagarlamudi et al. EACL (2012) @ryanjgallag NAACL 2018, New Orleans, LA 7

  8. Hierarchical LDA “Hierarchical Topic Models and the Nested Chinese Restaurant Process.” Gri ffi ths et al. Neural Information Processing Systems (2003). @ryanjgallag NAACL 2018, New Orleans, LA 8

  9. A Generative Modeling Tradeoff The Good: Priors explicitly encode your beliefs about what topics can be, and easily allow for iterative development of new topic models The Bad: Each additional prior takes a very specific view of the problem at hand, which both limits what a topic can be and makes it harder to justify in applications and to domain experts @ryanjgallag NAACL 2018, New Orleans, LA 9

  10. Proposed Work We propose a topic model that We propose a topic model that learns topics through learns topics through information-theoretic criteria, information-theoretic criteria, rather than a generative model, rather than a generative model, … within a framework that yields within a framework that yields hierarchical and semi- hierarchical and semi- supervised extensions with no supervised extensions with no additional assumptions additional assumptions … @ryanjgallag NAACL 2018, New Orleans, LA 10

  11. Proposed Work We propose a topic model that learns topics through information-theoretic criteria, rather than a generative model, … within a framework that yields hierarchical and semi- supervised extensions with no additional assumptions … @ryanjgallag NAACL 2018, New Orleans, LA 11

  12. A Different Perspective on “Topics” Consider three documents: LDA: a topic is a distribution over words @ryanjgallag NAACL 2018, New Orleans, LA 12

  13. A Different Perspective on “Topics” Consider three documents: CorEx: a topic is a binary latent factor LDA: a topic is a distribution over words @ryanjgallag NAACL 2018, New Orleans, LA 13

  14. A Different Perspective on “Topics” Consider three documents: CorEx: a topic is a binary latent factor LDA: a topic is a distribution over words @ryanjgallag NAACL 2018, New Orleans, LA 14

  15. CorEx Objective (example) Documents Probability table 1/2 0 0 1/2 @ryanjgallag NAACL 2018, New Orleans, LA 15

  16. CorEx Objective (example) Documents Probability table 1/2 0 0 1/2 Words 1 and 2 are related: @ryanjgallag NAACL 2018, New Orleans, LA 16

  17. CorEx Objective (example) Documents Probability table 1/2 0 0 1/2 Words 1 and 2 are related: Hypothesize a latent factor: @ryanjgallag NAACL 2018, New Orleans, LA 17

  18. CorEx Objective (example) Documents Probability table 1/2 0 0 1/2 Words 1 and 2 are related: Hypothesize a latent factor: Then conditioned on , words 1 and 2 are independent @ryanjgallag NAACL 2018, New Orleans, LA 18

  19. CorEx Objective (example) Documents Probability table 1/2 0 0 1/2 Words 1 and 2 are related: Hypothesize a latent factor: Then conditioned on , words 1 and 2 are independent Goal: find latent factors that make words conditionally independent @ryanjgallag NAACL 2018, New Orleans, LA 19

  20. CorEx Objective Goal: find latent factors that make words conditionally independent @ryanjgallag NAACL 2018, New Orleans, LA 20

  21. CorEx Objective Goal: find latent factors that make words conditionally independent Total correlation conditioned on Y @ryanjgallag NAACL 2018, New Orleans, LA 21

  22. CorEx Objective Goal: find latent factors that make words conditionally independent if and only if the topic “explains” all the dependencies (total correlation) Hence, “Total Cor relation Ex planation” (CorEx) @ryanjgallag NAACL 2018, New Orleans, LA 22

  23. CorEx Objective Goal: find latent factors that make words conditionally independent In order to maximize the information between a group of words in topic we consider a tractable lower bound: @ryanjgallag NAACL 2018, New Orleans, LA 23

  24. CorEx Objective Goal: find latent factors that make words conditionally independent In order to maximize the information between a group of words in topic we consider a tractable lower bound: We maximize this lower bound over topics @ryanjgallag NAACL 2018, New Orleans, LA 24

  25. CorEx Objective We can now rewrite the objective: @ryanjgallag NAACL 2018, New Orleans, LA 25

  26. CorEx Objective We can now rewrite the objective: We transform this from a combinatorial to a continuous optimization by introducing variables and “relaxing” words into informative topics @ryanjgallag NAACL 2018, New Orleans, LA 26

  27. CorEx Objective We can now rewrite the objective: We transform this from a combinatorial to a continuous optimization by introducing variables and “relaxing” words into informative topics This relaxation yields a set of update equations which we can iterate through until convergence @ryanjgallag NAACL 2018, New Orleans, LA 27

  28. 
 CorEx Objective We can now rewrite the objective: We transform this from a combinatorial to a continuous optimization by introducing variables and “relaxing” words into informative topics Under the hood: 1. We introduce a sparsity optimization for the update equations, 
 by assuming words are represented by binary random variables 
 2. The current relaxation scheme places each word in one topic, resulting in a partition of the vocabulary, rather than mixed membership topics @ryanjgallag NAACL 2018, New Orleans, LA 28

  29. 
 CorEx Objective We can now rewrite the objective: We transform this from a combinatorial to a continuous optimization by introducing variables and “relaxing” words into informative topics Under the hood: 1. We introduce a sparsity optimization for the update equations, 
 by assuming words are represented by binary random variables 
 2. The current relaxation scheme places each word in one topic, resulting in a partition of the vocabulary, rather than mixed membership topics @ryanjgallag NAACL 2018, New Orleans, LA 29

  30. 
 CorEx Objective We can now rewrite the objective: We transform this from a combinatorial to a continuous optimization by introducing variables and “relaxing” words into informative topics Under the hood: 1. We introduce a sparsity optimization for the update equations, 
 by assuming words are represented by binary random variables 
 2. The current relaxation scheme places each word in one topic, resulting in a partition of the vocabulary, rather than mixed membership topics @ryanjgallag NAACL 2018, New Orleans, LA 30

  31. 
 CorEx Objective We can now rewrite the objective: We transform this from a combinatorial to a continuous optimization by introducing variables and “relaxing” words into informative topics Under the hood: 1. We introduce a sparsity optimization for the update equations, 
 by assuming words are represented by binary random variables 
 2. The current relaxation scheme places each word in one topic, resulting in a partition of the vocabulary, rather than mixed membership topics These are issues of speed, not theory @ryanjgallag NAACL 2018, New Orleans, LA 31

  32. CorEx Topic Examples Data: news articles about Hillary Clinton’s presidential campaign, up to August 2016 Work by Abigail Ross and the Computational Story Lab, University of Vermont @ryanjgallag NAACL 2018, New Orleans, LA 32

  33. CorEx Topic Examples Data: news articles about Hillary Clinton’s presidential campaign, up to August 2016 Clinton Article Topics 1: server, department, classified, information, private, investigation, fbi, email, emails, secretary 3: sanders, bernie, primary, vermont, win, voters, race, nomination, vote, polls 6: crowd, woman, speech, night, women, stage, man, mother, audience, life 8: percent, poll, points, percentage, margin, survey, according, 10, polling, university 9: federal, its, o ffi cials, law, including, committee, sta ff , statement, director, group 13: islamic, foreign, military, terrorism, war, syria, iraq, isis, u, terrorist 14: trump, donald, trump’s, republican, nominee, party, convention, top, election, him Work by Abigail Ross and the Computational Story Lab, University of Vermont @ryanjgallag NAACL 2018, New Orleans, LA 33

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend