logistics
play

Logistics Class webpage: - PDF document

School of Computer Science Introduction Probabilistic Graphical Models (10- Probabilistic Graphical Models (10 -708) 708) Lecture 0, Sep 10, 2007 Receptor A Receptor A X 1 X 1 X 1 Receptor B Receptor B X 2 X 2 X 2 Eric Xing Eric Xing


  1. School of Computer Science Introduction Probabilistic Graphical Models (10- Probabilistic Graphical Models (10 -708) 708) Lecture 0, Sep 10, 2007 Receptor A Receptor A X 1 X 1 X 1 Receptor B Receptor B X 2 X 2 X 2 Eric Xing Eric Xing Kinase C Kinase C X 3 X 3 X 3 Kinase D Kinase D X 4 X 4 X 4 Kinase E Kinase E X 5 X 5 X 5 TF F TF F X 6 X 6 X 6 Reading: Gene G Gene G X 7 X 7 X 7 X 8 X 8 X 8 Gene H Gene H 1 Logistics � Class webpage: http://www.cs.cmu.edu/~epxing/Class/10708-07/ � Eric Xing 2 1

  2. Logistics No formal text book, but draft chapters will be handed out in class: � M. I. Jordan, An Introduction to Probabilistic Graphical Models � Daphne Koller and Nir Friedman, Structured Probabilistic Models � Mailing Lists: � To contact the instructors: 10708-07-instr@cs.cmu.edu � Class announcements list: 10708-07-announce@cs.cmu.edu. � TA: � Hetunandan Kamichetty, Doherty 4302C, Office hours: Wednesdays, 5:00-6:00 pm � Dr. Ramesh Nallapati � Class Assistant: � � Monica Hopes, Wean Hall 4616, x8-5527 Eric Xing 3 Logistics � 4 homework assignments: 45% of grade Theory exercises � Implementation exercises � � Final project: 30% of grade Applying PGM to your research area � NLP, IR, Computational biology, vision, robotics … � Theoretical and/or algorithmic work � a more efficient approximate inference algorithm � a new sampling scheme for a non-trivial model … � � Take home final: 25% of grade Theory exercises and/or analysis � � Policies … Eric Xing 4 2

  3. Past projects: Winner of the 2005 project: � J. Yang, Y. Liu, E. P. Xing and A. Hauptmann, Harmonium-Based Models for Semantic Video Representation and Classification , Proceedings of The Seventh SIAM International Conference on Data Mining (SDM 2007) . ( Recipient of the BEST PAPER Award ) Other projects: � Andreas Krause, Jure Leskovec and Carlos Guestrin, Data Association for Topic Intensity Tracking, 23rd International Conference on Machine Learning (ICML 2006). Y. Shi, F. Guo, W. Wu and E. P. Xing, GIMscan: A New Statistical Method for We will have a prize for the � Analyzing Whole-Genome Array CGH best project(s) … Data, The Eleventh Annual International Conference on Research in Computational Molecular Biology (RECOMB 2007) . Eric Xing 5 What is this? Classical AI and ML research ignored this phenomena � The Problem (an example): � you want to catch a flight at 10:00am from Pitt to SF, can I make it if I leave at � 7am and take a 28X at CMU? � partial observability (road state, other drivers' plans, etc.) � noisy sensors (radio traffic reports) uncertainty in action outcomes (flat tire, etc.) � immense complexity of modeling and predicting traffic � Reasoning under uncertainty ! � Eric Xing 6 3

  4. A universal task … Information retrieval Information retrieval Speech recognition Computer vision Speech recognition Computer vision Games Games Robotic control Robotic control Pedigree Pedigree Evolution Evolution Planning Planning Eric Xing 7 The Fundamental Questions � Representation How to capture/model uncertainties in possible worlds? � How to encode our domain knowledge/assumptions/constraints? � � Inference How do I answers questions/queries � X 9 ? according to my model and/or based given data? X 8 D e.g. : ( | ) P X ? i ? � Learning ? X 7 X 6 What model is "right" � X 1 X 2 X 3 X 4 X 5 for my data? D M = M e.g. : arg max ( ; ) F M ∈ M Eric Xing 8 4

  5. X 9 X 8 Graphical Models X 7 X 6 X 1 X 2 X 3 X 4 X 5 Graphical models are a marriage between graph theory and � probability theory One of the most exciting developments in machine learning � (knowledge representation, AI, EE, Stats,…) in the last two decades… Some advantages of the graphical model point of view � Inference and learning are treated together � Supervised and unsupervised learning are merged seamlessly � Missing data handled nicely � A focus on conditional independence and computational issues � Interpretability (if desired) � Are having significant impact in science, engineering and beyond! � Eric Xing 9 What is a Graphical Model? � The informal blurb: It is a smart way to write/specify/compose/design exponentially-large � probability distributions without paying an exponential cost, and at the same time endow the distributions with structured semantics A A B B A A A B B B C C D D E E C C C D D D E E E F F F F F G G H H G G G H H H ( ) P X ,X ,X ,X ,X ,X ,X ,X ( ) = ( ) ( ) ( | ) ( | ) ( | ) P X P X P X P X X X P X X P X X 1 2 3 4 5 6 7 8 1 8 1 2 3 1 2 4 2 5 2 : ( , ) ( ) ( , ) P X X X P X X P X X X 6 3 4 7 6 8 5 6 � A more formal description: � It refers to a family of distributions on a set of random variables that are compatible with all the probabilistic independence propositions encoded by a graph that connects these variables Eric Xing 10 5

  6. Statistical Inference probabilistic probabilistic generative generative model model gene expression profiles gene expression profiles Eric Xing 11 Statistical Inference statistical statistical inference inference gene expression profiles gene expression profiles Eric Xing 12 6

  7. Multivariate Distribution in High-D Space � A possible world for cellular signal transduction: Receptor A Receptor B X 1 X 2 X 3 Kinase C Kinase D X 4 Kinase E X 5 TF F X 6 Gene G X 7 X 8 Gene H Eric Xing 13 Recap of Basic Prob. Concepts Representation: what is the joint probability dist. on multiple � variables? ( , , , , , , , , ) P X X X X X X X X 1 2 3 4 5 6 7 8 A A A A B B B B How many state configurations in total? --- 2 8 � C C C C D D D D E E E E Are they all needed to be represented? � F F F F � Do we get any scientific/medical insight? G G G G H H H H � Learning: where do we get all this probabilities? Maximal-likelihood estimation? but how many data do we need? � Where do we put domain knowledge in terms of plausible relationships between variables, and � plausible values of the probabilities? Inference: If not all variables are observable, how to compute the � conditional distribution of latent variables given evidence? Computing p ( H | A ) would require summing over all 2 6 configurations of the � unobserved variables Eric Xing 14 7

  8. What is a Graphical Model? --- example from a signal transduction pathway � A possible world for cellular signal transduction: Receptor A Receptor B X 1 X 2 X 3 Kinase C Kinase D X 4 Kinase E X 5 TF F X 6 Gene G X 7 X 8 Gene H Eric Xing 15 GM: Structure Simplifies Representation � Dependencies among variables Receptor A X 1 Receptor B X 2 Membrane X 3 Kinase C Kinase D X 4 Kinase E X 5 Cytosol TF F X 6 Gene G X 7 X 8 Gene H Nucleus Eric Xing 16 8

  9. Probabilistic Graphical Models � If X i 's are conditionally independent (as described by a PGM ), the joint can be factored to a product of simpler terms, e.g., Receptor A Receptor A Receptor B Receptor B X 1 X 1 X 1 X 2 X 2 X 2 P ( X 1 , X 2 , X 3 , X 4 , X 5 , X 6 , X 7 , X 8 ) = P ( X 1 ) P ( X 2 ) P ( X 3 | X 1 ) P ( X 4 | X 2 ) P ( X 5 | X 2 ) Kinase C Kinase C X 3 X 3 X 3 Kinase D Kinase D X 4 X 4 X 4 Kinase E Kinase E X 5 X 5 X 5 P ( X 6 | X 3 , X 4 ) P ( X 7 | X 6 ) P ( X 8 | X 5 , X 6 ) TF F TF F X 6 X 6 X 6 X 7 X 7 X 7 Gene G Gene G X 8 X 8 X 8 Gene H Gene H Stay tune for what are these independencies! � Why we may favor a PGM? � Incorporation of domain knowledge and causal (logical) structures 2+2+4+4+4+8+4+8=36, an 8-fold reduction from 2 8 in representation cost ! Eric Xing 17 GM: Data Integration Receptor A Receptor A X 1 X 1 Receptor B Receptor B X X X 2 X 2 X X 1 1 2 2 X 3 X 3 X X Kinase C Kinase C Kinase D Kinase D X 4 X 4 X X Kinase E Kinase E X 5 X 5 X X 3 3 4 4 5 5 TF F TF F X 6 X 6 X X 6 6 Gene G Gene G X 7 X X 7 X X 8 X 8 X X Gene H Gene H 7 7 8 8 Eric Xing 18 9

  10. Probabilistic Graphical Models � If X i 's are conditionally independent (as described by a PGM ), the joint can be factored to a product of simpler terms, e.g., Receptor A Receptor A Receptor B Receptor B X 1 X 1 X 1 X X X X 2 X 2 X 2 X X X 1 1 1 2 2 2 P ( X 1 , X 2 , X 3 , X 4 , X 5 , X 6 , X 7 , X 8 ) = P ( X 2 ) P ( X 4 | X 2 ) P ( X 5 | X 2 ) P ( X 1 ) P ( X 3 | X 1 ) Kinase C Kinase C X X 3 X 3 X X 3 X Kinase D Kinase D X 4 X X 4 X X 4 X Kinase E Kinase E X X 5 X X 5 X X 5 3 3 3 4 4 4 5 5 5 P ( X 6 | X 3 , X 4 ) P ( X 7 | X 6 ) P ( X 8 | X 5 , X 6 ) TF F TF F X 6 X X X X 6 X 6 6 6 6 Gene G Gene G X X 7 X X 7 X X 7 Gene H Gene H X X 8 X X X 8 X 8 7 7 7 8 8 8 � Why we may favor a PGM? � Incorporation of domain knowledge and causal (logical) structures 2+2+4+4+4+8+4+8=36, an 8-fold reduction from 2 8 in representation cost ! � Modular combination of heterogeneous parts – data fusion Eric Xing 19 Rational Statistical Inference The Bayes Theorem: Likelihood Prior Posterior probability probability ( | ) ( ) p d h p h = ( | ) h h p h d ∑ ′ ′ ( | ) ( ) p d h p h ′ ∈ h H Sum over space d d of hypotheses This allows us to capture uncertainty about the model in a principled way � But how can we specify and represent a complicated model? � Typically the number of genes need to be modeled are in the order of thousands! � Eric Xing 20 10

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend