probabilistic graphical models
play

Probabilistic Graphical Models David Sontag New York University - PowerPoint PPT Presentation

Probabilistic Graphical Models David Sontag New York University Lecture 4, February 21, 2013 David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 1 / 29 Conditional random fields (CRFs) A CRF is a Markov network on variables X


  1. Probabilistic Graphical Models David Sontag New York University Lecture 4, February 21, 2013 David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 1 / 29

  2. Conditional random fields (CRFs) A CRF is a Markov network on variables X ∪ Y , which specifies the conditional distribution 1 � P ( y | x ) = φ c ( x , y c ) Z ( x ) c ∈ C with partition function � � Z ( x ) = φ c ( x , ˆ y c ) . y ˆ c ∈ C As before, two variables in the graph are connected with an undirected edge if they appear together in the scope of some factor The only difference with a standard Markov network is the normalization term – before marginalized over X and Y , now only over Y David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 2 / 29

  3. Parameterization of CRFs We typically parameterize each factor as a log-linear function, φ c ( x , y c ) = exp { w c · f c ( x , y c ) } f c ( x , y c ) is a feature vector w c are weights that are typically learned – we will discuss this extensively in later lectures This is without loss of generality: any discrete CRF can be parameterized like this (why?) Conditional random fields are in the exponential family: �� � 1 � P ( y | x ) = φ c ( x , y c ) = exp w c · f c ( x , y c ) − ln Z ( w , x ) Z ( x ) c ∈ C c ∈ C = exp { w · f ( x , y ) − ln Z ( w , x ) } . David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 3 / 29

  4. NLP example: named-entity recognition Given a sentence, determine the people and organizations involved and the relevant locations: “Mrs. Green spoke today in New York. Green chairs the finance committee.” Entities sometimes span multiple words. Entity of a word not obvious without considering its context CRF has one variable X i for each word, which encodes the possible labels of that word The labels are, for example, “B-person, I-person, B-location, I-location, B-organization, I-organization” Having beginning (B) and within (I) allows the model to segment adjacent entities David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 4 / 29

  5. NLP example: named-entity recognition The graphical model looks like (called a skip-chain CRF ): There are three types of potentials: φ 1 ( Y t , Y t +1 ) represents dependencies between neighboring target variables [analogous to transition distribution in a HMM] φ 2 ( Y t , Y t ′ ) for all pairs t , t ′ such that x t = x t ′ , because if a word appears twice, it is likely to be the same entity φ 3 ( Y t , X 1 , · · · , X T ) for dependencies between an entity and the word sequence [e.g., may have features taking into consideration capitalization] Notice that the graph structure changes depending on the sentence! David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 5 / 29

  6. Today’s lecture 1 Worst-case complexity of probabilistic inference 2 Elimination algorithm 3 Running-time analysis of elimination algorithm ( treewidth ) David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 6 / 29

  7. Probabilistic inference Today we consider exact inference in graphical models In particular, we focus on conditional probability queries, p ( Y | E = e ) = p ( Y , e ) p ( e ) (e.g., the probability of a patient having a disease given some observed symptoms) Let W = X − Y − E be the random variables that are neither the query nor the evidence. Each of these joint distributions can be computed by marginalizing over the other variables: � � p ( Y , e ) = p ( Y , e , w ) , p ( e ) = p ( y , e ) w y Naively marginalizing over all unobserved variables requires an exponential number of computations Does there exist a more efficient algorithm? David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 7 / 29

  8. Computational complexity of probabilistic inference Here we show that, unless P=NP, there does not exist a more efficient algorithm We show this by reducing 3-SAT, which is NP-hard, to probabilistic inference in Bayesian networks 3-SAT asks about the satisfiability of a logical formula defined on n literals Q 1 , . . . , Q n , e.g. ( ¬ Q 3 ∨ ¬ Q 2 ∨ Q 3 ) ∧ ( Q 2 ∨ ¬ Q 4 ∨ ¬ Q 5 ) · · · Each of the disjunction terms is called a clause , e.g. C 1 ( q 1 , q 2 , q 3 ) = ¬ q 3 ∨ ¬ q 2 ∨ q 3 In 3-SAT, each clause is defined on at most 3 literals. Our reduction also proves that inference in Markov networks is NP-hard (why?) David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 8 / 29

  9. Reducing satisfiability to MAP inference Input: 3-SAT formula with n literals Q 1 , . . . Q n and m clauses C 1 , . . . , C m Q 1 Q 2 Q 3 Q 4 Q n C 1 C 2 C 3 C m– 1 C m . . . A 1 A 2 A m– 2 X . . . One variable Q i ∈ { 0 , 1 } for each literal, p ( Q i = 1) = 0 . 5. One variable C i ∈ { 0 , 1 } for each clause, whose parents are the literals used in the clause. C i = 1 if the clause is satisfied, and 0 otherwise: p ( C i = 1 | q pa ( i ) ) = 1[ C i ( q pa ( i ) )] Variable X which is 1 if all clauses satisfied, and 0 otherwise: p ( A i = 1 | pa ( A i )) = 1[ pa ( A i ) = 1 ] , for i = 1 , . . . , m − 2 p ( X = 1 | a m − 2 , c m ) = 1[ a m − 2 = 1 , c m = 1] David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 9 / 29

  10. Reducing satisfiability to MAP inference Input: 3-SAT formula with n literals Q 1 , . . . Q n and m clauses C 1 , . . . , C m Q 1 Q 2 Q 3 Q 4 Q n C 1 C 2 C 3 C m– 1 C m . . . A 1 A 2 A m– 2 X . . . p ( q , c , a , X = 1) = 0 for any assignment q which does not satisfy all clauses 1 p ( Q = q , C = 1 , A = 1 , X = 1) = 2 n for any satisfying assignment q Thus, we can find a satisfying assignment (whenever one exists) by constructing this BN and finding the maximum a posteriori (MAP) assignment: argmax q , c , a p ( Q = q , C = c , A = a | X = 1) This proves that MAP inference in Bayesian networks and MRFs is NP-hard David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 10 / 29

  11. Reducing satisfiability to marginal inference Input: 3-SAT formula with n literals Q 1 , . . . Q n and m clauses C 1 , . . . , C m Q 1 Q 2 Q 3 Q 4 Q n C 1 C 2 C 3 C m– 1 C m . . . A 1 A 2 A m– 2 X . . . p ( X = 1) = � q , c , a p ( Q = q , C = c , A = a , X = 1) is equal to the number 1 of satisfying assignments times 2 n Thus, p ( X = 1) > 0 if and only if the formula has a satisfying assignment This shows that marginal inference is also NP-hard David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 11 / 29

  12. Reducing satisfiability to approximate marginal inference Might there exist polynomial-time algorithms that can approximately answer marginal queries, i.e. for some ǫ , find ρ such that ρ − ǫ ≤ p ( Y | E = e ) ≤ ρ + ǫ ? Suppose such an algorithm exists, for any ǫ ∈ (0 , 1 2 ). Consider the following: Start with E = { X = 1 } 1 For i = 1 , . . . , n : 2 Let q i = arg max q p ( Q i = q | E ) 3 E ← E ∪ ( Q i = q i ) 4 At termination, E is a satisfying assignment (if one exists). Pf by induction: In iteration i , if ∃ satisfying assignment extending E for both q i = 0 and q i = 1, then choice in line 3 does not matter Otherwise, suppose ∃ satisfying assignment extending E for q i = 1 but not for q i = 0. Then, p ( Q i = 1 | E ) = 1 and p ( Q i = 0 | E ) = 0 Even if approximate inference returned p ( Q i = 1 | E ) = 0 . 501 and p ( Q i = 0 | E ) = . 499, we would still choose q i = 1 Thus, it is even NP-hard to approximately perform marginal inference! David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 12 / 29

  13. Probabilistic inference in practice NP-hardness simply says that there exist difficult inference problems Real-world inference problems are not necessarily as hard as these worst-case instances The reduction from SAT created a very complex Bayesian network: Q 1 Q 2 Q 3 Q 4 Q n C 1 C 2 C 3 C m– 1 C m . . . A 1 A 2 A m– 2 X . . . Some graphs are easy to do inference in! For example, inference in hidden Markov models Y6 Y3 Y4 Y5 Y1 Y2 X1 X2 X3 X4 X5 X6 and other tree-structured graphs can be performed in linear time David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 13 / 29

  14. Variable elimination (VE) Exact algorithm for probabilistic inference in any graphical model Running time will depend on the graph structure Uses dynamic programming to circumvent enumerating all assignments First we introduce the concept for computing marginal probabilities, p ( X i ), in Bayesian networks After this, we will generalize to MRFs and conditional queries David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 14 / 29

  15. Basic idea Suppose we have a simple chain, A → B → C → D , and we want to compute p ( D ) p ( D ) is a set of values, { p ( D = d ) , d ∈ Val ( D ) } . Algorithm computes sets of values at a time – an entire distribution By the chain rule and conditional independence, the joint distribution factors as p ( A , B , C , D ) = p ( A ) p ( B | A ) p ( C | B ) p ( D | C ) In order to compute p ( D ), we have to marginalize over A , B , C : � p ( D ) = p ( A = a , B = b , C = c , D ) a , b , c David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 15 / 29

  16. Let’s be a bit more explicit... There is structure to the summation, e.g., repeated P ( c 1 | b 1 ) P ( d 1 | c 1 ) Let’s modify the computation to first compute P ( a 1 ) P ( b 1 | a 1 ) + P ( a 2 ) P ( b 1 | a 2 ) David Sontag (NYU) Graphical Models Lecture 4, February 21, 2013 16 / 29

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend