em over binary decision diagrams for probabilistic logic
play

EM over Binary Decision Diagrams for Probabilistic Logic Programs - PowerPoint PPT Presentation

EM over Binary Decision Diagrams for Probabilistic Logic Programs Elena Bellodi Fabrizio Riguzzi ENDIF University of Ferrara, Italy elena.bellodi@unife.it, fabrizio.riguzzi@unife.it Bellodi, Riguzzi (University of Ferrara) EMBLEM 1 / 20


  1. EM over Binary Decision Diagrams for Probabilistic Logic Programs Elena Bellodi Fabrizio Riguzzi ENDIF – University of Ferrara, Italy elena.bellodi@unife.it, fabrizio.riguzzi@unife.it Bellodi, Riguzzi (University of Ferrara) EMBLEM 1 / 20

  2. Outline Probabilistic Logic Languages 1 Inference with Decision Diagrams 2 Weight Learning for LPADs 3 EM over BDDs 4 Experiments and results 5 Conclusions and future works 6 References 7 Bellodi, Riguzzi (University of Ferrara) EMBLEM 2 / 20

  3. Probabilistic Logic Languages Probabilistic Logic Programming Logic + Probability: useful to model domains with complex and uncertain relationships among entities Many approaches proposed in: Logic Programming, Uncertainty in AI, Machine Learning, Databases Logic Programming : Distribution Semantics [Sato, 1995] Independent Choice Logic,PRISM, ProbLog, Logic Programs with Annotated Disjunctions (LPADs)[Vennekens et al., 2004],... They define a probability distribution over normal logic programs (possible worlds) They differ in the definition of the probability distribution The distribution is extended to a joint distribution over worlds and queries The probability of a query is obtained from this distribution by marginalization Bellodi, Riguzzi (University of Ferrara) EMBLEM 3 / 20

  4. Probabilistic Logic Languages Logic Programs with Annotated Disjunctions (LPAD) Example: development of an epidemic or pandemic, if somebody has the flu and the climate is cold. C 1 = epidemic : 0 . 6 ; pandemic : 0 . 3 ; null:0.1 : − flu ( X ) , cold . C 2 = cold : 0 . 7 ; null:0.3 . C 3 = flu ( david ) . C 4 = flu ( robert ) . Worlds obtained by selecting only one atom from the head of every grounding of each rule Bellodi, Riguzzi (University of Ferrara) EMBLEM 4 / 20

  5. Inference with Decision Diagrams Inference Explanation: set of probabilistic choices that ensure the entailment of the goal Covering set of explanations: every world where the query is true is consistent with at least one explanation A covering set of explanations for :- epidemic. is { κ 1 , κ 2 } κ 1 = { ( C 1 , θ 1 = { X / david } , 1 ) , ( C 2 , {} , 1 ) } κ 2 = { ( C 1 , θ 2 = { X / robert } , 1 ) , ( C 2 , {} , 1 ) } Explanations are not mutually exclusive From a covering set of explanations the probability of the query Q is computed by means of Decision Diagrams Bellodi, Riguzzi (University of Ferrara) EMBLEM 5 / 20

  6. Inference with Decision Diagrams Multivalued Decision Diagrams (MDD) Multivalued Decision Diagrams (MDDs) represent a Boolean function f ( X ) on a set of multivalued variables X ij → ground clause C i θ j , with domain 1,..., | head ( C i ) | In a MDD a path to a 1-leaf corresponds to an explanation for Q The various paths are mutually exclusive f ( X ) = ( X 11 = 1 ∧ X 21 = 1 ) ∨ ( X 12 = 1 ∧ X 21 = 1 ) �� �� X 11 �� �� 3 2 1 �� �� �� �� X 12 �� �� �� �� 1 ������������� 3 1 �� �� X 21 �� �� 2 3 2 � � � 2 � � � 1 � � � � � � � 1 0 Bellodi, Riguzzi (University of Ferrara) EMBLEM 6 / 20

  7. � � � � � � Inference with Decision Diagrams Binary Decision Diagrams (BDD) MDDs can be converted into Binary Decision Diagrams with Boolean variables multivalued variable X ij with n i values → n i − 1 Boolean variables X ij 1 ,..., X ijn i − 1 from f ( X ) = ( X 11 = 1 ∧ X 21 = 1 ) ∨ ( X 12 = 1 ∧ X 21 = 1 ) to f ( X ) = (( X 111 ∧ X 112 ) ∧ X 211 ) ∨ (( X 121 ∧ X 122 ) ∧ X 211 ) �� �� n 1 � � X 111 �� �� � �� �� n 2 X 121 �� �� �� �� n 3 X 211 �� �� � � � � � � 1 0 Bellodi, Riguzzi (University of Ferrara) EMBLEM 7 / 20

  8. Weight Learning for LPADs Weight Learning for LPADs Problem: model of the domain known VS weights (numeric parameters) unknown Weight learning: inference of weights from data Given a LPAD: a probabilistic logical model with unknown probabilities data: a set of interpretations Find the values of the probabilities that maximize the probability of the data given the model Expectation Maximization (EM) algorithm iterative method for problems with incomplete data Expectation step: estimates missing data given observed data + current estimate of parameters Maximization step: computes the parameters using estimates of E step Bellodi, Riguzzi (University of Ferrara) EMBLEM 8 / 20

  9. EM over BDDs EMBLEM : EM over Bdds for probabilistic Logic programs Efficient Mining EM over BDDs proposed in [Ishihata et al., 2008] Input: a LPAD; logical interpretations (data); target predicate(s) all ground atoms in the interpretations for the target predicate(s) correspond to as many queries BDDs encode the disjunction of explanations for each query Q EM algorithm directly over the BDDs missing data : the number of times that i-th head atom has been selected from groundings of the clauses used in the proof of the queries Bellodi, Riguzzi (University of Ferrara) EMBLEM 9 / 20

  10. EM over BDDs EM Algorithm Expectation step (synthesis) Computes P ( X ijk = x , Q ) and P ( Q ) 1 � j ∈ g ( i ) P ( X ijk = x , Q ) expected counts E [ c ikx ] = 2 P ( Q ) for all rules C i and k = 1 , ..., n i − 1, where c ikx is the number of times a binary variable X ijk takes value x ∈ { 0 , 1 } , and for all values of j ∈ g ( i ) = { j | θ j is a substitution grounding C i } Maximization step Updates parameters π ik representing P ( X ijk = 1 ) π ik = E [ c ik 1 | Q ] / ( E [ c ik 0 | Q ] + E [ c ik 1 | Q ] ) Bellodi, Riguzzi (University of Ferrara) EMBLEM 10 / 20

  11. EM over BDDs Expectation Computation n ∈ N ( Q ) , v ( n )= X ijk e x ( n ) P ( X ijk = x , Q ) = � n ∈ N ( Q ) , v ( n )= X ijk F ( n ) B ( child x ( n )) π ikx = � π ikx is π ik if x = 1 and ( 1 − π ik ) if x = 0 F ( n ) is the forward probability, the probability mass of the paths from the root to n B ( n ) is the backward probability, the probability mass of paths from n to the 1-leaf e x ( n ) is the probability mass of paths from the root to the 1 leaf passing through the x branch of n Bellodi, Riguzzi (University of Ferrara) EMBLEM 11 / 20

  12. EM over BDDs Computation of the forward probability 1: procedure G ET F ORWARD ( root ) 2: F ( root ) = 1 F ( n ) = 0 for all nodes ⊲ BDD traversed from root to leaves 3: for l = 1 to levels do ⊲ BDD levels 4: for all node ∈ Nodes ( l ) do ⊲ Nodes of one level 5: Let X ijk be v ( node ) , the variable associated to node 6: if child 0 ( node ) is not terminal then ⊲ node ’s child connected by 0-branch 7: F ( child 0 ( node )) = F ( child 0 ( node )) + F ( node ) · ( 1 − π ik ) ⊲ π ik : probability 8: Add child 0 ( node ) to Nodes ( level ( child 0 ( node ))) 9: end if 10: if child 1 ( node ) is not terminal then 11: F ( child 1 ( node )) = F ( child 1 ( node )) + F ( node ) · π ik 12: Add child 1 ( node ) to Nodes ( level ( child 1 ( node ))) 13: end if 14: end for 15: end for 16: end procedure For all nodes of a level the forward probabilities of their children are computed, by using probabilities π ik associated to the outgoing edges Bellodi, Riguzzi (University of Ferrara) EMBLEM 12 / 20

  13. EM over BDDs Computation of the backward probability function G ET B ACKWARD ( node ) if node is a terminal then ⊲ BDD traversed recursively from root up to leaves return value ( node ) ⊲ leaves return 0 or 1 else Let X ijk be v ( node ) B ( child 0 ( node )) = G ET B ACKWARD ( child 0 ( node ) ) ⊲ recursive calls B ( child 1 ( node )) = G ET B ACKWARD ( child 1 ( node ) ) e 0 ( node ) = F ( node ) · B ( child 0 ( node )) · ( 1 − π ik ) ⊲ F ( node ) from GetForward e 1 ( node ) = F ( node ) · B ( child 1 ( node )) · π ik η 0 ( i , k ) = η 0 t ( i , k ) + e 0 ( node ) ⊲ update of η x ( i , k ) to build P ( X ijk = x , Q ) η 1 ( i , k ) = η 1 t ( i , k ) + e 1 ( node ) return B ( child 0 ( node )) · ( 1 − π ik ) + B ( child 1 ( node )) · π ik end if end function at the end of all recursive calls, the function returns B(root) = probability of the query P(Q) Bellodi, Riguzzi (University of Ferrara) EMBLEM 13 / 20

  14. Experiments and results Experiments - settings EMBLEM is implemented in Yap Prolog Comparison with other systems for learning and inference under the distribution semantics: RIB [Riguzzi and di Mauro, 2011] CEM [Riguzzi, 2007] LeProblog [De Raedt et al., 2007] for learning and inference in Markov Logic Networks: Alchemy Datasets composed of 5 mega-interpretations → Five-fold cross validation Performance evaluation Area Under the PR (Precision-Recall) Curve Bellodi, Riguzzi (University of Ferrara) EMBLEM 14 / 20

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend