possibilistic graphical models and how to learn them from
play

Possibilistic Graphical Models and How to Learn Them from Data - PowerPoint PPT Presentation

Possibilistic Graphical Models and How to Learn Them from Data Christian Borgelt Dept. of Knowledge Processing and Language Engineering Otto-von-Guericke-University of Magdeburg Universit atsplatz 2, D-39106 Magdeburg, Germany E-mail:


  1. Possibilistic Graphical Models and How to Learn Them from Data Christian Borgelt Dept. of Knowledge Processing and Language Engineering Otto-von-Guericke-University of Magdeburg Universit¨ atsplatz 2, D-39106 Magdeburg, Germany E-mail: borgelt@iws.cs.uni-magdeburg.de Christian Borgelt Possibilistic Graphical Models and How to Learn Them from Data 1

  2. Contents ✎ Possibility Theory ✍ Axiomatic Approach ✍ Semantical Considerations ✎ Graphical Models / Inference Networks ✍ relational ✍ probabilistic ✍ possibilistic ✎ Learning Possibilistic Graphical Models from Data ✍ Computing Maximum Projections ✍ Naive Possibilistic Classifiers ✍ Learning the Structure of Graphical Models ✎ Summary Christian Borgelt Possibilistic Graphical Models and How to Learn Them from Data 2

  3. Possibility Theory: Axiomatic Approach Definition: Let Ω be a (finite) sample space. A possibility measure Π on Ω is a function Π : 2 Ω ✦ [0 ❀ 1] satisfying 1. Π( ❀ ) = 0 and 2. ✽ ❊ 1 ❀ ❊ 2 ✒ Ω : Π( ❊ 1 ❬ ❊ 2 ) = max ❢ Π( ❊ 1 ) ❀ Π( ❊ 2 ) ❣ . ✎ Similar to Kolmogorov’s axioms of probability theory. ✎ From the axioms follows Π( ❊ 1 ❭ ❊ 2 ) ✔ min ❢ Π( ❊ 1 ) ❀ Π( ❊ 2 ) ❣ ✿ ✎ Attributes are introduced as random variables (as in probability theory). ✎ Π( ❆ = ❛ ) is an abbreviation of Π( ❢ ✦ ✷ Ω ❥ ❆ ( ✦ ) = ❛ ❣ ) ✎ If an event ❊ is possible without restriction, then Π( ❊ ) = 1 ✿ If an event ❊ is impossible, then Π( ❊ ) = 0 ✿ Christian Borgelt Possibilistic Graphical Models and How to Learn Them from Data 3

  4. Possibility Theory and the Context Model Interpretation of Degrees of Possibility [Gebhardt and Kruse 1993] ✎ Let Ω be the (nonempty) set of all possible states of the world, ✦ 0 the actual (but unknown) state. ✎ Let ❈ = ❢ ❝ 1 ❀ ✿ ✿ ✿ ❀ ❝ ♥ ❣ be a set of contexts (observers, frame conditions etc.) and ( ❈❀ 2 ❈ ❀ P ) a finite probability space (context weights). ✎ Let Γ : ❈ ✦ 2 Ω be a set-valued mapping, which assigns to each context the most specific correct set-valued specification of ✦ 0 . The sets Γ( ❝ ) are called the focal sets of Γ. ✎ Γ is a random set (i.e., a set-valued random variable) [Nguyen 1978]. The basic possibility assignment induced by Γ is the mapping ✙ : Ω ✦ [0 ❀ 1] ✙ ( ✦ ) ✼✦ P ( ❢ ❝ ✷ ❈ ❥ ✦ ✷ Γ( ❝ ) ❣ ) ✿ Christian Borgelt Possibilistic Graphical Models and How to Learn Them from Data 4

  5. � ☎ ✁ ✂ ✄ Example: Dice and Shakers shaker 1 shaker 2 shaker 3 shaker 4 shaker 5 tetrahedron hexahedron octahedron icosahedron dodecahedron 1 – 4 1 – 6 1 – 8 1 – 10 1 – 12 numbers degree of possibility 1 5 + 1 5 + 1 5 + 1 5 + 1 1 – 4 = 1 5 1 5 + 1 5 + 1 5 + 1 4 5 – 6 = 5 5 1 5 + 1 5 + 1 3 7 – 8 = 5 5 5 + 1 1 2 9 – 10 = 5 5 1 1 11 – 12 = 5 5 Christian Borgelt Possibilistic Graphical Models and How to Learn Them from Data 5

  6. From the Context Model to Possibility Measures Definition: Let Γ : ❈ ✦ 2 Ω be a random set. The possibility measure induced by Γ is the mapping Π : 2 Ω ✦ [0 ❀ 1] ❀ ✼✦ P ( ❢ ❝ ✷ ❈ ❥ ❊ ❭ Γ( ❝ ) ✻ = ❀❣ ) ✿ ❊ Problem: From the given interpretation it follows only: � � � ✽ ❊ ✒ Ω : max ✦ ✷ ❊ ✙ ( ✦ ) ✔ Π( ❊ ) ✔ min 1 ❀ ✙ ( ✦ ) ✿ ✦ ✷ ❊ 1 2 3 4 5 1 2 3 4 5 ❝ 1 : 1 ❝ 1 : 1 ✎ ✎ 2 2 ❝ 2 : 1 ❝ 2 : 1 ✎ ✎ ✎ ✎ ✎ 4 4 ❝ 3 : 1 ❝ 3 : 1 ✎ ✎ ✎ ✎ ✎ ✎ ✎ 4 4 1 1 1 1 1 1 1 1 ✙ 0 1 ✙ 2 2 4 4 4 2 4 4 Christian Borgelt Possibilistic Graphical Models and How to Learn Them from Data 6

  7. From the Context Model to Possibility Measures (cont.) Attempts to solve the indicated problem: ✎ Require the focal sets to be consonant : Definition: Let Γ : ❈ ✦ 2 Ω be a random set with ❈ = ❢ ❝ 1 ❀ ✿ ✿ ✿ ❀ ❝ ♥ ❣ . The focal sets Γ( ❝ ✐ ), 1 ✔ ✐ ✔ ♥ , are called consonant , iff there exists a sequence ❝ ✐ 1 ❀ ❝ ✐ 2 ❀ ✿ ✿ ✿ ❀ ❝ ✐ n , 1 ✔ ✐ 1 ❀ ✿ ✿ ✿ ❀ ✐ ♥ ✔ ♥ , ✽ 1 ✔ ❥ ❁ ❦ ✔ ♥ : ✐ ❥ ✻ = ✐ ❦ , so that Γ( ❝ ✐ 1 ) ✒ Γ( ❝ ✐ 2 ) ✒ ✿ ✿ ✿ ✒ Γ( ❝ ✐ n ) ✿ ✦ mass assignment theory [Baldwin et al. 1995] Problem: The “voting model” is not sufficient to justify consonance. ✎ Use the lower bound as the “most pessimistic” choice. [Gebhardt 1997] Problem: Basic possibility assignments represent negative information, the lower bound is actually the most optimistic choice. ✎ Justify the lower bound from decision making purposes. [Borgelt 1995, Borgelt 2000] Christian Borgelt Possibilistic Graphical Models and How to Learn Them from Data 7

  8. From the Context Model to Possibility Measures (cont.) ✎ Assume that in the end we have to decide on a single event. ✎ Each event is described by the values of a set of attributes. ✎ Then it can be useful to assign to a set of events the degree of possibility of the “most possible” event in the set. Example: � 36 18 18 28 0 40 0 40 28 0 0 0 28 28 40 0 0 40 18 18 0 0 0 18 0 0 20 20 18 18 0 0 0 18 36 0 18 18 0 18 max 40 40 20 max 18 18 18 28 Christian Borgelt Possibilistic Graphical Models and How to Learn Them from Data 8

  9. Possibility Distributions Definition: Let ❳ = ❢ ❆ 1 ❀ ✿ ✿ ✿ ❀ ❆ ♥ ❣ be a set of attributes defined on a (finite) sample space Ω with respective domains dom( ❆ ✐ ), ✐ = 1 ❀ ✿ ✿ ✿ ❀ ♥ . A possibility distribution ✙ ❳ over ❳ is the restriction of a possibility measure Π on Ω to the set of all events that can be defined by stating values for all attributes in ❳ . That is, ✙ ❳ = Π ❥ ❊ X , where � � � ❊ ✷ 2 Ω ❊ ❳ � ✾ ❛ 1 ✷ dom( ❆ 1 ) : ✿ ✿ ✿ ✾ ❛ ♥ ✷ dom( ❆ ♥ ) : = � � ❊ � = ❆ ❥ = ❛ ❥ ❆ j ✷ ❳ � � � ❊ ✷ 2 Ω � ✾ ❛ 1 ✷ dom( ❆ 1 ) : ✿ ✿ ✿ ✾ ❛ ♥ ✷ dom( ❆ ♥ ) : = � � �� � � ✦ ✷ Ω ❊ = ❆ ❥ ( ✦ ) = ❛ ❥ ✿ � ❆ j ✷ ❳ ✎ Corresponds to the notion of a probability distribution. ✎ Advantage of this formalization: No index transformation functions are needed for projections, there are just fewer terms in the conjunctions. Christian Borgelt Possibilistic Graphical Models and How to Learn Them from Data 9

  10. Conditional Possibility and Independence Definition: Let Ω be a (finite) sample space, Π a possibility measure on Ω, and ❊ 1 ❀ ❊ 2 ✒ Ω events. Then Π( ❊ 1 ❥ ❊ 2 ) = Π( ❊ 1 ❭ ❊ 2 ) is called the conditional possibility of ❊ 1 given ❊ 2 . Definition: Let Ω be a (finite) sample space, Π a possibility measure on Ω, and ❆❀ ❇❀ and ❈ attributes with respective domains dom( ❆ ) ❀ dom( ❇ ) ❀ and dom( ❈ ). ❆ and ❇ are called conditionally possibilistically independent given ❈ , written ❆ ❄ ❄ Π ❇ ❥ ❈ , iff ✽ ❛ ✷ dom( ❆ ) : ✽ ❜ ✷ dom( ❇ ) : ✽ ❝ ✷ dom( ❈ ) : Π( ❆ = ❛❀ ❈ = ❝ ❥ ❇ = ❜ ) = min ❢ Π( ❆ = ❛ ❥ ❇ = ❜ ) ❀ Π( ❈ = ❝ ❥ ❇ = ❜ ) ❣ ✿ ✎ Similar to the corresponding notions of probability theory. Christian Borgelt Possibilistic Graphical Models and How to Learn Them from Data 10

  11. Graphical Models / Inference Networks ✎ Decomposition: Under certain conditions a distribution ✍ (e.g. a probability distribution) on a multi-dimensional domain, which encodes prior or generic knowledge about this domain, can be decomposed into a set ❢ ✍ 1 ❀ ✿ ✿ ✿ ❀ ✍ s ❣ of (overlapping) distributions on lower-dimensional subspaces. ✎ Simplified Reasoning: If such a decomposition is possible, it is sufficient to know the distributions on the subspaces to draw all inferences in the domain under consideration that can be drawn using the original distribution ✍ . ✎ Since such a decomposition is usually represented as a network and since it is used to draw inferences, it can be called an inference network . The edges of the network indicate the paths along which evidence has to be propagated. ✎ Another popular name is graphical model , where “graphical” indicates that it is based on a graph in the sense of graph theory. Christian Borgelt Possibilistic Graphical Models and How to Learn Them from Data 11

  12. ✢ ✔ ✕ ✖ ✗ ✘ ✙ ✚ ✛ ✜ A Simple Example Example World Relation ☛ ✍ color shape size ☛ ✍ small ✡ ✍ medium ✡ ✍ small ✡ ☞ medium ✡ ☞ medium ✟ ✌ large ✠ ✌ medium ✠ ☞ medium ✠ ☞ medium ✎ 10 simple geometric objects, 3 attributes large ✎ One object is chosen at random and examined. ✎ Inferences are drawn about the unobserved attributes. Christian Borgelt Possibilistic Graphical Models and How to Learn Them from Data 12

  13. ★ ★ ★ ★ ✦ ★ ★ ✥ ✧ ✩ ✩ ✦ ✦ ★ The Reasoning Space Relation Geometric Interpretation ☛ ✡ ✟ ✠ ☛ ✍ color shape size ☞ ☛ ✍ small ✡ ✍ ✌ medium ✡ ✍ small ✍ ✡ ☞ medium ✡ ☞ medium ✟ ✌ large large ✠ ✌ medium medium small ✠ ☞ medium ✠ ☞ medium large Each cube represents one tuple. Christian Borgelt Possibilistic Graphical Models and How to Learn Them from Data 13

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend