Learning Terminological Na¨ ıve Bayesian Classifiers Under Different Assumptions
- n Missing Knowledge
Learning Terminological Na ve Bayesian Classifiers Under Different - - PowerPoint PPT Presentation
Learning Terminological Na ve Bayesian Classifiers Under Different Assumptions on Missing Knowledge Pasquale Minervini Claudia dAmato Nicola Fanizzi Department of Computer Science University of Bari URSW 2011 Bonn, October 23,
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works
1
2
3
4
5
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works
usually probabilistic information is assumed to be available the CWA is adopted
the OWA has to be taken into account
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works
Missing Completely at Random Missing at Random Informatively Missing
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works The Reference Representation Language Bayesian Networks
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works The Reference Representation Language Bayesian Networks
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works The Reference Representation Language Bayesian Networks
n
it is assumed that the presence (or absence) of a particular feature (random variable) of a class is unrelated to the presence (or absence) of any other feature, given the class variable (random variable)
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works Defining a Terminological Na¨ ıve Bayesian Network Learning a TBN: Problem Definition TBN: the Learning Algorithm The Ignorance Models
A Terminological Bayesian Network (TBN) NK, w.r.t. a DL KB K, is defined as a pair G, ΘG, where: G = V, E is a directed acyclic graph, in which: V = {F1, . . . , Fn, C} is a set of vertices, each Fi representing a DL (eventually complex) concepts defined over K and C representing a target concept E ⊆ V × V is a set of edges, modeling the dependence relations between the elements of V; ΘG is a set of conditional probability distributions (CPD), one for each V ∈ V, representing the conditional probability of the feature concept given its parents in the graph
In the case of a Terminological Na¨ ıve Bayesian Network, E = {C, Fi | i ∈ {1, . . . , n}}, namely ∀i, j ∈ {1, . . . , n} and i = j Fi is independent
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works Defining a Terminological Na¨ ıve Bayesian Network Learning a TBN: Problem Definition TBN: the Learning Algorithm The Ignorance Models
Pr(Female|Father) Pr(Female|¬Father) Pr(HasChild|Father) Pr(HasChild|¬Father)
Father Female HasChild
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works Defining a Terminological Na¨ ıve Bayesian Network Learning a TBN: Problem Definition TBN: the Learning Algorithm The Ignorance Models
C (A), Ind− C (A) and Ind0 C(A), so that:
∀a ∈ Ind+
C (A) : K |
= C(a), ∀a ∈ Ind−
C (A) : K |
= ¬C(a), ∀a ∈ Ind0
C(A) : K |
= C(a) ∧ K | = ¬C(a);
K maximizing the scoring function
K ← arg max NK
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works Defining a Terminological Na¨ ıve Bayesian Network Learning a TBN: Problem Definition TBN: the Learning Algorithm The Ignorance Models
function learn(K, IndC(A)) {The TBN is initialized as containing only the target concept node} N ∗
K = G, ΘG;
G = V ← {C}, E ← ∅; NK ← ∅; repeat NK ← N ∗
K;
{A new network is created, having one more node and different parameters than the previous one} Network = c′, N ′
K, s′ ← extend(NK, IndC(A));
N ∗
K ← N ′ K;
{Possible stopping conditions: a) improvements in score below a threshold; b) reaching a maximum number of nodes} until stopping criterion on Network; return NK;
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works Defining a Terminological Na¨ ıve Bayesian Network Learning a TBN: Problem Definition TBN: the Learning Algorithm The Ignorance Models
function extend(NK, IndC(A)) Concept ← Start; Best ← ∅; repeat Concepts ← ∅; for c′ ∈ {c′ ∈ ρcl
↓ (Concept) | |c′| ≤ min(|c| + d, maxLen)} do
V′ ← V ∪ {c′}; N ′
K ← optimalNetwork(V′, IndC(A));
s′ ← score(N ′
K, IndC(A));
Concepts ← Concepts ∪ {c′, N ′
K, s′};
end for Best ← arg maxc′,N ′
K,s′∈Concepts∪{Best} s′;
Concept ← c : c, NK, s = Best; {Possible stopping conditions: a) exceeding a maximum number of iterations; b) exceeding a maximum number of refinement steps} until Stopping criterion on Best; return Best;
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works Defining a Terminological Na¨ ıve Bayesian Network Learning a TBN: Problem Definition TBN: the Learning Algorithm The Ignorance Models
Initial Network Searching for the first feature Adding the first feature to the network Searching for the second feature Adding the second feature to the network Father ∃hasChild.Person Father ∃hasParent.Person ∃hasChild.Person Female Female ∃hasChild.⊤ Male Person ∃hasParent.⊤ ∃married.⊤ Female Mammal Person ⊤ ⊤ ∃hasSibling.⊤ ∃married.⊤ Father
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works Defining a Terminological Na¨ ıve Bayesian Network Learning a TBN: Problem Definition TBN: the Learning Algorithm The Ignorance Models
To learn the TBN, different assumptions (ignorance models) on the nature of the missing information are considered, given an ideal KB K∗ having additional knowledge: MCAR (Missing Completely At Random) – the probability that a ∈ C I is missing is independent of any kind of (additional) knowledge: Pr(K | = C(a) ∧ K | = ¬C(a) | K∗) = Pr(K | = C(a) ∧ K | = ¬C(a)); MAR (Missing At Random) – the probability that a ∈ C I is missing depends only from K and does not depend on additional knowledge: Pr(K | = C(a)∧K | = ¬C(a) | K∗) = Pr(K | = C(a)∧K | = ¬C(a) | K); NMAR (Not Missing At Random or IM, Informatively Missing) – the probability that a ∈ C I is missing could be not the same if additional knowledge is available Pr(K | = C(a)∧K | = ¬C(a) | K∗) = Pr(K | = C(a)∧K | = ¬C(a) | K).
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works Defining a Terminological Na¨ ıve Bayesian Network Learning a TBN: Problem Definition TBN: the Learning Algorithm The Ignorance Models
C (A)
C (A)
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works Defining a Terminological Na¨ ıve Bayesian Network Learning a TBN: Problem Definition TBN: the Learning Algorithm The Ignorance Models
C (A)
C (A)
C (A)
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works Defining a Terminological Na¨ ıve Bayesian Network Learning a TBN: Problem Definition TBN: the Learning Algorithm The Ignorance Models
probability intervals are determined instead of single probability values
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works
Pr(Father(a)) = Pr(Father) Pr(HasChild | Father)
Pr(Father′) Pr(HasChild | Father′) ;
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works
A na¨ ıve TBN using Robust Bayesian Estimation for inferring posterior probability intervals in presence of NMAR assumption is s.t. conditional probability contain probability intervals (defined by upper and lower bound) instead of probability values
[Pr(Female|Father),Pr(Female|Father)] [Pr(Female|¬Father),Pr(Female|¬Father)] [Pr(HasChild|Father),Pr(HasChild|Father)] [Pr(HasChild|¬Father),Pr(HasChild|¬Father)]
Father Female HasChild
Inference on a instance of Father given that K | = HasChild(a), is given by the interval [Pr(Father | HasChild), Pr(Father | HasChild)], where: Pr(Fa(a)) = Pr(Fa | HC) = Pr(HC | Fa)Pr(Fa) Pr(HC | Fa)Pr(Fa) + Pr(HC | ¬Fa)Pr(¬Fa) ; Pr(Fa(a)) = Pr(Fa | HC) = Pr(HC | Fa)Pr(Fa) Pr(HC | Fa)Pr(Fa) + Pr(HC | ¬Fa)Pr(¬Fa) ;
Learning Terminological Na¨ ıve Bayesian Classifiers
Introduction & Motivation Background Learning a Terminological Na¨ ıve Bayesian Network Classifying individuals with a TBN Conclusions and Future Works
Learning Terminological Na¨ ıve Bayesian Classifiers