learning from description logics
play

Learning from Description Logics Part 2 of the Tutorial on Semantic - PowerPoint PPT Presentation

Learning from Description Logics Part 2 of the Tutorial on Semantic Data Mining Agnieszka Lawrynowicz, Jedrzej Potoniec Poznan University of Technology Semantic Data Mining Tutorial (ECML/PKDD11) 1 Athens, 9 September 2011 Outline


  1. Learning from Description Logics Part 2 of the Tutorial on Semantic Data Mining Agnieszka Lawrynowicz, Jedrzej Potoniec Poznan University of Technology Semantic Data Mining Tutorial (ECML/PKDD’11) 1 Athens, 9 September 2011

  2. Outline Description logics in a nutshell 1 Learning in description logic - definition 2 DL learning methods and techniques: 3 Concept learning Refinement operators Pattern mining Similarity-based approaches Tools 4 Applications 5 Presentation of a tool: RMonto 6 Semantic Data Mining Tutorial (ECML/PKDD’11) 2 Athens, 9 September 2011

  3. Learning in DLs Definition Learning in description logics: a machine learning approach that adopts Inductive Logic Programming as the methodology and description logic as the language of data and hypotheses. Description logics theoretically underpin the state-of-art Web ontology representation language, OWL , so description logic learning approaches are well suited for semantic data mining. Semantic Data Mining Tutorial (ECML/PKDD’11) 3 Athens, 9 September 2011

  4. Description logic Definition Description Logics, DL s = family of first order logic-based formalisms suitable for representing knowledge, especially terminologies, ontologies. Semantic Data Mining Tutorial (ECML/PKDD’11) 4 Athens, 9 September 2011

  5. Description logic Definition Description Logics, DL s = family of first order logic-based formalisms suitable for representing knowledge, especially terminologies, ontologies. subset of first order logic (decidability, efficiency, expressivity) root: semantic networks, frames Semantic Data Mining Tutorial (ECML/PKDD’11) 4 Athens, 9 September 2011

  6. Basic building blocks DL concepts roles constructors individuals Examples Atomic concepts : Artist , Movie Role: creates Constructors: ⊓ ⊓ ⊓ , ∃ ∃ ∃ Concept definition: Director ≡ ≡ ≡ Artist ⊓ ⊓ ⊓ ∃ ∃ ∃ creates.Movie ⊑ Axiom (”each director is an artist”) : Director ⊑ ⊑ Artist Asertion: creates(sofiaCoppola, lostInTranslation) Semantic Data Mining Tutorial (ECML/PKDD’11) 5 Athens, 9 September 2011

  7. DL knowledge base K = ( T Box, A Box ) T Box = { CreteHolidaysOffer ≡ Offer ⊓∃ in.Crete ⊓∀ in.Crete SantoriniHolidaysOffer ≡ Offer ⊓∃ in.Santorini ⊓∀ in.Santorini TromsøyaHolidaysOffer ≡ Offer ⊓∃ in.Tromsøya ⊓∀ in.Tromsøya Crete ⊑ ∃ partOf.Greece Santorini ⊑ ∃ partOf.Greece Tromsøya ⊑ ∃ partOf.Norway }. A Box = { Offer(o1). in(Crete). SantoriniHolidaysOffer(o2). Offer(o3). in(Santorini). hasPrice(o3, 300) }. Semantic Data Mining Tutorial (ECML/PKDD’11) 6 Athens, 9 September 2011

  8. DL reasoning services satisfiability inconsistency subsumption instance checking Semantic Data Mining Tutorial (ECML/PKDD’11) 7 Athens, 9 September 2011

  9. Concept learning Given new target concept name C knowledge base K as background knowledge a set E + of positive examples, and a set E − of negative examples the goal is to learn a concept definition C ≡ D such that = E + and K ∪ { C ≡ D } | K ∪ { C ≡ D } | = E − Semantic Data Mining Tutorial (ECML/PKDD’11) 8 Athens, 9 September 2011

  10. Negative examples and Open World Assumption But what are negative examples in the context of the Open World Assumption? Semantic Data Mining Tutorial (ECML/PKDD’11) 9 Athens, 9 September 2011

  11. Semantics: ”closed world” vs ”open world” Closed world (Logic programming LP , databases) complete knowledge of instances lack of information is by default negative information ( negation-as-failure ) Semantic Data Mining Tutorial (ECML/PKDD’11) 10 Athens, 9 September 2011

  12. Semantics: ”closed world” vs ”open world” Closed world (Logic programming LP , databases) complete knowledge of instances lack of information is by default negative information ( negation-as-failure ) Open world (description logic DL , Semantic Web) incomplete knowledge of instances negation of some fact has to be explicitely asserted ( monotonic negation ) Semantic Data Mining Tutorial (ECML/PKDD’11) 10 Athens, 9 September 2011

  13. ”Closed world” vs ”open world” example Let data base contain the following data : OscarMovie(lostInTranslation) Director(sofiaCoppola) creates(sofiaCoppola, lostInTranslation) Semantic Data Mining Tutorial (ECML/PKDD’11) 11 Athens, 9 September 2011

  14. ”Closed world” vs ”open world” example Let data base contain the following data : OscarMovie(lostInTranslation) Director(sofiaCoppola) creates(sofiaCoppola, lostInTranslation) Are all of the movies of Sofia Coppola Oscar movies? Semantic Data Mining Tutorial (ECML/PKDD’11) 11 Athens, 9 September 2011

  15. ”Closed world” vs ”open world” example Let data base contain the following data : OscarMovie(lostInTranslation) Director(sofiaCoppola) creates(sofiaCoppola, lostInTranslation) Are all of the movies of Sofia Coppola Oscar movies? YES - closed world Semantic Data Mining Tutorial (ECML/PKDD’11) 11 Athens, 9 September 2011

  16. ”Closed world” vs ”open world” example Let data base contain the following data : OscarMovie(lostInTranslation) Director(sofiaCoppola) creates(sofiaCoppola, lostInTranslation) Are all of the movies of Sofia Coppola Oscar movies? YES - closed world DON’T KNOW - open world Semantic Data Mining Tutorial (ECML/PKDD’11) 12 Athens, 9 September 2011

  17. ”Closed world” vs ”open world” example Let data base contain the following data : OscarMovie(lostInTranslation) Director(sofiaCoppola) creates(sofiaCoppola, lostInTranslation) Are all of the movies of Sofia Coppola Oscar movies? YES - closed world DON’T KNOW - open world Different conclusions! Semantic Data Mining Tutorial (ECML/PKDD’11) 12 Athens, 9 September 2011

  18. OWA and machine learning OWA is problematic for machine learning since an individual is rarely deduced to belong to a complement of a concept unless explicitely asserted so. Semantic Data Mining Tutorial (ECML/PKDD’11) 13 Athens, 9 September 2011

  19. Dealing with OWA in learning Solution1: alternative problem setting Solution2: K operator Solution3: new performance measures Semantic Data Mining Tutorial (ECML/PKDD’11) 14 Athens, 9 September 2011

  20. Dealing with OWA in learning: alternative problem setting ”Closing” the knowledge base to allow performing instance checks under the Closed World Assumption (CWA). By default: Positive examples of the form C ( a ) , and negative examples of the form ¬ C ( a ) , where a is an individual and holding: = E + and K ∪ { C ≡ D } | K ∪ { C ≡ D } | = E − Alternatively: = E + and Examples of the form C ( a ) and holding: K ∪ { C ≡ D } | K ∪ { C ≡ D } �| = E − Semantic Data Mining Tutorial (ECML/PKDD’11) 15 Athens, 9 September 2011

  21. Dealing with OWA in learning: K operator epistemic K –operator allows for querying for known properties of known individuals w.r.t. the given knowlege base K the K operator alters constructs like ∀ in a way that they operate on a Closed World Assumption. Consider two queries: Q1: K | = {( ∀ creates.OscarMovie) (sofiaCoppola)} Q2: K | = {( ∀ K creates.OscarMovie) (sofiaCoppola)} Badea and Nienhuys-Cheng (ILP 2000) considered the K operator from a theoretical point of view. not easy to implement in reasoning systems, non-standard Semantic Data Mining Tutorial (ECML/PKDD’11) 16 Athens, 9 September 2011

  22. Dealing with OWA in learning: new performance measures d’Amato et al (ESWC 2008) – overcoming unknown answers from the reasoner (as a reference system) – correspondence between the classification by the reasoner for the instances w.r.t. the test concept C and the definition induced by a learning system match rate: number of individuals with exactly the same classification by both the inductive and the deductive classifier w.r.t the overall number of individuals; omission error rate: number of individuals not classified by inductive method, relevant to the query w.r.t. the reasoner; commission error rate: number of individuals found relevant to C , while they (logically) belong to its negation or vice-versa; induction rate: number of individuals found relevant to C or to its negation, while either case not logically derivable from K ; Semantic Data Mining Tutorial (ECML/PKDD’11) 17 Athens, 9 September 2011

  23. Concept learning - algorithms supervised: YINYANG (Iannone et al, Applied Intelligence 2007) DL-Learner (Lehmann & Hitzler, ILP 2007) DL-FOIL (Fanizzi et al, ILP 2008) TERMITIS (Fanizzi et al, ECML/PKDD 2010) unsupervised: KLUSTER (Kietz & Morik, MLJ 1994) Semantic Data Mining Tutorial (ECML/PKDD’11) 18 Athens, 9 September 2011

  24. DL-learning as search learning in DLs can be seen as search in space of concepts it is possible to impose ordering on this search space using subsumption as natural quasi-order , and generality measure between concepts if D ⊑ C then C covers all instances that are covered by D refinement operators may be applied to traverse the space by computing a set of specializations (resp. generalizations) of a concept Semantic Data Mining Tutorial (ECML/PKDD’11) 19 Athens, 9 September 2011

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend