Ontology Consumer Analysis Tool Onto CAT Valerie Cross and Anindita - - PowerPoint PPT Presentation

ontology consumer analysis tool
SMART_READER_LITE
LIVE PREVIEW

Ontology Consumer Analysis Tool Onto CAT Valerie Cross and Anindita - - PowerPoint PPT Presentation

Ontology Consumer Analysis Tool Onto CAT Valerie Cross and Anindita Pal Computer Science and Systems Analysis Miami University, Oxford OH 2006 Protg Conference Stanford University 2006 Protege Conference 1 Agenda Motivation


slide-1
SLIDE 1

2006 Protege Conference 1

Ontology Consumer Analysis Tool OntoCAT

Valerie Cross and Anindita Pal Computer Science and Systems Analysis Miami University, Oxford OH 2006 Protégé Conference Stanford University

slide-2
SLIDE 2

2006 Protege Conference 2

Agenda

  • Motivation
  • Perspectives on Ontology Evaluation
  • Some Current Approaches
  • Ontology Consumer Analysis Tool
  • Some Experiments Using OntoCAT
  • Conclusion
  • Future Plans
slide-3
SLIDE 3

2006 Protege Conference 3

CAT on a Log

Evaluating

OWL on a Log

Note that OWL and CAT are not only on two separate logs But also in two separate worlds!

slide-4
SLIDE 4

2006 Protege Conference 4

Motivation

  • Ontologies the “backbone of the

Semantic Web”

  • Development and deployment
  • ntology-based software solutions

requires considerable time and effort

  • Numerous existing ontologies in

libraries available on the WWW

  • Why reinvent the wheel? Reuse of
  • ntologies
slide-5
SLIDE 5

2006 Protege Conference 5

What is ontology evaluation?

  • Ontology evaluation - key problem in

the field of ontology development and reuse.

  • Selection vs. Evaluation
  • Two separate tasks?
  • How related?
  • When does it occur?
  • Selection Evaluation?
  • Ontology Selection: Ontology

Evaluation on the Real Semantic Web

(Sabou, Lopez, Motta,Uren EON 2006)

slide-6
SLIDE 6

2006 Protege Conference 6

What kinds of selection criteria?

  • Popularity
  • metrics account solely for the links between

different ontologies.

  • same principle as Web search engines, often use

a modified version of the PageRank algorithm.

  • Semantic data richness
  • determine richness of the ontology’s

conceptualization

  • Topic coverage
  • level to which ontology covers a certain topic.
  • ontology concept labels compared to a set of

query terms representing the domain.

slide-7
SLIDE 7

2006 Protege Conference 7

What are we evaluating?

  • From U.S. National Center for

Ontology Research (NCOR) position paper at EON 2006:

  • well-defined ontology design

techniques, i.e., quality of design

  • principled measurement methods, i.e.,

quality of evaluation

  • higher quality ontologies, i.e., quality of

content

slide-8
SLIDE 8

2006 Protege Conference 8

Some Approaches to E valuating Ontologies

  • One-T [Bouillon et al 2002] :
  • Ontology Group at Universidad Politécnica de Madrid

(UPM)

  • Content for completeness, consistency and correctness
  • OntoClean [Guarino and Welty 2002] :
  • The Ontology Group at the Italian National Research

Council (CNR).

  • Methodologies to evaluate during its entire lifetime
  • Formal analysis of taxonomies
slide-9
SLIDE 9

2006 Protege Conference 9

Some Approaches to E valuating Ontologies

ONTOMETRIC [Lozano-Tello and Gómez-Pérez 2004]

  • Ontology Group at Universidad Politécnica de Madrid

(UPM)

  • method to quantify the suitability of ontologies for the

users’ systems,

  • uses a taxonomy of 160 ontology characteristics,
  • Content, language, development methodology, built by software tool,

cost of use.

  • not fully automated, based on AHP (Saaty 1977)
  • Application Use of ontology to assess application’s

performance, merits of

  • competency questions,
  • use cases,
  • scenarios
slide-10
SLIDE 10

2006 Protege Conference 10

Consumer Perspective Approach

  • Noy [2004] suggests for ontology re-use need more

research from consumer perspective

  • Somewhat analogous to reviewing Table of Contents and

Index, number of pages, etc. for the usefulness of book before deciding whether to check out or purchase.

  • AKTiveRank [Alani and Brewster 2005]
  • AKT (Advanced Knowledge Technologies) consortium of

British universities: Southampton, Edinburgh, Aberdeen, Sheffield and The Open University.

  • ranks ontologies retrieved by an ontology search engine

based on set of query terms and measures

  • OntoQA Analysis tool [Tartir 2005]
  • LSDIS (Large Scale Distributed Information Systems) Lab,

University of Georgia

  • analyzes ontology schemas and their populations and

describes them through a set of metrics.

slide-11
SLIDE 11

2006 Protege Conference 11

AKTiveRank

  • Ranks ontologies retrieved by search engine

(EON 2005)

  • Class match: coverage of query terms
  • Centrality: more central a class
  • Density: degree of details
  • Semantic similarity measure: closeness of classes
  • Produces overall rank
  • Extensions (EON 2006 and Protégé Conference)
  • Collect vocabulary for domain interest
  • Ranking based on number of class labels that match

with terminology for domain of interest

  • New Centrality based on high “betweenness”
slide-12
SLIDE 12

2006 Protege Conference 12

OntoQA

  • Schema:
  • Relationship richness
  • Attribute richness
  • Inheritance richnness
  • Instances:
  • Class Richness
  • Average Population
  • Connectivity
  • Cohesion
  • Importance
  • Relationship Richness
  • Fullness
slide-13
SLIDE 13

2006 Protege Conference 13

Ontology Consumer Analysis Tool

  • plug-in for OWL Protégé
  • very parameterized
  • Intensional and extensional
  • View metrics interested in
  • Size
  • Structure
  • User selectable root for analysis
  • User selectable relation for

establishing extensional structure

slide-14
SLIDE 14

2006 Protege Conference 14

WordNet

  • Princeton University
  • Terminological ontology of English
  • Organizes nouns, verbs, adjectives and adverbs

into synonym sets

  • Simple intensional structure: 10 classes
slide-15
SLIDE 15

2006 Protege Conference 15

WordNet

  • Complex extensional structure based on

hypernymOf /hyponymOf

  • Example Root Instance “entity, physical thing”, one
  • f the nine noun roots
slide-16
SLIDE 16

2006 Protege Conference 16

OntoCAT User Interface

slide-17
SLIDE 17

2006 Protege Conference 17

OntoCAT Buttons

  • Metrics Button
  • Display result of selected metrics
  • Report Button
  • Report result of selected to file
  • Button
  • Generate tree of hub concept to visualize
  • Click hub for individual hub visualization
slide-18
SLIDE 18

2006 Protege Conference 18

OntoCat Selection Class/ E xtensional Relation

slide-19
SLIDE 19

2006 Protege Conference 19

OntoCAT Hub Analysis

slide-20
SLIDE 20

2006 Protege Conference 20

OntoCAT Intensional Report

slide-21
SLIDE 21

2006 Protege Conference 21

E xtensional Hub Summary for WordNet

slide-22
SLIDE 22

2006 Protege Conference 22

OntoCAT Root Summary

slide-23
SLIDE 23

2006 Protege Conference 23

UMLS Hub Summaries

slide-24
SLIDE 24

2006 Protege Conference 24

Visualizing Hubs

Figure 6.6 ICD9CM Information Visualization of Hubs with Connecting Concepts.

slide-25
SLIDE 25

2006 Protege Conference 25

Summary

  • Many flavors of ontology evaluation
  • r selection
  • OntoCat - one of several tools to

begin addressing needs of ontology evaluation for the purpose of re-use

  • Structural and size analysis just one

set of parameters.

  • Challenge specifying parameters or

structural properties for evaluation

  • user preference
  • purpose for reusing ontology
slide-26
SLIDE 26

2006 Protege Conference 26

Possible Future Work

  • Interface with filtering/selection

approaches such as AKTiveRank before perform evaluation

  • Comparison metrics/charts for multiple
  • ntologies in addition to ranking
  • Current Visualization
  • Hubs visualization Improvement
  • Individual hub visualization
  • Top-level summary
  • Bottom-up level summary