Knowledge Provenance: An Approach to Modeling and Maintaining - - PowerPoint PPT Presentation

knowledge provenance
SMART_READER_LITE
LIVE PREVIEW

Knowledge Provenance: An Approach to Modeling and Maintaining - - PowerPoint PPT Presentation

ICEIMT04, Toronto Knowledge Provenance: An Approach to Modeling and Maintaining Evolution and Validity of Knowledge Jingwei Huang & Mark S. Fox Enterprise Integration Laboratory, University of Toronto {msf,jingwei}@eil.utoronto.ca


slide-1
SLIDE 1

1

Knowledge Provenance:

An Approach to Modeling and Maintaining Evolution and Validity of Knowledge

Jingwei Huang & Mark S. Fox Enterprise Integration Laboratory, University of Toronto {msf,jingwei}@eil.utoronto.ca http://www.eil.utoronto.ca/ ICEIMT’04, Toronto

slide-2
SLIDE 2

2

Outline

  • Problem
  • Motivating example
  • Static KP Ontology
  • Implementation
  • Progress
  • Dynamic KP
  • Uncertainty-oriented KP
  • Conclusion and future work
slide-3
SLIDE 3

3

Problem to be addressed

  • How to determine the validity and origin
  • f Web information
  • Why?
  • In general
  • Anyone can publish info on Internet
  • Web info may be true, false, or outdated
  • Need tools to discern difference
  • In the context of enterprise integration
  • Dynamic partnerships
  • Uncertain information flow
  • Knowledge management
slide-4
SLIDE 4

4

Questions KP attempts to answer

  • Can this information be believed to be

true?

  • Who created it?
  • Can its creator be trusted?
  • What does it depend on?
  • Can the information it depends on be

believed to be true?

slide-5
SLIDE 5

5

Research Objectives

Goal: to create an approach to determining the validity and origin of knowledge/information on the Web

  • Level 1: Static Knowledge Provenance
  • Level 2: Dynamic KP
  • Level 3: Uncertainty-oriented KP
  • Level 4: Judgment-based KP
slide-6
SLIDE 6

6

Relevant Research

  • Internet information resource evaluation criteria

([Alexander&Tate99])

  • - Authority, Accuracy, Objectivity, Currency and Coverage
  • Trust and Trust Management

[Marsh94] [Blaze96] [Yu&Singh00][Golbeck02]

  • - formalizing trust
  • - decentralized trust management for web security
  • - trust propagation in social networks
  • Social Networks ([Milgram67],[Watts99] )
  • - six degrees of separation
  • Semantic Web ([Berners-Lee03])
  • Technologies from AI/uncertainty logic/Digital Signature
slide-7
SLIDE 7

7

Motivating Scenario

“Arctic sea ice coverage fell from around 6.5 million

square kilometres to around 5.5 million square kilometres in 2002.”

  • --- a news from internet

Is it true?

  • Who said this?
  • -- NASA
  • How can we know the creator is as claimed?
  • -- digital signature and digital certification
  • Can we trust the creator on this topic?
  • -- NASA has authority on the topic
  • -- trust relationships
slide-8
SLIDE 8

8

Motivating Scenario (2):

  • ------ Dependent Information

“In 2002, a satellite-based survey [NASA 2003] found that ‘Arctic sea ice coverage fell from around 6.5 million square kilometres to around 5.5 million square kilometres in one year’. The melting sea ice threatens to drive polar bears extinct within 100 years”.

  • Quotation
  • Derived conclusion
slide-9
SLIDE 9

9

Findings in Scenario

  • Basic information unit ---- proposition
  • Proposition taxonomy

KP_prop Asserted_prop Dependent_prop Equivalent_prop Derived_prop Composite_prop AND_prop OR_prop NEG_prop

  • How to determine truth?
  • Information dependencies
  • Trust relationships
slide-10
SLIDE 10

10

Approach

Modeling and maintaining

  • information sources
  • Information dependencies
  • trust relationships
slide-11
SLIDE 11

11

Static KP Ontology(1)

  • Motivating scenario
  • Informal competency questions/requirements
  • Terminology
  • Axioms in FOL
slide-12
SLIDE 12

12

Informal competency questions

  • Is this proposition true, false, or unknown?
  • Who created this proposition?
  • What is the digital signature verification status?
  • Which knowledge fields does it belong to?
  • Can the infor-creator be trusted in the fields?
  • Is its truth dependent on other propositions?

If so, which ones?

slide-13
SLIDE 13

13

Terminology

  • assigned_truth_value (x,v):

x has a truth value v assigned by info-creator

  • trusted_truth_value(a,x,v):

KP agent a trusts: x has truth value v.

  • is_dependent_on(x, y):

Proposition x is dependent on proposition y.

  • trusted(x, a):

x is trusted by a

  • trusted_in(a, c, f):

a trusts information creator c in knowledge field f.

slide-14
SLIDE 14

14

Axioms

  • 14 axioms defined in FOL
  • A proposition is "trusted", if:
  • its creator is trusted in the topic
  • digital signature is verified.
  • An asserted-prop is trusted to be true, if:
  • It is trusted
  • An equivalent-prop is trusted to be true, if:
  • has same content as its source
  • its source is trusted to be true.
  • A derived-prop is trusted to be true, if:
  • it is trusted,
  • the propositions it depends on are trusted to be true.
slide-15
SLIDE 15

15

Axiom (2)

Axiom SKP-12: for-all (a, x, y, v)((type(x, "derived_prop") ^ trusted(x, a) ^ assigned_truth_value(x, v) ^ is_dependent_on(x, y) ^ trusted_truth_value(a, y, "True"))

  • >trusted_truth_value(a, x, v)).
slide-16
SLIDE 16

16

Implementation

  • Information creators:

annotate web info with KP tags (RDFS data)

  • Information users (provenance requesters):

define trust relationships

  • KP reasoner:

traces kp tags across web pages and deduce validity of info

  • -- has implemented in Prolog
  • -- web implementation
slide-17
SLIDE 17

17

Example: annotate web page

<kp:Derived_prop rdf:id="ReduceDelay" is_dependent_on = "#ProblemOfDelay" in_field = "CRM”> To reduce response-delay to less than one minute may increase customer loyalty. </kp:Derived_prop>

slide-18
SLIDE 18

18

Example: KP Process

Derived_prop:"ReduceDelay" creator:"Tim Levy" is_dependent_on:"ProblemOfDelay" in_field:"CustomerRelationManagement" assigned_truth_value:"True" Derived_prop:"ReduceDelay" creator:"Tim Levy" is_dependent_on:"ProblemOfDelay" in_field:"CustomerRelationManagement" assigned_truth_value:"True" Equivalent_prop:"ProblemOfDelay" creator:"Tim Levy" is_dependent_on:"doc2;ProblemOfDelay" in_field:"CustomerRelationManagement" Equivalent_prop:"ProblemOfDelay" creator:"Tim Levy" is_dependent_on:"doc2;ProblemOfDelay" in_field:"CustomerRelationManagement" Asserted_prop:"ProblemOfDelay" creator:"Bill Cox" in_field:"CustomerRelationManagement" assigned_truth_value:"True" Asserted_prop:"ProblemOfDelay" creator:"Bill Cox" in_field:"CustomerRelationManagement" assigned_truth_value:"True" Is it true ? Who said ? applying axiom SKP-2 ==> This proposition is trusted applying axiom SKP-1 ==> Trusted_truth_value=True applying axiom SKP-5 ==> Trusted_truth_value=True applying axiom SKP-2&12 ==> Trusted_truth_value=True Trust Relationships: "Tim Levy" is trusted in "CRM" ... "Bill Cox" is trusted in "CRM" ... Trust Relationships: "Tim Levy" is trusted in "CRM" ... "Bill Cox" is trusted in "CRM" ... has_source KP_metadata depends_on KP_metadata KP_metadata Yeah, you could believe it ! because ...

KP software agent

Digital Signature Verification

slide-19
SLIDE 19

19

Dynamic KP

  • Truth values and trust relationships

may change over time

  • Dynamic KP extends static KP by

attaching an effective period to each proposition and each trust relationship

  • Dynamic KP determine the validity of

a proposition at a give time point

slide-20
SLIDE 20

20

Uncertainty-oriented KP(1)

  • Uncertainty in KP
  • uncertain truth values
  • uncertain trust relationships
  • Introduce subjective probability to

represent uncertain truth values and uncertain trust relationships

slide-21
SLIDE 21

21

Uncertainty-oriented KP(2)

  • Trust Degree:

td = Pr(trusted_in(a,c,f))

  • -- uncertain trust relationship
  • -- subject probability for an info user to trust an info-creator
  • Assigned Certainty Degree:

acd = Pr(assigned_truth_value(x,”True”))

  • -- uncertain assigned truth value given by creator
  • -- subjective probability of a proposition being true
  • Certainty Degree:

cd = Pr(trusted_truth_value(a,x,”True”))

  • -- uncertain trusted truth value
  • -- probability of the trusted truth value being true
slide-22
SLIDE 22

22

Uncertainty-oriented KP(3)

Certainty degree of asserted proposition:

cd = td*(acd – 0.5) + 0.5

td = 1.0 1 1 0.5 0.5 acd cd (a) when td=1.0, cd=acd td = 0.0 1 1 0.5 0.5 acd cd (b) when td=0.0, cd=0.5

td1 1 1 0.5 0.5 acd cd td2 < td1 (c) with the decrease of td, cd close to 0.5

slide-23
SLIDE 23

23

Uncertainty-oriented KP(4)

  • Certainty degree of a derived proposition

cdx = tdx*cdy*(acdx – 0.5) + 0.5

  • Certainty degree of an equivalent proposition

cdx = tdx * (cdy – 0.5) + 0.5

  • Certainty degree of an AND proposition

Z= x ^ y cdz = tdz*(Pr(x|y)* cdy – 0.5) + 0.5

slide-24
SLIDE 24

24

Conclusion

  • Formal models of (static, dynamic,

uncertain) Knowledge Provenance has been built and implemented in Prolog.

  • The KP models could be used to help

users to determine the validity and

  • rigin of web information annotated

with KP metadata.

slide-25
SLIDE 25

25

Future Work

  • Level 4: Judgment-based KP:

Trust propagation and update in social networks

  • Web implementation
  • Applications
  • E-business
  • Health consulting
slide-26
SLIDE 26

26

Thank You ! & Questions?