Monte Carlo Semantics McPIET at RTE-4: Robust Inference and Logical - - PowerPoint PPT Presentation

monte carlo semantics
SMART_READER_LITE
LIVE PREVIEW

Monte Carlo Semantics McPIET at RTE-4: Robust Inference and Logical - - PowerPoint PPT Presentation

Monte Carlo Semantics McPIET at RTE-4: Robust Inference and Logical Pattern Processing Based on Integrated Deep and Shallow Semantics Richard Bergmair University of Cambridge Computer Laboratory Natural Language Information Processing Text


slide-1
SLIDE 1

Monte Carlo Semantics

McPIET at RTE-4: Robust Inference and Logical Pattern Processing Based on Integrated Deep and Shallow Semantics Richard Bergmair

University of Cambridge Computer Laboratory Natural Language Information Processing

Text Analysis Conference, Nov-17 2008

slide-2
SLIDE 2

Desiderata for a Theory of RTE

◮ Does it describe the relevant aspects of the

systems we have now?

◮ Does it suggest ways of building better

systems in the future?

slide-3
SLIDE 3

A System for RTE

◮ informativity: Can it take into account all available

relevant information?

◮ robustness: Can it proceed on reasonable assumptions,

where it is missing relevant information.

slide-4
SLIDE 4

Current RTE Systems

A spectrum between

◮ shallow inference

(e.g. bag-of-words)

◮ deep inference

(e.g. FOPC theorem proving, see Bos & Markert)

slide-5
SLIDE 5

The Informativity/Robustness Tradeoff

informativity robustness

slide-6
SLIDE 6

The Informativity/Robustness Tradeoff

informativity

deep

robustness

slide-7
SLIDE 7

The Informativity/Robustness Tradeoff

informativity

deep shallow

robustness

slide-8
SLIDE 8

The Informativity/Robustness Tradeoff

informativity

deep shallow intermediate

robustness

slide-9
SLIDE 9

The Informativity/Robustness Tradeoff

informativity robustness

?

slide-10
SLIDE 10

Outline

Informativity, Robustness & Graded Validity Propositional Model Theory & Graded Validity Shallow Inference: Bag-of-Words Encoding Deep Inference: Syllogistic Encoding Computation via the Monte Carlo Method

slide-11
SLIDE 11

Outline

Informativity, Robustness & Graded Validity Propositional Model Theory & Graded Validity Shallow Inference: Bag-of-Words Encoding Deep Inference: Syllogistic Encoding Computation via the Monte Carlo Method

slide-12
SLIDE 12

Informative Inference.

predicate/argument structures ⊤ > The cat chased the dog. → The dog chased the cat. monotonicity properties, upwards entailing Some (grey X) are Y → Some X are Y ≥ ⊤ ⊤ > Some X are Y → Some (grey X) are Y

slide-13
SLIDE 13

Robust Inference.

monotonicity properties, upwards entailing Some X are Y → Some (grey X) are Y > Some X are Y → Some (clean (grey X)) are Y graded standards of proof Socrates is a man → Socrates is a man > Socrates is a man → Socrates is mortal Socrates is a man → Socrates is mortal > Socrates is a man → Socrates is not a man

slide-14
SLIDE 14

. . . classically

(i) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; ENTAILED / valid (ii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; CONTRADICTION / unsatisfiable (iii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; UNKNOWN / possible (iv) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ. UNKNOWN / possible

slide-15
SLIDE 15

. . . classically

(i) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; ENTAILED / valid (ii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; CONTRADICTION / unsatisfiable (iii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; UNKNOWN / possible (iv) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ. UNKNOWN / possible

slide-16
SLIDE 16

. . . classically

(i) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; ENTAILED / valid (ii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; CONTRADICTION / unsatisfiable (iii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; UNKNOWN / possible (iv) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ. UNKNOWN / possible

slide-17
SLIDE 17

. . . classically

(i) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; ENTAILED / valid (ii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; CONTRADICTION / unsatisfiable (iii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; UNKNOWN / possible (iv) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ. UNKNOWN / possible

slide-18
SLIDE 18

. . . classically

(i) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; ENTAILED / valid (ii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; CONTRADICTION / unsatisfiable (iii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; UNKNOWN / possible (iv) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ. UNKNOWN / possible

slide-19
SLIDE 19

. . . classically

(i) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; ENTAILED / valid (ii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; CONTRADICTION / unsatisfiable (iii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; UNKNOWN / possible (consistency) (iv) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ. UNKNOWN / possible

slide-20
SLIDE 20

. . . classically

(i) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; ENTAILED / valid (ii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; CONTRADICTION / unsatisfiable (iii) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ; UNKNOWN / possible (consistency) (iv) T ∪ {ϕ} | = ψ and T ∪ {ϕ} | = ¬ψ. UNKNOWN / possible (completeness)

slide-21
SLIDE 21

. . . instead

(i) T ∪ {ϕ} | =1.0 ψ and T ∪ {ϕ} | =0.0 ¬ψ; (ii) T ∪ {ϕ} | =0.0 ψ and (iii) T ∪ {ϕ} | =t ψ and T ∪ {ϕ} | =t′ ¬ψ, for 0 < t, t′ < 1.0.

(a) t > t′ (b) t < t′

More generally, for any two candidate entailments

◮ T ∪ {ϕi} |

=ti ¬ψi,

◮ T ∪ {ϕj} |

=tj ¬ψj, decide whether ti > tj, or ti < tj.

slide-22
SLIDE 22

. . . instead

(i) T ∪ {ϕ} | =1.0 ψ and T ∪ {ϕ} | =0.0 ¬ψ; (ii) T ∪ {ϕ} | =0.0 ψ and (iii) T ∪ {ϕ} | =t ψ and T ∪ {ϕ} | =t′ ¬ψ, for 0 < t, t′ < 1.0.

(a) t > t′ (b) t < t′

More generally, for any two candidate entailments

◮ T ∪ {ϕi} |

=ti ¬ψi,

◮ T ∪ {ϕj} |

=tj ¬ψj, decide whether ti > tj, or ti < tj.

slide-23
SLIDE 23

Outline

Informativity, Robustness & Graded Validity Propositional Model Theory & Graded Validity Shallow Inference: Bag-of-Words Encoding Deep Inference: Syllogistic Encoding Computation via the Monte Carlo Method

slide-24
SLIDE 24

Model Theory: Classical Bivalent Logic

Definition

◮ Let Λ = p1, p2, . . . , pN be a propositional language. ◮ Let w = [w1, w2, . . . , wN] be a model.

The truth value · Λ

w is:

⊥ Λ

w = 0;

piΛ

w = wi for all i;

ϕ → ψΛ

w =

           1 if ϕΛ

w = 1 and ψΛ w = 1,

if ϕΛ

w = 1 and ψΛ w = 0,

1 if ϕΛ

w = 0 and ψΛ w = 1,

1 if ϕΛ

w = 0 and ψΛ w = 0;

for all formulae ϕ and ψ over Λ.

slide-25
SLIDE 25

Model Theory: Satisfiability, Validity

Definition

◮ ϕ is valid iff ϕw = 1 for all w ∈ W. ◮ ϕ is satisfiable iff ϕw = 1 for some w ∈ W.

Definition

ϕW = 1 |W|

  • w∈W

ϕw.

Corollary

◮ ϕ is valid iff ϕW = 1. ◮ ϕ is satisfiable iff ϕW > 0.

slide-26
SLIDE 26

Outline

Informativity, Robustness & Graded Validity Propositional Model Theory & Graded Validity Shallow Inference: Bag-of-Words Encoding Deep Inference: Syllogistic Encoding Computation via the Monte Carlo Method

slide-27
SLIDE 27

Bag-of-Words Inference (1)

assume strictly bivalent valuations; Λ = {socrates, is, a, man, so, every}, |W| = 26; (T) socrates ∧ is ∧ a ∧ man ∴ (H) so ∧ every ∧ man ∧ is ∧ socrates ; ΛT = {a}, |WT| = 21; ΛO = {socrates, is, man}, |WO| = 23; ΛH = {so, every}, |WH| = 22; 21 ∗ 23 ∗ 22 = 26;

slide-28
SLIDE 28

Bag-of-Words Inference (2)

How to make this implication false?

◮ Choose the 1 out of 24 = 16 valuations from WT × WO

which makes the antecedent true.

◮ Choose any of the 22 − 1 = 3 valuations from WH which

make the consequent false. ...now compute an expected value. Count zero for the 1 ∗ (22 − 1) = 3 valuations that make this implication false. Count one, for the other 26 − 3. Now T → HW = 26 − 3 26 = 0.95312,

  • r, more generally,

T → HW = 1 − 2|ΛH| − 1 2|ΛT|+|ΛH|+|ΛO| .

slide-29
SLIDE 29

Outline

Informativity, Robustness & Graded Validity Propositional Model Theory & Graded Validity Shallow Inference: Bag-of-Words Encoding Deep Inference: Syllogistic Encoding Computation via the Monte Carlo Method

slide-30
SLIDE 30

Language: Syllogistic Syntax

Let Λ = {x1, x2, x3, y1, y2, y3}; All X are Y =(x1 → y1) ∧ (x2 → y2) ∧ (x3 → y3) Some X are Y =(x1 ∧ y1) ∨ (x2 ∧ y2) ∨ (x3 ∧ y3) All X are not Y =¬ Some X are Y, Some X are not Y =¬ All X are Y,

slide-31
SLIDE 31

Proof theory: A Modern Syllogism

∴ All X are X (S1), Some X are Y ∴ Some X are X (S2), All Y are Z All X are Y ∴ All X are Z (S3), All Y are Z Some Y are X ∴ Some X are Z (S4), Some X are Y ∴ Some Y are X (S5);

slide-32
SLIDE 32

Proof theory: “Natural Logic”

∴ All (red X) are X (NL1), ∴ All cats are animals (NL2), Some X are (red Y) ∴ Some X are Y , Some X are cats ∴ Some X are animals , Some (red X) are Y ∴ Some X are Y , Some cats are Y ∴ Some animals are Y , All X are (red Y) ∴ All X are Y , All X are cats ∴ All X are animals , All X are Y ∴ All (red X) are Y , All animals are Y ∴ All cats are Y ;

slide-33
SLIDE 33

Natural Logic Robustness Properties

Some X are Y ∴ Some X are (red Y) > Some X are Y ∴ Some X are (big (red Y)) , Some X are Y ∴ Some (red X) are Y > Some X are Y ∴ Some (big (red X)) are Y , All X are Y ∴ All X are (red Y) > All X are Y ∴ All X are (big (red Y)) , All (red X) are Y ∴ All X are Y > All (big (red X)) are Y ∴ All X are Y .

slide-34
SLIDE 34

Preliminary Conclusions

(a) “. . . you must be very naive to believe you can reason about language in logic. Even if you could, you’re missing the knowledge to prove things. Even if you had that, logic would still be too computationally complex.” WRONG! (b) “. . . you must be rather ignorant to believe a machine learner will get you anywhere, if all you do is to feed it bags

  • f words. It’s just wrong from the point of view of logic,

epistemology, linguistics, and whatever other theory you should care about.” WRONG!

slide-35
SLIDE 35

Preliminary Conclusions

(a) “. . . you must be very naive to believe you can reason about language in logic. Even if you could, you’re missing the knowledge to prove things. Even if you had that, logic would still be too computationally complex.” WRONG! (b) “. . . you must be rather ignorant to believe a machine learner will get you anywhere, if all you do is to feed it bags

  • f words. It’s just wrong from the point of view of logic,

epistemology, linguistics, and whatever other theory you should care about.” WRONG!

slide-36
SLIDE 36

Preliminary Conclusions

(a) “. . . you must be very naive to believe you can reason about language in logic. Even if you could, you’re missing the knowledge to prove things. Even if you had that, logic would still be too computationally complex.” WRONG! (b) “. . . you must be rather ignorant to believe a machine learner will get you anywhere, if all you do is to feed it bags

  • f words. It’s just wrong from the point of view of logic,

epistemology, linguistics, and whatever other theory you should care about.” WRONG!

slide-37
SLIDE 37

Outline

Informativity, Robustness & Graded Validity Propositional Model Theory & Graded Validity Shallow Inference: Bag-of-Words Encoding Deep Inference: Syllogistic Encoding Computation via the Monte Carlo Method

slide-38
SLIDE 38

Model Theory: Satisfiability, Validity, Expectation

Definition

ϕW = 1 |W|

  • w∈W

ϕw. How do we compute this in general?

Observation

◮ Draw w randomly from a uniform distribution over W.

Now ϕ is the probability that ϕ is true in w.

◮ If W ⊆ W is a random sample over population W, the

sample mean ϕW approaches the population mean ϕW as |W| approaches W.

slide-39
SLIDE 39

Outline

Informativity, Robustness & Graded Validity Propositional Model Theory & Graded Validity Shallow Inference: Bag-of-Words Encoding Deep Inference: Syllogistic Encoding Computation via the Monte Carlo Method