the core method
play

The Core Method Steffen H olldobler International Center for - PowerPoint PPT Presentation

The Core Method Steffen H olldobler International Center for Computational Logic Technische Universit at Dresden Germany The Very Idea The Propositional CORE Method Human Reasoning Steffen H olldobler The Core Method 1


  1. The Core Method Steffen H¨ olldobler International Center for Computational Logic Technische Universit¨ at Dresden Germany ◮ The Very Idea ◮ The Propositional CORE Method ◮ Human Reasoning Steffen H¨ olldobler The Core Method 1

  2. The Very Idea ◮ Various semantics for logic programs coincide with fixed points of associated immediate consequence operators (Apt, van Emden: Contributions to the Theory of Logic Programming. Journal of the ACM 29, 841-862: 1982). ◮ Banach Contraction Mapping Theorem A contraction mapping f defined on a complete metric space ( X , d ) has a unique fixed point. The sequence y , f ( y ) , f ( f ( y )) , . . . converges to this fixed point for any y ∈ X . ⊲ Consider programs whose immediate consequence operator is a contraction. (Fitting: Metric Methods – Three Examples and a Theorem. Journal of Logic Programming 21, 113-127: 1994). ◮ Every continuous function on the reals can be uniformly approximated by feed- forward connectionist networks (Funahashi: On the Approximate Realization of Continuous Mappings by Neural Networks. Neural Networks 2, 183-192: 1989). ⊲ Consider programs whose immediate consequence operator is continuous. (H., Kalinke, St¨ orr: Approximating the Semantics of Logic Programs by Recurrent Neural Networks. Applied Intelligence 11, 45-59: 1999). Steffen H¨ olldobler The Core Method 2

  3. First Ideas ◮ H., Kalinke: Towards a New Massively Parallel Computational Model for Logic Programming In: Proceedings of the ECAI94 Workshop on Combining Symbolic and Connectionist Processing, 68-77: 1994. Steffen H¨ olldobler The Core Method 3

  4. Interpretations ◮ Let L be a propositional language and {⊤ , ⊥} the set of truth values. ◮ An interpretation I is a mapping L → {⊤ , ⊥} . ◮ For a given program P , an interpretation I can be represented by the set of atoms occurring in P which are mapped to ⊤ under I , i.e. I ⊆ R P . ◮ 2 R P is the set of all interpretations for P . ◮ ( 2 R P , ⊆ ) is a complete lattice. ◮ An interpretation I for P is a model for P iff I ( P ) = ⊤ . Steffen H¨ olldobler The Core Method 4

  5. Immediate Consequence Operator P : 2 R P → 2 R P : ◮ Immediate consequence operator T P ( I ) = { A | there is a clause A ← L 1 ∧ . . . ∧ L n ∈ P T such that I | = L 1 ∧ . . . ∧ L n } . ◮ I is a supported model iff T P ( I ) = I . ◮ Let lfp T P be the least fixed point of T P if it exists. Steffen H¨ olldobler The Core Method 5

  6. The Propositional CORE Method ◮ Let L be a propositional logic language. ◮ Given a logic program P together with immediate consequence operator T P . ◮ Let |R P | = m and 2 R P be the set of interpretations for P . ◮ Find a mapping rep : 2 R P → R m . ◮ Construct a feed-forward network computing f P : R m → R m , called the core, such that the following holds: P ( I ) = J then f P ( rep ( I )) = rep ( J ) , where I , J ∈ 2 R P . ⊲ If T P ( rep − 1 ( s )) = rep − 1 ( t ) , where s , t ∈ R m . ⊲ If f P ( s ) = t then T ◮ Connect the units in the output layer recursively to the units in the input layer. ◮ Show that the following holds ⊲ I = lfp T P iff the recurrent network converges to rep ( I ) , i.e. it reaches a stable state with input and output layer representing rep ( I ) . ◮ Connectionist model generation using recurrent networks with feed-forward core. Steffen H¨ olldobler The Core Method 6

  7. 3-Layer Recurrent Networks core 1 . . . . . . . . . 1 input layer hidden layer output layer ◮ At each point in time all units do: ⊲ apply activation function to obtain potential, ⊲ apply output function to obtain output. Steffen H¨ olldobler The Core Method 7

  8. Propositional CORE Method using Binary Threshold Units ◮ Let L be the language of propositional logic. ◮ Let P be a propositional logic program, e.g., P = { p , r ← p ∧ ∼ q , r ← ∼ p ∧ q } . ◮ T P ( I ) = { A | A ← L 1 ∧ . . . ∧ L m ∈ P such that I | = L 1 ∧ . . . ∧ L m } . T P ( ∅ ) = { p } P ( { p } ) { p , r } T = P ( { p , r } ) { p , r } T = = lfp T P Steffen H¨ olldobler The Core Method 8

  9. Representing Interpretations ◮ 2 R P ◮ Let m = |R P | and identify R P with { 1 , . . . , m } . ◮ Define rep : 2 R P → R m such that for all 1 ≤ j ≤ m we find: � 1 if j ∈ I , rep ( I )[ j ] = 0 if j �∈ I . E.g., if R P = { p , q , r } = { 1 , 2 , 3 } and I = { p , r } then rep ( I ) = ( 1 , 0 , 1 ) . ◮ Other encodings are possible, e.g., � 1 if j ∈ I , rep ′ ( I )[ j ] = − 1 if j �∈ I . ◮ We can represent interpretations by arrays of binary or bipolar threshold units. Steffen H¨ olldobler The Core Method 9

  10. Computing the Core ◮ Theorem For each program P , there exists a core of logical threshold units computing T P . ◮ Proof Let P be a program, m = |R P | , and ω ∈ R + . Wlog we assume that all occurrences of ⊤ in P have been eliminated. ⊲ Translation Algorithm 1 Input and output layer: vector of length m of binary threshold units with threshold 0 . 5 and ω/ 2 in the input and output layer, respectively. 2 For each clause of the form A ← L 1 ∧ . . . ∧ L k ∈ P , k ≥ 0, do: 2.1 Add a binary threshold unit u h to the hidden layer. 2.2 Connect u h to the unit representing A in the output layer with weight ω . 2.3 For each literal L j , 1 ≤ j ≤ k , connect the unit representing L j in the input layer to u h and, if L j is an atom then set the weight to ω otherwise set the weight to − ω . 2.4 Set the threshold of u h to ( p − 0 . 5 ) ω , where p is the number of positive literals occurring in L 1 ∧ . . . ∧ L k . Steffen H¨ olldobler The Core Method 10

  11. Computing the Core (Continued) ◮ Theorem For each program P , there exists a core of logical threshold units computing T P . ◮ Proof (Continued) Some observations: ⊲ u h becomes active at time t + 1 iff L 1 ∧ . . . ∧ L k is mapped to ⊤ by the interpretation represented by the state of the input layer at time t . ⊲ The unit representing A in the output layer becomes active at time t + 2 iff there is a rule of the form A ← L 1 ∧ . . . ∧ L k ∈ P and the unit u h in the hidden layer corresponding to this rule is active at time t + 1. ⊲ The result follows immediately from these observations. qed Steffen H¨ olldobler The Core Method 11

  12. Computing the Core (Example) ◮ Consider again P = { p , r ← p ∧ ∼ q , r ← ∼ p ∧ q } . ◮ The translation algorithm yields: ω - ω ω 1 p p 2 2 2 ω − ω 1 ω ω q q 2 2 2 − ω ω ω ω 1 ω ω r r 2 2 2 Steffen H¨ olldobler The Core Method 12

  13. Hidden Layers are Needed ◮ The XOR can be represented by the program { r ← p ∧ ∼ q , r ← ∼ p ∧ q } . ◮ Proposition 2-layer networks cannot compute T P for definite P . ◮ Proof Suppose there exist 2-layer networks for definite P computing T P . ⊲ Consider P = { p ← q , p ← r ∧ s , p ← t ∧ u } . p 1 θ 7 7 p . 5 w 72 ω q 2 8 q . 5 2 ω . 5 r 3 9 r 2 ω s 4 . 5 10 s w 76 2 ω t 5 . 5 11 t 2 ω u 6 . 5 12 u 2 ⊲ Let v be the state of the input layer; v is an interpretation. Steffen H¨ olldobler The Core Method 13

  14. Hidden Layers are Needed (Continued) ◮ Proposition 2-layer networks cannot compute T P for definite P . ◮ Proof (Continued) Consider P = { p ← q , p ← r ∧ s , p ← t ∧ u } . ⊲ We have to find θ 7 and w 7 j , 2 ≤ j ≤ 6, such that p ∈ T P ( v ) iff w 72 v 2 + w 73 v 3 + w 74 v 4 + w 75 v 5 + w 76 v 6 ≥ θ 7 . ⊲ Because conjunction is commutative we find p ∈ T P ( v ) iff w 72 v 2 + w 74 v 3 + w 73 v 4 + w 76 v 5 + w 75 v 6 ≥ θ 7 . ⊲ Consequently, p ∈ T P ( v ) iff w 72 v 2 + w 1 ( v 3 + v 4 ) + w 2 ( v 5 + v 6 ) ≥ θ 7 , where w 1 = 1 2 ( w 73 + w 74 ) and w 2 = 1 2 ( w 75 + w 76 ) . Steffen H¨ olldobler The Core Method 14

  15. Hidden Layers are Needed (Continued) ◮ Proposition 2-layer networks cannot compute T P for definite P . ◮ Proof (Continued) Consider P = { p ← q , p ← r ∧ s , p ← t ∧ u } . ⊲ Likewise, because disjunction is commutative we find p ∈ T P ( v ) iff w · x ≥ θ 7 , 3 ( w 72 + w 1 + w 2 ) and x = � 7 where w = 1 j = 2 v i . ⊲ For the network to compute T P the following must hold: ◮ If x = 0 ( v 2 = . . . = v 6 = 0) then w · x − θ 7 < 0. ◮ ◮ If x = 1 ( v 2 = 1 , v 3 = . . . = v 6 = 0) then w · x − θ 7 ≥ 0. ◮ ◮ If x = 2 ( v 2 = v 4 = v 6 = 0 , v 3 = v 5 = 1) then w · x − θ 7 < 0. ◮ ⊲ However, d ( w · x − θ 7 ) = w cannot change its sign; contradiction. dx ⊲ Consequently, 2-layer feed-forward networks cannot compute T qed P Steffen H¨ olldobler The Core Method 15

  16. Adding Recurrent Connections ◮ Recall P = { p , r ← p ∧ ∼ q , r ← ∼ p ∧ q } . - ω - ω ω ω . 5 . 5 2 2 2 2 ω ω ω . 5 2 2 2 ω ω ω . 5 . 5 2 2 2 Steffen H¨ olldobler The Core Method 16

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend