Lifted Relational Neural Networks
Gustav Sourek, Vojtech Aschenbrenner, Filip Zelezny & Ondrej Kuzelka
Lifted Relational Neural Networks Gustav Sourek, Vojtech - - PowerPoint PPT Presentation
Lifted Relational Neural Networks Gustav Sourek, Vojtech Aschenbrenner, Filip Zelezny & Ondrej Kuzelka Outline Motivation From Neural Nets point of view (possibly) From Markov Logic point of view What are Lifted Relational
Gustav Sourek, Vojtech Aschenbrenner, Filip Zelezny & Ondrej Kuzelka
2
3
4
5
(mutagenicity) classification problem
6
the problem in advance?
We may design anonymous predicates for these patterns
different contexts (rules) (Muggleton,1988) Neural learning of latent (non ground) patterns
7
aforementioned limitation of propositionalization
Markov Logic Networks(Richardson, Domingos,2005), Bayesian Logic Programs(Kersting, De Raedt,2000)
KBANN(Towel, Shavlik,1994), CILP(Franca, Zaverucha, Garcez,1999)
8
9
the presence of uncertainty?
10
11
12
a given LRNN model corresponds to an atom neuron
derived * from a given LRNN corresponds to a rule neuron
rule’s ground head {h(b1
1, …, b1 k), …, h(bn 1, …, bn k)}
there is an aggregation neuron * meaning it is present in the least Herbrand model
13
14
as their neurons are tied to the same template rules
activations so as to reflect the (fuzzy) logic of disjunction, conjunction, and different forms
aggregative reasoning over relational patterns
15
16
sets of corresponding facts (typically ground unit clauses)
sample can thus be though of simply as another LRNN
17
Different samples result in different LRNNs
these are 3 different types of neurons : atom neurons, rule neurons, aggregation neurons
18
given LRNN corresponds to an atom neuron
Set of all atom neurons:
NbondOH(h1,o1),NbondOH(h2,o1) } (* Meaning present in the least Herbrand model of it)
19
given LRNN corresponds to an atom neuron
Set of all atom neurons:
20
be derived* from a given LRNN corresponds to a rule neuron
Set of all rule neurons: NbondOH(h1,o1) H(h1), O(o1), bond(h1,o1) , NbondOH(h2,o1) H(h2), O(o1), bond(h2,o1) (*Meaning the atoms are true in the least Herbrand model)
21
be derived* from a given LRNN corresponds to a rule neuron
22
atom neurons (rule’s body) have high outputs
23
rule having the same ground literal in the head. For each such aggregation there is an aggregation neuron.
1.0 : bondOH(X,Y) :- H(X), O(Y), bond(X,Y)
24
different logic of aggregation neurons
25
yet with different weights
weights:
0.2 : NGroup1 :- hasHCl
26
atom naturally corresponds to disjunction
at least one of the rule neurons has high output
27
28
a regular neural network with shared weights
template’s clause and exploit sample regularities
not pose any problem to weight learning
adaptions can be efficiently used for training
29
30
0.0 atomGroup1(X) :- o(X). 0.0 atomGroup1(X) :- cl(X). .... 0.0 atomGroup3(X) :- cl(X). …. 0.0 bondGroup3(X) :- 2=(X). …. graphlet0 :- atomGroup2(X), bond(X,Y,B1), bondGroup1(B1), atomGroup3(Y)… …. 0.0 class1 :- graphlet0. …. 0.0 class1 :- graphlet242.
31
32
33
relational autoencoders,…
34
Gustav Sourek1 , Suresh Manandhar2 , Filip Zelezny1 , Steven Schockaert3 , and Ondrej Kuzelka3
1) Czech Technical University in Prague, Czech Republic {souregus, zelezny}@fel.cvut.cz 2) Department of Computer Science, University of York, UK suresh.manandhar@york.ac.uk 3) School of CS & Informatics, Cardiff University, UK {SchockaertS1, KuzelkaO}@cardiff.ac.uk
36
Learning Predictive Categories with LRNNs
We consider a following (learning) scenario with latent categories:
1.
Entities
2.
Properties
37
1.
Given : a set of entities and corresponding lists of their properties
2.
Assumption : there exists some latent hierarchy of categories that are predictive of their corresponding
3.
Goal : Learn suitable category structures from data
38
w’cecp : HasProperty(A, B) ← IsA(A, ce), IsA(B, cp), HasProperty(ce, cp)
39
and their targets {1/0 HasProperty(e, p)} via SGD
𝑙
bi − k + b0 )
𝑙
bi + b0 )
categories for both objects and properties
40
reported in (Statistical Predicate Invention, Kok and Domingos, 2007)
clustering objects and relations
41
42
43
category membership degrees
wl : HasProperty(A, B) ← HasProperty(C, B), Similar(A, C, l)
externally obtained embeddings
considered to learn soft categories of predicates, too
44
non-trivial SRL scenarios
entities and their properties
with just mild extensions of the template
incorporation of LRNNs into NLP tasks pipelines
45
graphical models (e.g., Markov Logic Networks)
corresponding ground models
convolutional NN), Explicit variable binding
Structure learning inspired by meta-interpretive learning
46
See “Lifted Relational Neural Networks” at arXiv.org for more details
47
48