1
A Probabilistic Model of Cross- situational Word Learning from Noisy and Ambiguous Data
Afra Alishahi
Joint work with Afsaneh Fazly and Suzanne Stevenson, University of Toronto
A Probabilistic Model of Cross- situational Word Learning from - - PowerPoint PPT Presentation
A Probabilistic Model of Cross- situational Word Learning from Noisy and Ambiguous Data Afra Alishahi Joint work with Afsaneh Fazly and Suzanne Stevenson, University of Toronto 1 Word Learning Word learning: a mapping between a word and
1
Joint work with Afsaneh Fazly and Suzanne Stevenson, University of Toronto
2
Word learning: a mapping between a word and its
Mappings are learned from exposure to word usages
the chimp eats apples
3
Which aspect of a scene is described by a
a black chimp is sitting
the chimp eats apples there are two red apples in his hands
4
What word refers to what part of the meaning?
the chimp eats apples
5
What word refers to what part of the meaning?
{black, animal, living, chimp, eyes, hands, feet, red, apple, fruit, edible, food, rock,
action, consume, sit, hold, …} the chimp eats apples
6
Meaning of a word is learned by detecting meaning
daddy is picking apples the chimp eats apples
7
Cross-situational learning does not explain various
Many specific principles are proposed to explain
A unified model of word learning is needed to
8
Implement an incremental probabilistic account of
Explain observed patterns without incorporating
Handle referential uncertainty and ambiguity. Learn word–meaning mappings from naturally
9
Input is a sequence of utterance–scene pairs: Meaning of each word is represented as a set of
10
An adaptation of a model for finding corresponding
[Brown et al.’93]
Each input pair is processed in two steps:
11
Alignment probabilities: Meaning probabilities:
wk ∈U (t )
s=1 t
s=1 t
m j ∈M
12
13
14
15
daddy human glasses pick …
16
daddy human glasses pick …
17
daddy human glasses pick mommy I desire plate green …
18
A word is learned when most of its probability mass
Comprehension score:
m j ∈Tw
19
Utterances from Manchester corpus in CHILDES
20
… paired with meaning primitives extracted from
touch, deed, …
21
… and subsequent primitive sets are combined to
touch, deed, …
22
Change in proportion of learned words over time:
23
24
25
26
27
Young children can easily determine the meaning of
28
29
Not clear whether children “learn” the meaning of a
30
Fast mapping is due to a specialized mechanism for
[Markman & Wachtel’88; Golinkoff et al.’92; Gopnik & Meltzoff’87]
Fast mapping arises from general processes of
[Clark’90; Diesendruck & Markson’01, Halberda’06]
31
Input: a sequence of utterance–scene pairs: Output: a probability distribution over meaning
{ THE, CHIMP, EAT, AN, APPLE, SIT, ON, ROCK, HAND, LEAF }
{ DADDY, PICK, APPLE, TREE, SUNGLASSES, LEAF }
{ SEE, THE, RED, APPLE, ON, ROCK, GREEN, PLATE}
32
Familiar target: Novel target:
33
Familiar target:
34
Familiar target:
35
Familiar target:
Use meaning probability p(.|apple)
36
Familiar target:
Use meaning probability p(.|apple)
37
Novel target:
38
Novel target:
39
Novel target:
Meaning probabilities are not informative:
40
Novel target:
Meaning probabilities are not informative: Use referent probability rf (dax |.):
41
Novel target:
Use referent probability rf (dax |.):
42
43
Perform induction over recently-acquired knowledge
The model correctly maps dax to its referent.
44
45
Induction over two recently fast-mapped and one
The presence of a third novel object can be
46
47
Developed an incremental probabilistic model that
Future directions: