An introduction to computational psycholinguistics: Modeling human sentence processing
Shravan Vasishth University of Potsdam, Germany http://www.ling.uni-potsdam.de/∼vasishth vasishth@acm.org September 2005, Bochum
Neural structure
1
An introduction to computational psycholinguistics: Modeling human - - PDF document
An introduction to computational psycholinguistics: Modeling human sentence processing Shravan Vasishth University of Potsdam, Germany http://www.ling.uni-potsdam.de/ vasishth vasishth@acm.org September 2005, Bochum Neural structure 1 A
1
2
3
4
5
n
j=
6
n
j=
7
8
9
10
11
12
13
14
15
16
j (tk−ok) k 17
18
19
hidden, z(t) input, x(t) hidden, z(t-1) copy z
20
21
22
23
24
25
26
27
28
29
j∈W
j∈W
30
31
i∈G
i∈U
i∈G
32
j∈G
33
34
35
36
Aside 37
Aside 38
400 500 600 700 800 900 1000 English self−paced reading experiment Position
Mean Reading Time (msecs)
Det N1 who Det N2 that Det N3 had V3 to the NP had V2 Adv V1 the NP
Grammatical, Similar NPs Grammatical, Dissimilar NPs Ungrammatical, Similar NPs Ungrammatical, Dissimilar NPs Similar NPs Dissimilar NPs Encoding interference Storage and Retrieval interference (Un−)Grammaticality
Aside 39
1000 1500 2000 German self−paced reading experiment Position
Mean Reading Time (msecs)
NP1 who NP2 who NP3 V3 V2 V1 NP
Grammatical, Similar NPs Grammatical, Dissimilar NPs Ungrammatical, Similar NPs Ungrammatical, Dissimilar NPs Similar NPs Dissimilar NPs Encoding interference Storage and Retrieval interference (Un−)Grammaticality
40
1. Making “experience” more realistic: The corpora do not reflect frequencies of occurrence of center embeddings and/or center-embedding types. The technology for computing these is now available (e.g. (Korthals, 2001; Kurz, 2000)). Modify the training corpora frequencies to reflect our best possible estimate of these frequencies and then examine the performance
2. Probabilistic models of processing versus SRNs: C compare their model’s performance against simple n-gram models. Replicate their experiment results, and try to find a measure using probabilistic methods that can outperform their system or do as well. 3. Extending empirical coverage: modeling Suckow et al.’s data: Without messing with the architecture, carry out simulations to model the interference effects and the verb reading pattern asymmetries for English and German found by (Suckow, Vasishth, & Lewis, 2005) The poster handout of this paper is available from my home page. The simulations should be able to account for all the data covered by C as well as the new data (i.e.regression testing is necessary for a meaningful step forward). 4. The Chinese and German relative clause puzzle: (Konieczny & M¨ uller, 2004) showed that word order may not be able to explain the German S/O relative clause facts: in German, the verb is final in both S and O relatives, and yet the ORs are
should be able to explain this in terms of experience (word order). How to reconcile these two opposing results in an SRN architecture? 5. Extending the coverage with Japanese double embeddings: Will the network perform well with 5-clause NPs and 2 verbs (ditransitive plus transitive)? We’d probably have to add HUs. What does this mean? Relevant empirical work: (Lewis & Nakayama, 2001) and references cited there. 41
42
43
44
45
46
47
48
49
50
51
52
53
54