a framework for representing
play

A Framework for Representing Language Acquisition in a Population - PowerPoint PPT Presentation

A Framework for Representing Language Acquisition in a Population Setting Jordan Kodner Christopher Cerezo Falco University of Pennsylvania ACL - July 16, 2018 Melbourne Language Change Languages change over time Both an internal and


  1. A Framework for Representing Language Acquisition in a Population Setting Jordan Kodner Christopher Cerezo Falco University of Pennsylvania ACL - July 16, 2018 Melbourne

  2. Language Change Languages change over time Both an internal and external process ● Fundamentally social ● Individuals acquire language and transmit it to future generations ● New variants propagate through populations ● Modelling Change Must model how the individual reacts to linguistic input and to the ● community 2

  3. Example - The Cot-Caught Merger /ɒ/ “ cot ” is pronounced the same ● as /ɔ/ “ caught ” Minimal pairs distinguished by ● /ɒ/~/ɔ/ become homophones /ɒ/ /ɔ/ cot caught Don Dawn collar caller Merged Unmerged knotty naughty odd awed pond pawned 3

  4. Example - The Cot-Caught Merger /ɒ/ “ cot ” is pronounced the same ● as /ɔ/ “ caught ” Present in many dialects of North ● American English Eastern New England ○ Western Pennsylvania ○ Lower Midwest ○ West ○ Canada (all) ○ Merged Unmerged 4

  5. Example - The Cot-Caught Merger /ɒ/ “ cot ” is pronounced the same ● as /ɔ/ “ caught ” Present in many dialects of North ● American English Eastern New England ○ Western Pennsylvania ○ Lower Midwest ○ West ○ Canada (all) ○ Merged Spreading into Rhode Island ● Unmerged (Johnson 2007) 5

  6. Example - The Cot-Caught Merger /ɒ/ “ cot ” is pronounced the same ● as /ɔ/ “ caught ” Present in many dialects of North ● American English Eastern New England ○ Western Pennsylvania ○ Lower Midwest ○ West ○ Canada (all) ○ Merged Spreading into Rhode Island ● Unmerged Rapid! Families with Non-merged ● parents and older siblings but merged younger siblings 6

  7. Existing Frameworks 7

  8. Three Classes of Framework 1. Swarm Frameworks 2. Network Frameworks 3. Algebraic Frameworks 8

  9. Three Classes of Framework 1. Swarm Frameworks Individual agents on a grid moving randomly and interacting (ABM) ○ e.g., Harrison et al. 2002, Satterfield 2001, Schulze et al. 2008, Stanford & ○ Kenny 2013 9

  10. Three Classes of Framework 1. Swarm Frameworks Individual agents on a grid moving randomly and interacting (ABM) ○ e.g., Harrison et al. 2002, Satterfield 2001, Schulze et al. 2008, Stanford & ○ Kenny 2013 + Bloomfield (1933)’s Principle of Density for free + Diffusion is straightforward - Not a lot of control over the network - Thousands of degrees of freedom -> should run many many times -> slow 10

  11. Three Classes of Framework 1. Swarm Frameworks 2. Network Frameworks Speakers are nodes in a graph, edges are possibility of interaction ○ e.g., Baxter et al. 2006, Baxter et al. 2009, Blythe & Croft 2012, Fagyal et ○ al. 2010, Minett & Wang 2008, Kauhanen 2016 11

  12. Three Classes of Framework 1. Swarm Frameworks 2. Network Frameworks Speakers are nodes in a graph, edges are possibility of interaction ○ e.g., Baxter et al. 2006, Baxter et al. 2009, Blythe & Croft 2012, Fagyal et ○ al. 2010, Minett & Wang 2008, Kauhanen 2016 + Much more control over network structure + Easy to model concepts from the sociolinguistic lit. (e.g., Milroy & Milroy) - Nodes only interact with immediate neighbours -> slow and less realistic? - Practically implemented as random interactions between neighbours -> same problem as #1 12

  13. Three Classes of Framework 1. Swarm Frameworks 2. Network Frameworks 3. Algebraic Frameworks Expected outcome of interactions is calculated analytically ○ e.g., Abrams & Stroganz 2003, Baxter et al. 2006, Minett & Wang 2008, ○ Niyogi & Berwick 1997, Yang 2000, Niyogi & Berwick 2009 13

  14. Three Classes of Framework 1. Swarm Frameworks 2. Network Frameworks 3. Algebraic Frameworks Expected outcome of interactions is calculated analytically ○ e.g., Abrams & Stroganz 2003, Baxter et al. 2006, Minett & Wang 2008, ○ Niyogi & Berwick 1997, Yang 2000, Niyogi & Berwick 2009 + Closed-form solution rather than simulation -> faster and more direct - No network structure! Always implemented over perfectly mixed populations 14

  15. Three Classes of Framework 1. Swarm Frameworks 2. Network Frameworks 3. Algebraic Frameworks This proliferation of “boutique” frameworks is a problem An ad hoc framework risks “overfitting” the pattern ● Comparison between frameworks is challenging ● 15

  16. Our Framework 16

  17. Best of All Worlds Impose density effects on a network structure and calculate the outcome of each iteration analytically 17

  18. Best of All Worlds Impose density effects on a network structure and calculate the outcome of each iteration analytically Swarm + Captures the Principle of Density Network + Models key facts about social networks Algebraic + No random process in the core algorithm 18

  19. The Model Language change as a two-step loop 1. Propagation: Variants distribute through the network 2. Acquisition: Individuals internalize them 19

  20. Vocabulary L : That which is transmitted Language ≈ Variant ≈ Sample G : That which generates/describes/distinguishes L That which is learned/influenced by L Grammar ≈ Variety ≈ Latent Variable 20

  21. Binary G Examples G : {Merged grammar, Non-merged grammar} L : Merged or non-merged instances of cot and caught words G : { Dived -generating grammar, Dove -generating grammar} L : Instances of the past tense of dive as dived or dove G : { have +NEG = haven’t got grammar, have +NEG = don’t have grammar} L : Instances of haven’t got and instances of don’t have 21

  22. The Model Language change as a two-step loop 1. Propagation: L distributes through the network 2. Acquisition: Individuals react to L to create G If this were a linear chain, L 0 → G 1 → L 1 → G 2 → L 2 → … → L n → G n+1 → ... 22

  23. The Model Language change as a two-step loop 1. Propagation: L distributes through the network 2. Acquisition: Individuals react to L to create G Generic. Not problem-specific. 23

  24. Intuition behind Propagation Algorithm For T iterations, For the individual at each node Begin travelling ; While travelling Randomly select outgoing edge by weight and follow it OR stop; Increase chance of stopping next time; End Interact with the individual at the current Node; End End 24

  25. Intuition behind Propagation Algorithm For T iterations, For the individual at each node Nodes are not individuals. Begin travelling ; Individuals “stand on” nodes While travelling Randomly select outgoing edge by weight and follow it OR stop; Increase chance of stopping next time; End Interact with the individual at the current node; End End 25

  26. Intuition behind Propagation Algorithm For T iterations, Weighted or unweighted, For the individual at each node Directed or undirected Begin travelling ; Individuals “travel” along While travelling edges and find someone to Randomly select outgoing edge interact with by weight and follow it OR stop; Increase chance of stopping next time; End Interact with the individual at the current node; End End 26

  27. Intuition behind Propagation Algorithm For T iterations, Weighted or unweighted, For the individual at each node Directed or undirected Begin travelling ; Determine who this node While travelling Individuals connected by Randomly select outgoing edge shorter or higher weighted by weight and follow it OR stop; paths are more likely to Increase chance of stopping next time; interact. End Interact with the individual at the current node; End End 27

  28. Intuition behind Propagation Algorithm For T iterations, Weighted or unweighted, For the individual at each node Directed or undirected Begin travelling ; While travelling Rather than simulating Randomly select outgoing edge interactions in a loop, by weight and follow it OR stop; calculate a closed-form Increase chance of stopping next time; solution End Interact with the individual at the current node; End End 28

  29. The Propagation Function E = G T α(I - (1 - α) A) -1 29

  30. The Propagation Function E = G T α(I - (1 - α) A) -1 The Linguistic Environment E is a g x n matrix: n individuals, g possible grammars ● For each individual, the proportion of input drawn from each grammar ● 30

  31. The Propagation Function E = G T α(I - (1 - α) A) -1 The Linguistic Environment Distribution of Grammars Of the previous generation ● G is an n x g matrix ● Proportions by which each individual produces L ● 31

  32. The Propagation Function E = G T α(I - (1 - α) A) -1 The Linguistic Environment Distribution of Grammars Interaction Probabilities A is an n x n adjacency matrix ● The probabilities that nodes i , j interact given that the number of ● steps travelled declines by a geometric distribution α parameter from that distribution [0,1] ● 32

  33. The Acquisition Function ● Problem-specific ● Should take E t as input and produce G t+1 as output T ● In the simplest case ( neutral change ), G t+1 = E t ● The following case study uses a variational learner 33

  34. Case Study Spread of the Cot-Caught Merger 34

  35. Model for Merger Acquisition (Yang 2009) Learners will acquire the merged grammar iff more than ~17% of their environment is merged 35

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend