Learning biases in
- paque interactions
BRANDON PRICKETT UNIVERSITY OF MASSACHUSETTS, AMHERST NECPHON 2017
Learning biases in opaque interactions BRANDON PRICKETT UNIVERSITY - - PowerPoint PPT Presentation
Learning biases in opaque interactions BRANDON PRICKETT UNIVERSITY OF MASSACHUSETTS, AMHERST NECPHON 2017 Overview 1. Background 4. Conclusions 1. Phonological interactions 1. Which learner predicted the data best? 2. Opaque
BRANDON PRICKETT UNIVERSITY OF MASSACHUSETTS, AMHERST NECPHON 2017
1. Background
1. Phonological interactions 2. Opaque interactions 3. Kiparskian biases 4. Recent approaches to interaction learning
2. Experimental data
1. Methods 2. Results
3. Comparing the models’ performance
1. Expectation Driven Learning with SMR constraints 2. MaxEnt Stratal OT learner 3. A Sequence-to-Sequence neural net
4. Conclusions
1. Which learner predicted the data best? 2. Future work
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
2
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
4
/ti/ /kai/ /tai/
ti
T[+palatal]/_[+front] tʃi
[tʃi] [ki] [tʃi]
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
5
/ti/ /kai/ /tai/
ti
T[+palatal]/_[+front] tʃi
[tʃi] [ki] [tʃi] Deletion moves the /i/ so that it can trigger the /t/ [tʃ] process.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
6
/ti/ /kai/ /tai/
ti
T[+palatal]/_[+front] tʃi
[tʃi] [ki] [tʃi]
/ti/ /kia/ /tia/
ta
T[+palatal]/_[+front] tʃi
[tʃi] [ka] [ta]
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
7
/ti/ /kai/ /tai/
ti
T[+palatal]/_[+front] tʃi
[tʃi] [ki] [tʃi]
/ti/ /kia/ /tia/
ta
T[+palatal]/_[+front] tʃi
[tʃi] [ka] [ta] Deletion removes the /i/ so that it can’t trigger the /t/ [tʃ] process.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
8
applying when it seems like it shouldn’t (counterbleeding).
/ti/ /kai/ /tai/
T[+palatal]/_[+front] tʃi
ti
[tʃi] [ki] [ti]
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
9
applying when it seems like it shouldn’t (counterbleeding).
/ti/ /kai/ /tai/
T[+palatal]/_[+front] tʃi
ti
[tʃi] [ki] [ti] Deletion moves the /i/ next to the /t/, but does so too late to cause the /t/[tʃ] process.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
10
applying when it seems like it shouldn’t (counterbleeding).
/ti/ /kai/ /tai/
T[+palatal]/_[+front] tʃi
ti
[tʃi] [ki] [ti]
/ti/ /kia/ /tia/
T[+palatal]/_[+front] tʃi
tʃa
[tʃi] [ka] [tʃa]
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
11
applying when it seems like it shouldn’t (counterbleeding).
/ti/ /kai/ /tai/
T[+palatal]/_[+front] tʃi
ti
[tʃi] [ki] [ti]
/ti/ /kia/ /tia/
T[+palatal]/_[+front] tʃi
tʃa
[tʃi] [ka] [tʃa] Deletion removes the /i/, but does so too late to stop the /t/[tʃ] process.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
12
learning biases that could influence the likelihood of these interactions arising diachronically:
interactions we just discussed, this bias favors feeding and counterbleeding.
transparent from the surface form (i.e. there are no surface violations of rules that applied somewhere in the grammar). Out of the interactions we just discussed, this bias favors feeding and bleeding.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
13
learning biases that could influence the likelihood of these interactions arising diachronically:
interactions we just discussed, this bias favors feeding and counterbleeding.
transparent from the surface form (i.e. there are no surface violations of rules that applied somewhere in the grammar). Out of the interactions we just discussed, this bias favors feeding and bleeding. Application No Application Transparent Feeding Bleeding Opaque Counterbleeding Counterfeeding
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
14
Application No Application Transparent Feeding Bleeding Opaque Counterbleeding Counterfeeding
learning biases that could influence the likelihood of these interactions arising diachronically:
interactions we just discussed, this bias favors feeding and counterbleeding.
transparent from the surface form (i.e. there are no surface violations of rules that applied somewhere in the grammar). Out of the interactions we just discussed, this bias favors feeding and bleeding. Very favored
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
15
Application No Application Transparent Feeding Bleeding Opaque Counterbleeding Counterfeeding
learning biases that could influence the likelihood of these interactions arising diachronically:
interactions we just discussed, this bias favors feeding and counterbleeding.
transparent from the surface form (i.e. there are no surface violations of rules that applied somewhere in the grammar). Out of the interactions we just discussed, this bias favors feeding and bleeding. Somewhat favored
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
16
Application No Application Transparent Feeding Bleeding Opaque Counterbleeding Counterfeeding
learning biases that could influence the likelihood of these interactions arising diachronically:
interactions we just discussed, this bias favors feeding and counterbleeding.
transparent from the surface form (i.e. there are no surface violations of rules that applied somewhere in the grammar). Out of the interactions we just discussed, this bias favors feeding and bleeding. Not favored
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
17
Application No Application Transparent Feeding Bleeding Opaque Counterbleeding Counterfeeding
learning biases that could influence the likelihood of these interactions arising diachronically:
interactions we just discussed, this bias favors feeding and counterbleeding.
transparent from the surface form (i.e. there are no surface violations of rules that applied somewhere in the grammar). Out of the interactions we just discussed, this bias favors feeding and bleeding.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
18
learnability.
constraints.
style rules and the underlying representations for a toy language.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
19
that directly tested the learnability different phonological interactions.
either individually or as a pair:
/t/ + /-i/ /t/ + /-a/ /t/ + /-i/ and /-a/ /k/ + …
BBleeding
[tʃi] [ta] /tia/ [ta] [ki], [ka], [ka]
FFeeding
[tʃi] [ta] /tai/ [tʃi] [ki], [ka], [ki] Counterbleeding [tʃi] [ta] /tia/ [tʃa] [ki], [ka], [ka]
C Counterfeeding
[tʃi] [ta] /tai/ [ti] [ki], [ka], [ki]
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
21
Experiment Results
participants’ accuracy was ordered in the same order that Kiparsky’s (1968, 1971) biases would predict.
condition that included less explicit instructions for the participants.
model that supported this result.
palatalization, and deletion (the latter two represent which kind of trial subjects’ responses were for).
intercepts were included.
transparency were both significant (p=.025 and .019, respectively).
transparency was not.
22 Application No Application Transparent Feeding Bleeding Opaque Counter- bleeding Counter- feeding
participants’ accuracy was ordered in the same order that Kiparsky’s (1968, 1971) biases would predict.
condition that included less explicit instructions for the participants.
model that supported this result.
palatalization, and deletion (the latter two represent which kind of trial subjects’ responses were for).
intercepts were included.
transparency were both significant (p=.025 and .019, respectively).
transparency was not.
Experiment Results
23
*
Experiment Results
24
participants’ accuracy was ordered in the same order that Kiparsky’s (1968, 1971) biases would predict.
condition that included less explicit instructions for the participants.
model that supported this result.
palatalization, and deletion (the latter two represent which kind of trial subjects’ responses were for).
intercepts were included.
transparency were both significant (p=.025 and .019, respectively).
transparency was not.
*
model the acquisition of different phonological interactions.
markedness constraints.
likely to have produced the training data.
either of the Kiparskian biases.
condition), there was an application bias.
condition), there was a transparency bias.
but feeding had no advantage over bleeding and counterbleeding.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
26
separate runs on each kind of interaction (frequency for each learning datum=18).
average error for each iteration
experiment, the learner struggled the most with the counterfeeding interaction.
learner failed to learn feeding better than bleeding and only had a slight advantage for feeding over counterbleeding.
EDL-HS Results
27
acquisition of bleeding and counterbleeding interactions.
constraints to represent a phonological derivation. It can represent all four kinds of interactions discussed here because of the independence of rankings across strata (although see McCarthy 2007 on the limitations of this theory).
Wilson 2008) version of this framework (see Staubs and Pater 2016 for more on learning MaxEnt grammars in a derivational framework).
in the training data and the probabilities given to these mappings by the model’s grammar.
than counterbleeding.
grammar.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
28
(σ2 = 4000, number of strata=2) for 20 runs on each kind of interaction.
average KL-Divergence on each call to the objective function (i.e. the function that was being minimized by the optimizer).
show a preference for B>F,CF>CB.
the human behavior.
since
Stratal MaxEnt Results
29
Stratal MaxEnt Results
30
(σ2 = 4000, number of strata=2) for 20 runs on each kind of interaction.
average KL-Divergence on each call to the objective function (i.e. the function that was being minimized by the optimizer).
show a preference for B>F,CF>CB.
the human behavior.
since
that converged you get F,B>CB>CF.
Pater (to appear) use in their paper.
(figure adapted from https://medium.com/@devnag/seq2seq-the-clown-car-of-deep-learning-f88e1204dac3) [ a tʃ a ] / a t i a /
such a task.
mine will map from UR to SR.
simulations.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
31
(number of hidden layers=4, number
number of epochs=100) for 20 runs
average mean squared error after 15 epochs.
the subjects in my experiment.
learning the feeding pattern well.
experimental human results, which found feeding to be the most learnable kind of pattern.
Seq2Seq Results
32
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
33
learning curves.
do.
bleeding/counterbleeding.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
34
communication).
surface forms like the subjects were tasked with doing; see Gasser 1993 for a similar approach).
O’Hara 2017 for various methods of learning UR’s in a MaxEnt framework).
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
35
Baković, Eric (2011). Opacity and ordering. In John Goldsmith, Jason Riggle, and Alan Yu (eds.), The Handbook of Phonological Theory (2nd ed). Malden, MA: Wiley-Blackwell. 40-67. Brooks, K. M., Pajak, B., & Bakovic, E. (2013). Learning biases for phonological interactions. Poster presented at 2013 Meeting on Phonology. Rasin, Ezer, Iddo Berger, Nur Lan, and Roni Katzir (2017). Acquiring opaque phonological interactions using minimum description length. Poster preseted at the 2017 Annual Meeting on Phonology Goldwater, Sharon, and Mark Johnson (2003). Learning OT Constraint Rankings Using a Maximum Entropy Model. In Jennifer Spenader, Anders Eriksson, and Osten Dahl (eds.), Proceedings of the Stockholm Workshop on Variation within Optimality Theory. 111–20. Hayes, Bruce, and Colin Wilson (2008). A Maximum Entropy Model of Phonotactics and Phonotactic Learning. Linguistic Inquiry 39, 379–440. Jarosz, Gaja (2015). Expectation Driven Learning of Phonology. Ms., University of Massachusetts Amherst. Jarosz, Gaja (2014). Serial Markedness Reduction. Proceedings of the 2013 Meeting on Phonology 1(1), Amherst, MA. Jarosz, Gaja (2016). Learning Opaque and Transparent Interactions in Harmonic Serialism. In Proceedings of the Annual Meetings on Phonology (Vol. 3). Kim, Yun Jung (2012). Do learners prefer transparent rule ordering? An artificial language learning study. In Proceedings from the Annual Meeting of the Chicago Linguistic Society, 48(1), 375-386. Chicago Linguistic Society. Kiparsky, Paul (1968). Linguistic universals and linguistic change. In Emmon Bach & Robert T. Harms (eds.), Universals in linguistic theory. New York : Holt, Reinhart & Winston. 170–202. Kiparsky, Paul (1971). Historical linguistics. In W. O. Dingwall (ed.) A Survey of Linguistic Science. College Park: University of Maryland Linguistics Program. 576-642. Kirov, Christo (2017). Recurrent Neural Networks as a Strong Domain-General Baseline for Morpho-Phonological Learning. Poster presented at the 2017 Meeting of the Linguistic Society of America. McCarthy, John J. 2007. Hidden generalizations: phonological opacity in Optimality Theory. London: Equinox Press. Nazarov, Aleksei and Joe Pater (to appear). Learning opacity in Stratal Maximum Entropy Grammar. Phonology. Pater, Joe, Karen Jesney, Robert Staubs, and Brian Smith (2012). Learning Probabilities over Underlying Representations. In Proceedings of the Twelfth Meeting of the Special Interest Group on Computational Morphology and Phonology, 62–
Rahman, Fariz (2016). Seq2Seq. Documentation and download available at: https://github.com/farizrahman4u/seq2seq Staubs, Robert, and Pater, Joe (2016). Learning serial constraint-based grammars. In John McCarthy and Joe Pater (eds.), Harmonic Grammar and Harmonic Serialism. London: Equinox Press.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
36
Thanks to the UMass Sound Workshop and Grant Group for helpful feedback on this
about their respective learning models. And a big thanks to Joe Pater and Gaja Jarosz for advising me throughout this project. Experiment subjects were paid using funds from NSF Grant #BCS-1650957.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
37
5 10 15 20 25 30 35
Counterbleeding Counterfeeding
Number of Processes by Type On Environment On Focus
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
40
(Data collection funded by NSF grant BCS-424077; for more information, contact Joe Pater)
plural+diminutive), for a total of 216 trials (108 per phase).
American English).
plural+diminutive). Each picture was randomly paired with a stem.
HITS, with a percentage of acceptance >= 85.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
42
example, the 'f' in 'hoof' changes to a 'v' in the plural form ('hooves’),” and were encouraged to find similar rules in the alien language.
were allowed to proceed.
the circled picture (choice of ‘cat’ vs. ‘dog’).
had to choose an audio file that corresponded to the circled picture (choice of ‘lock’ vs. ‘locks’).
audio file that corresponded to the circled picture (choice of ‘statue’ vs. ‘statuette’).
circled picture that they were shown.
experience in the experiment.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
43
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
44
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
45
subjects and stimuli on the intercepts.
"binomial", glmerControl(optimizer="bobyqa", optCtrl = list(maxfun = 100000)), data = exp_data)
Estimate Std. Error z value Pr(>|z|) (Intercept) 0.84542 0.11646 7.259 3.89e-13 *** Trans ranspare parent nt 0.26 0.26731 731 0 0.114 .11429 29 2.3 2.339 0 39 0.019 .019342 342 * * Appli pplicati cation
0.25 0.25538 538 0 0.114 .11426 26 2.2 2.235 0 35 0.025 025415 415 * * Deletion 0.39945 0.04411 9.057 < 2e-16 *** Palatalization -0.20457 0.04397 -4.652 3.29e-06 *** Transparent:Application 0.02355 0.11513 0.205 0.837909 Transparent:Deletion 0.18515 0.03845 4.816 1.47e-06 *** Application:Deletion -0.01127 0.03845 -0.293 0.769538 Transparent:Palatalization -0.06703 0.03835 -1.748 0.080468 . Application:Palatalization 0.06670 0.03837 1.738 0.082153 . Deletion:Palatalization -0.10457 0.04397 -2.378 0.017388 * Transparent:Application:Deletion 0.06628 0.04095 1.619 0.105515 Transparent:Application:Palatalization -0.06774 0.04086 -1.658 0.097386 . Transparent:Deletion:Palatalization -0.13848 0.03837 -3.609 0.000308 *** Application:Deletion:Palatalization -0.24473 0.03841 -6.372 1.86e-10 *** Transparent:Application:Deletion:Palatalization -0.11768 0.04090 -2.877 0.004011 **
+Delet.
+Palat. Interacting Palatal.
Hiatus Faith +Transpar.
+Applic. Feeding C.B.
Bleeding C.F.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
46
phonological patterns.
2015):
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
48
phonological patterns.
2015):
Constraints
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
49
phonological patterns.
2015):
Probability that Constraint 1 >> Constraint 2
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
50
phonological patterns.
2015):
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
51
phonological patterns.
2015):
P(data | A>>B, grammar) * P(A>>B | grammar) (data | grammar)
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
52
phonological patterns.
2015):
P(data | A>>B, grammar) * P(A>>B | grammar) (data | grammar)
Measured by seeing how
produces each output in the learning data
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
53
phonological patterns.
2015):
P(data | A>>B, grammar) * P(A>>B | grammar) (data | grammar)
This is taken from the current grammar
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
54
phonological patterns.
2015):
P(data | A>>B, grammar) * P(A>>B | grammar) (data | grammar)
This is the sum of the numerator and its complement.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
55
derivation.
shown):
/tia/ SM(*VV, *ti) SM(*ti, *VV) *ti *VV MAX IDENT tia, [tia]
*, * *, *
tʃia, [tʃa]
*, -- ta, [ta]
*, --
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
57
derivation.
shown):
/tia/ SM(*VV, *ti) SM(*ti, *VV) *ti *VV MAX IDENT tia, [tia]
*, * *, *
tʃia, [tʃa]
*, -- ta, [ta]
*, --
Candidates in the first step of the derivation
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
58
derivation.
shown):
/tia/ SM(*VV, *ti) SM(*ti, *VV) *ti *VV MAX IDENT tia, [tia]
*, * *, *
tʃia, [tʃa]
*, -- ta, [ta]
*, --
Candidates in the final step of the derivation
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
59
(weights are analogous to OT constraint rankings).
constraint weights and the input-output probabilities given to the learner.
that output.
values.
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
61
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
62
UR Stratum 1 Output Stratum 2 Output (SR) A A A B B
UR Stratum 1 Output Stratum 2 Output (SR) A A A B B
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
63
UR Stratum 1 Output Stratum 2 Output (SR) A A A B B
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
64
P(Path 1)
UR Stratum 1 Output Stratum 2 Output (SR) A A A B B
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
65
P(Path 1)
P(Path 1) P(Path 2)
UR Stratum 1 Output Stratum 2 Output (SR) A A A B B
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
66
UR Stratum 1 Output Stratum 2 Output (SR) A A A B B
University of Massachusetts, Amherst BPRICKETT@UMASS.EDU - HTTP://PEOPLE.UMASS.EDU/BPRICKETT/
67
P(Path 1) P(Path 2)
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, a segment) at a time.
[ / b a t # / ]
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
b [ / b a t # / ]
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
b [ / b a t # / ]
[
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
b b / b a t # / ]
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
b [ b / b a t # / ]
a
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
b [ b / b a t # / ]
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
b [ b a / b a t # / ]
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
b a [ b a / b a t # / ]
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
b a [ b a / b a t # / ]
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
b a [ b a t / b a t # / ]
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
b a [ b a t / b a t # / ]
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
/ b a t # / b a t [ b a t ]
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
/ b a t # / b a t [ b a ] t
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
/ b a t # / b a t [ b a ] t #
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
/ b a t # / b a t [ b a ] t #
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
/ b a t # / b a t # [ b a ] t s
architecture of our neural net.
through a datum (in our case, a word) one moment (in our case, segments) at a time.
/ b a t # / [ b a ] t s
/tia/ *VV *ti Max Ident tia W* W* L ta * tʃia W* L W* tʃa * W*
ta *VV *ti Max Max [ta] [tʃa] W*
/tai/ *VV *ti Ident Max tai W* L L ti W* L * tʃai W* L L tʃi * *
tʃi *VV *ti Ident Max [tʃi] [ti] W* W*
/tia/ *ti Max
*VV
Ident tia W*
*
L ta W* L L tʃia * * tʃa W* L *
tʃia *ti *VV Max Ident [tʃia] W* L [tʃa] * [tia] W* W* L W* [ta] * W*
/tai/ *ti Ident *VV Max tai * ti W* L W* tʃai W* * tʃi W* L W*
tai Ident *VV *ti Max [tai] W* L L [ti] * * [tʃai] W* W* L L [tʃi] W* L *
/ti/ *ti Ident *VV Max ti W* tʃi W* tʃi Ident *VV *ti Max [tʃi] [ti] W* W*
/tia/ *ti Ident *VV Max SM (*ti, *VV) SM (*VV, *ti) tia <> W* W* L L L ta <*ti+*VV> * * * tʃia <*ti> W* W* L L L
ta <*ti+*VV> *ti Ident *VV Max SM (*ti, *VV) SM (*VV, *ti) [ta] <*ti+*VV> * * tʃa <*ti+*VV> W* * *
/tai/ *VV *ti Ident Max SM (*ti, *VV) SM (*VV, *ti) tai <> W* L L ti <*VV> * * tʃai <*ti> W* L W* L
ti <*VV> *VV *ti Ident Max SM (*ti, *VV) SM (*VV, *ti) ti <*VV> W* L L tʃi <*VV, *ti> * *
/tia/ *ti SM (*ti, *VV) Ident *VV Max SM (*VV, *ti) tia <> W* L * ta <*ti+*VV> W* L L W* W* tʃia <*ti> * *
tʃia <*ti> *ti SM (*ti, *VV) Ident *VV Max SM (*VV, *ti) tʃia <*ti> W* L L tʃa <*ti, *VV> * *
/tai/ SM (*ti, *VV) *VV *ti Ident Max SM (*VV, *ti) tai <> W* L L ti <*VV> * * tʃai <*ti> W* L W* L
ti <*VV> SM (*ti, *VV) *VV *ti Ident Max SM (*VV, *ti) [ti] <*VV> * tʃi <*VV, *ti> W* L W*