Lecture 10
Discriminative Training, ROVER, and Consensus Michael Picheny, Bhuvana Ramabhadran, Stanley F . Chen
IBM T.J. Watson Research Center Yorktown Heights, New York, USA {picheny,bhuvana,stanchen}@us.ibm.com
Lecture 10 Discriminative Training, ROVER, and Consensus Michael - - PowerPoint PPT Presentation
Lecture 10 Discriminative Training, ROVER, and Consensus Michael Picheny, Bhuvana Ramabhadran, Stanley F . Chen IBM T.J. Watson Research Center Yorktown Heights, New York, USA {picheny,bhuvana,stanchen}@us.ibm.com 10 December 2012 General
IBM T.J. Watson Research Center Yorktown Heights, New York, USA {picheny,bhuvana,stanchen}@us.ibm.com
2 / 90
1
2
3
4
3 / 90
1
4 / 90
5 / 90
6 / 90
7 / 90
8 / 90
1
9 / 90
10 / 90
11 / 90
12 / 90
13 / 90
1
14 / 90
N
15 / 90
N
16 / 90
N
17 / 90
18 / 90
N
19 / 90
N
N
20 / 90
21 / 90
d
N
d
22 / 90
1
23 / 90
c
24 / 90
25 / 90
c
26 / 90
27 / 90
28 / 90
29 / 90
W SB
30 / 90
1
31 / 90
32 / 90
33 / 90
34 / 90
1
2
3
4
35 / 90
2
36 / 90
37 / 90
θ
38 / 90
2
39 / 90
θ
θ
θ
i)p(Sj i)
i refers to the jth possible sentence hypothesis given a
40 / 90
θ
41 / 90
42 / 90
2
43 / 90
44 / 90
i)p(Sj i)
45 / 90
mk
mki (t)
mk
mki (t)
mki (t) are the observation counts for state k, mixture
mki (t) are the counts computed across all the sentence
46 / 90
mk
mki (t)
mk
mki (t)
mki (t) are the observation counts for state k, mixture
mki (t) are the counts computed across all the sentence
47 / 90
mk − Θden mk + Dmkµ′ mk
mk − γden mk + Dmk
48 / 90
49 / 90
50 / 90
51 / 90
i)κp(Sj i)κ
52 / 90
i)κp(Sj i)κ
53 / 90
54 / 90
2
55 / 90
56 / 90
i)κp(Sj i)κ
57 / 90
i)κp(Sj i)κ exp(−bA(Sj i, Sref))
58 / 90
59 / 90
i)κp(Sj i)κ exp(−bA(Sj i, Si))
1 1+e2ρx
60 / 90
61 / 90
62 / 90
63 / 90
1
2
3
4
64 / 90
65 / 90
66 / 90
67 / 90
68 / 90
69 / 90
score(m, n) = min(score(m−1, n−1)+4∗no_match(m, n), score(m−1, n)+3, score(m, n−1)+3)
70 / 90
No so much a ROVER issue but will be important for confusion networks. Problem: How to score relative probabilities and deletions? Solution: no_match(s1,s2)= (1 - p1(winner(s2)) + 1 - p2(winner(s1)))/2
71 / 90
72 / 90
73 / 90
74 / 90
75 / 90
76 / 90
1
2
3
4
77 / 90
78 / 90
79 / 90
SIL SIL SIL SIL SIL SIL VEAL VERY HAVE HAVE HAVE MOVE MOVE HAVE VERY VERY VERY VERY VERY VEAL I I I FINE OFTEN OFTEN FINE IT IT FAST
I
VERY FINE OFTEN FAST HAVE
MOVE
80 / 90
81 / 90
SIL SIL SIL SIL SIL SIL VEAL VERY HAVE HAVE HAVE MOVE MOVE HAVE VERY VERY VERY VERY VERY VEAL I I I FINE OFTEN OFTEN FINE IT IT FAST
(0.45) (0.55) MOVE HAVE I
VERY FINE OFTEN FAST (0.39) IT (0.61)
(0.45) (0.55) MOVE HAVE I
VERY FINE OFTEN FAST (0.39) IT (0.61)
83 / 90
40K 70K 280K 40K MLLR 14 16 18 20 22 24 Acoustic Model Word Error Rate (%) LARGE sLM2 SMALL sLM2 LARGE sLM2+C SMALL sLM2+C LARGE sLM2+C+MX SMALL sLM2+C+MX 84 / 90
85 / 90
86 / 90
87 / 90
88 / 90
89 / 90
90 / 90