Automatic Speech Recognition (CS753)
Lecture 19: Search, Decoding and LatuicesAutomatic Speech Recognition (CS753)
Automatic Speech Recognition (CS753) Automatic Speech Recognition - - PowerPoint PPT Presentation
Automatic Speech Recognition (CS753) Automatic Speech Recognition (CS753) Lecture 19: Search, Decoding and La tu ices Instructor: Preethi Jyothi Mar 27, 2017 Recap: Static and Dynamic Networks Static network: Build compact decoding graph
Automatic Speech Recognition (CS753)
Lecture 19: Search, Decoding and LatuicesAutomatic Speech Recognition (CS753)
Recap: Static and Dynamic Networks
Static Network Decoding
Searching the graph
Viterbi beam search decoder
Trellis with full Viterbi & beam search
No beam search With beam searchBeam search algorithm
Initialization: current states := initial state while (current states do not contain the goal state) do: successor states := NEXT(current states) where NEXT is next state function score the successor states set current states to a pruned set of successor states using beam width δA* stack decoder
Recall A* algorithm
A* stack decoder
Which hypotheses should be extended?
A* stack decoder
Fast-match
A* stack decoder
Example (1)
Image from [JM]: Jurafsky & Martin, SLP 2nd edition, Chapter 10Example (2)
A* vs Beam search
Latuice Generation
Latuice Generation
Word Confusion Networks
Word confusion networks are normalised word latuices that provide alignments for a fraction of word sequences in the word latuice HAVE HAVE HAVE I I MOVE VERY VERY I SIL SIL VEAL OFTEN OFTEN SIL SIL SIL SIL FINE I T VERY FAST VERY MOVE HAVE IT (a) Word Lattice I HAVE IT VEAL FINEConstructing word confusion network