strongly incremental repair detection
play

Strongly Incremental Repair Detection Julian Hough 1 , 2 and Matthew - PowerPoint PPT Presentation

Strongly Incremental Repair Detection Julian Hough 1 , 2 and Matthew Purver 2 1 Dialogue Systems Group and CITEC, University of Bielefeld 2 Cognitive Science Research Group, Queen Mary University of London October 26th 2014, EMNLP Doha, Qatar


  1. Strongly Incremental Repair Detection Julian Hough 1 , 2 and Matthew Purver 2 1 Dialogue Systems Group and CITEC, University of Bielefeld 2 Cognitive Science Research Group, Queen Mary University of London October 26th 2014, EMNLP Doha, Qatar Hough and Purver EMNLP 2014

  2. Problem statement 1 STIR: Strongly Incremental Repair Detection 2 Edit terms Repair start Reparandum start Repair end Evaluation measures for repair 3 Experiments and results 4 Conclusions and Future 5 Hough and Purver EMNLP 2014

  3. Problem statement 1 STIR: Strongly Incremental Repair Detection 2 Edit terms Repair start Reparandum start Repair end Evaluation measures for repair 3 Experiments and results 4 Conclusions and Future 5 Hough and Purver EMNLP 2014

  4. Self-repairs “But one of the, the two things that I’m really . . . ” “Our situation is just a little bit, kind of the opposite of that” “and you know it’s like you’re, I mean, employments are contractual by nature anyway” [Switchboard examples] Hough and Purver EMNLP 2014

  5. Self-repairs: Annotation scheme John [ likes + { uh } loves ] Mary � �� � ���� � �� � interregnum repair reparandum [Shriberg, 1994, onwards] Terminology: edit terms , interruption point (+), repair onset Hough and Purver EMNLP 2014

  6. Self-repairs: classes “But one of [ the, + the ] two things that I’m really . . . ” [repeat] “Our situation is just [ a little bit, + kind of the opposite ] of that” [substitution] “and you know it’s like [ you’re + { I mean } ] employments are contractual by nature anyway” [delete] [Switchboard examples] Hough and Purver EMNLP 2014

  7. Self-repair detection: why do we care? Dialogue systems (parsing speech) Hough and Purver EMNLP 2014

  8. Self-repair detection: why do we care? Dialogue systems (parsing speech) Hough and Purver EMNLP 2014

  9. Self-repair detection: why do we care? Dialogue systems (parsing speech) Hough and Purver EMNLP 2014

  10. Self-repair detection: why do we care? Dialogue systems (parsing speech) Hough and Purver EMNLP 2014

  11. Self-repair detection: why do we care? Interpreting self-repair Preserving the reparandum and repair structure Evidence: [Brennan and Schober, 2001] showed subjects use the reparandum to make faster decisions: “Pick the yell-purple square” faster “Pick the uhh-purple square” Hough and Purver EMNLP 2014

  12. Self-repair detection: why do we care? Interpreting self-repair Preserving the reparandum and repair structure Evidence: [Brennan and Schober, 2001] showed subjects use the reparandum to make faster decisions: “Pick the yell-purple square” faster “Pick the uhh-purple square” Self-repairs have meaning! Dialogue systems should not filter out the reparandum! Hough and Purver EMNLP 2014

  13. Self-repair detection: why do we care? Interpreting self-repair Preserving the reparandum and repair structure Evidence: [Brennan and Schober, 2001] showed subjects use the reparandum to make faster decisions: “Pick the yell-purple square” faster “Pick the uhh-purple square” Self-repairs have meaning! Dialogue systems should not filter out the reparandum! Accuracy evaluation Standard evaluation F-score on reparandum words Also interested in repair structure assignment! Hough and Purver EMNLP 2014

  14. Self-repair detection: why do we care? Interpreting self-repair Preserving the reparandum and repair structure Evidence: [Brennan and Schober, 2001] showed subjects use the reparandum to make faster decisions: “Pick the yell-purple square” faster “Pick the uhh-purple square” Self-repairs have meaning! Dialogue systems should not filter out the reparandum! Accuracy evaluation Standard evaluation F-score on reparandum words Also interested in repair structure assignment! Hough and Purver EMNLP 2014

  15. Self-repair detection: Incrementality Non-incremental vs. Incremental Dialogue Systems [Schlangen and Skantze, 2011] Hough and Purver EMNLP 2014

  16. Self-repair detection: Incrementality We want good incremental performance : Timing - Low latency, short time to detect repairs Evolution over time - Responsiveness of the detection (incremental accuracy) - Stability of the output (low jitter) Computational complexity - Minimal processing overhead (fast) Hough and Purver EMNLP 2014

  17. Self-repair detection Problem statement A system that achieves: Interpretation of repair - repair structure tags rather than just reparandum words Strong incrementality - Give the best results possible as early as possible - Computationally fast Controllable trade-off between incrementality and overall accuracy Hough and Purver EMNLP 2014

  18. Previous approaches: Noisy channel model Best coverage generative model [Zwarts et al., 2010, Johnson and Charniak, 2004] S-TAG exploits ( ‘rough copy’ ) dependency with string alignment [Zwarts et al., 2010] utterance-final F-score = 0.778 Hough and Purver EMNLP 2014

  19. Previous approaches: Noisy channel model Best coverage generative model [Zwarts et al., 2010, Johnson and Charniak, 2004] S-TAG exploits ( ‘rough copy’ ) dependency with string alignment [Zwarts et al., 2010] utterance-final F-score = 0.778 Two incremental measures: - Time-to-detection : 7.5 words from reparandum onset - 4.6 words from repair onset - Delayed accuracy : slow rise up to 6 words back Complexity O ( n 5 ) Hough and Purver EMNLP 2014

  20. Previous approaches: Noisy channel model Why poor incremental performance? Hough and Purver EMNLP 2014

  21. Previous approaches: Noisy channel model Why poor incremental performance? - Inherently non-incremental string-alignment - Utterance global (c.f. spelling correction) - Sparsity of alignment forms [Hough and Purver, 2013] Hough and Purver EMNLP 2014

  22. SOLUTION: Information theory and strong incrementality Local measures of fluency for minimum latency in detection Does not just rely on string alignment Information theoretic measures of language models [Keller, 2004, Jaeger and Tily, 2011] Minimal complexity Hough and Purver EMNLP 2014

  23. Problem statement 1 STIR: Strongly Incremental Repair Detection 2 Edit terms Repair start Reparandum start Repair end Evaluation measures for repair 3 Experiments and results 4 Conclusions and Future 5 Hough and Purver EMNLP 2014

  24. STIR: Strongly Incremental Repair Detection John [ likes + { uh } loves ] Mary � �� � ���� � �� � reparandum interregnum repair ... [ rm start ... rm end + { ed } rp start ... rp end ] ... Hough and Purver EMNLP 2014

  25. STIR: Strongly Incremental Repair Detection ... [ rm start ... rm end + { ed } rp start ... rp end ] ... ... { ed } ... Hough and Purver EMNLP 2014

  26. STIR: Strongly Incremental Repair Detection Hough and Purver EMNLP 2014

  27. STIR: Strongly Incremental Repair Detection “John” S 0 S 1 Hough and Purver EMNLP 2014

  28. STIR: Strongly Incremental Repair Detection “John” “likes” S 0 S 1 S 2 Hough and Purver EMNLP 2014

  29. STIR: Strongly Incremental Repair Detection “John” “likes” “uh” ed S 0 S 1 S 2 S 3 ed Hough and Purver EMNLP 2014

  30. STIR: Strongly Incremental Repair Detection “John” “likes” “uh” “loves” rp start ? ed S 0 S 1 S 2 S 3 S 4 rp start ed Hough and Purver EMNLP 2014

  31. STIR: Strongly Incremental Repair Detection “John” “likes” “uh” “loves” rp start ? S 0 S 1 S 4 rm end ed S 2 S 3 rp start rm end ed Hough and Purver EMNLP 2014

  32. STIR: Strongly Incremental Repair Detection “John” “likes” “uh” “loves” rp start S 0 S 1 S 4 rm start rm end ed S 2 S 3 rp start rm start rm end ed Hough and Purver EMNLP 2014

  33. STIR: Strongly Incremental Repair Detection “John” “likes” “uh” “loves” rp start S 0 S 1 S 4 rm start rp sub end rm end ed S 2 S 3 rp start rp sub rm start rm end ed end Hough and Purver EMNLP 2014

  34. STIR: Strongly Incremental Repair Detection “Mary” “John” “likes” “uh” “loves” rp start S 0 S 1 S 4 S 5 rm start rp sub end rm end ed S 2 S 3 rp start rp sub rm start rm end ed end Hough and Purver EMNLP 2014

  35. STIR: fluency modelling using enriched n-gram LMs s ( w i − 2 , w i − 1 , w i ) (surprisal) WML( w i − 2 , w i − 1 , w i ) (syntactic fluency) H ( θ ( w | c )) (entropy) KL ( θ ( w | c a ) , θ ( w | c b )) (distribution divergence) Hough and Purver EMNLP 2014

  36. STIR: fluency modelling using enriched n-gram LMs s ( w i − 2 , w i − 1 , w i ) (surprisal) WML( w i − 2 , w i − 1 , w i ) (syntactic fluency) H ( θ ( w | c )) (entropy) KL ( θ ( w | c a ) , θ ( w | c b )) (distribution divergence) p lex (word) and p pos (POS) models Does not use lexical or POS values , but information theoretic measures [Keller, 2004, Jaeger and Tily, 2011, Clark et al., 2013] Hough and Purver EMNLP 2014

  37. STIR: fluency modelling using enriched n-gram LMs rp start local deviation from fluency : drop in WML lex 0.0 −0.2 −0.4 −0.6 WML −0.8 −1.0 −1.2 −1.4 any very care i havent had good really good experience with child Hough and Purver EMNLP 2014

  38. STIR: fluency modelling using enriched n-gram LMs Extend ‘rough copy’ dependency [Johnson and Charniak, 2004] to gradient measures Information content = entropy Parallelism = distributional similarity Hough and Purver EMNLP 2014

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend