Speech Question Answering TOEFL Listening Comprehension Test by - - PowerPoint PPT Presentation

speech question answering
SMART_READER_LITE
LIVE PREVIEW

Speech Question Answering TOEFL Listening Comprehension Test by - - PowerPoint PPT Presentation

Speech Question Answering TOEFL Listening Comprehension Test by Machine Wei Fang December 13, 2017 Speech Processing & Machine Learning Lab 1 Question Answering (QA) Understand spoken content Answer questions about spoken content


slide-1
SLIDE 1

Speech Question Answering

TOEFL Listening Comprehension Test by Machine

Wei Fang December 13, 2017

Speech Processing & Machine Learning Lab 1

slide-2
SLIDE 2

Question Answering (QA)

  • Understand spoken content
  • Answer questions about spoken content

2

slide-3
SLIDE 3

Question Answering (QA)

  • Understand spoken content
  • Answer questions about spoken content

2

slide-4
SLIDE 4

New Task: TOEFL Listening Comprehension Test by Machine

  • TOEFL: Test of English as a Foreign Language
  • Listening Section:
  • Listen to a 3 5 minute story
  • Answer question with a set of answer choices

3

slide-5
SLIDE 5

New Task: TOEFL Listening Comprehension Test by Machine

  • TOEFL: Test of English as a Foreign Language
  • Listening Section:
  • Listen to a 3 5 minute story
  • Answer question with a set of answer choices

3

slide-6
SLIDE 6

New Task: TOEFL Listening Comprehension Test by Machine

  • TOEFL: Test of English as a Foreign Language
  • Listening Section:
  • Listen to a 3∼5 minute story
  • Answer question with a set of answer choices

3

slide-7
SLIDE 7

New Task: TOEFL Listening Comprehension Test by Machine

  • TOEFL: Test of English as a Foreign Language
  • Listening Section:
  • Listen to a 3∼5 minute story
  • Answer question with a set of answer choices

3

slide-8
SLIDE 8

New Task: TOEFL Listening Comprehension Test by Machine

Dataset

  • Past exams collected from a TOEFL practice website
  • Splits - train/dev/test: 717/124/122
  • Audio stories with two transcriptions:

manual, ASR (CMU Sphinx with WER) Approach

4

slide-9
SLIDE 9

New Task: TOEFL Listening Comprehension Test by Machine

Dataset

  • Past exams collected from a TOEFL practice website
  • Splits - train/dev/test: 717/124/122
  • Audio stories with two transcriptions:

manual, ASR (CMU Sphinx with WER) Approach

4

slide-10
SLIDE 10

New Task: TOEFL Listening Comprehension Test by Machine

Dataset

  • Past exams collected from a TOEFL practice website
  • Splits - train/dev/test: 717/124/122
  • Audio stories with two transcriptions:

manual, ASR (CMU Sphinx with WER) Approach

4

slide-11
SLIDE 11

New Task: TOEFL Listening Comprehension Test by Machine

Dataset

  • Past exams collected from a TOEFL practice website
  • Splits - train/dev/test: 717/124/122
  • Audio stories with two transcriptions:

manual, ASR (CMU Sphinx with 34.32% WER) Approach

4

slide-12
SLIDE 12

New Task: TOEFL Listening Comprehension Test by Machine

Dataset

  • Past exams collected from a TOEFL practice website
  • Splits - train/dev/test: 717/124/122
  • Audio stories with two transcriptions:

manual, ASR (CMU Sphinx with 34.32% WER) Approach

4

slide-13
SLIDE 13

Neural Network Model Architecture

The entire model learned end-to-end.

5

slide-14
SLIDE 14

Baseline NN Model: LSTM

Hermann, Kočiský, Grefenstette, Espeholt, Kay, Suleyman, Blunsom. Teaching Machines to Read and Comprehend. NIPS 2015.

6

slide-15
SLIDE 15

Attending to Relevant Sentences in Story

Note: Bi-directional RNNs

7

slide-16
SLIDE 16

Attending to Relevant Sentences in Story

Note: Bi-directional RNNs

7

slide-17
SLIDE 17

Attending to Relevant Sentences in Story

Tseng, Shen, Lee, Lee. Towards Machine Comprehension of Spoken Content: Initial TOEFL Listening Comprehension Test by Machine. Interspeech 2016.

8

slide-18
SLIDE 18

Sentence Representations

Tai, Socher, Manning. Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks. ACL 2015.

9

slide-19
SLIDE 19

Sentence Representations

Tai, Socher, Manning. Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks. ACL 2015.

9

slide-20
SLIDE 20

Hierarchical Attention

Sequential Attention

 

  • Hierarchical Attention

Fang, Hsu, Lee, Lee. Hierarchical Attention Model for Improved Machine Comprehension of Spoken Content. SLT 2016.

10

slide-21
SLIDE 21

Experimental Results

Question-Choice Similarity Sliding Window LSTM +Attention Tree-LSTM Tree-LSTM +Attention 0.25 0.30 0.35 0.40 0.45 0.50

Random

Accuracy

11

slide-22
SLIDE 22

Analysis

There are 3 types of questions.

Type 3: Connecting Information

  • Understanding Organization
  • Connecting Content
  • Making Inferences

12

slide-23
SLIDE 23

Analysis

There are 3 types of questions.

Type 2: Pragmatic Understanding

  • Function of What is Said
  • Speaker’s Attitude

12

slide-24
SLIDE 24

Analysis

There are 3 types of questions.

Type 2: Pragmatic Understanding

  • Function of What is Said
  • Speaker’s Attitude

Example:

What is the purpose of the man’s response? What can be inferred about the student? 12

slide-25
SLIDE 25

Transfer Learning from Movie QA

Motivation

TOEFL is a small dataset; transfer from larger QA dataset (MovieQA) to improve performance.

Tapaswi, Zhu, Stiefelhagen, Torralba, Urtasun, Fidler. MovieQA: Understanding Stories in Movies through Question-Answering Tree-LSTM +Attention Chung et al. Chung et al. (transfer)

Accuracy

Chung, Lee, Glass. Supervised and Unsupervised Transfer Learning for Question Answering. arXiv 2017.

13

slide-26
SLIDE 26

Transfer Learning from Movie QA

Motivation

TOEFL is a small dataset; transfer from larger QA dataset (MovieQA) to improve performance.

Tapaswi, Zhu, Stiefelhagen, Torralba, Urtasun, Fidler. MovieQA: Understanding Stories in Movies through Question-Answering Tree-LSTM +Attention Chung et al. Chung et al. (transfer) 0.50 0.55

Accuracy

Chung, Lee, Glass. Supervised and Unsupervised Transfer Learning for Question Answering. arXiv 2017.

13

slide-27
SLIDE 27

Conclusion

  • Introduced a new task TOEFL Listening Comprehension Test

by Machine.

  • Proposed attention-based models to outperform previous

methods.

  • Performance can be improved by transfer learning from a

larger QA dataset.

14

slide-28
SLIDE 28

Thanks

Contact Wei Fang b40815@gmail.com

14