Combining Speech and Speaker Recognition - A Joint Modeling Approach - - PowerPoint PPT Presentation

combining speech and speaker recognition a joint modeling
SMART_READER_LITE
LIVE PREVIEW

Combining Speech and Speaker Recognition - A Joint Modeling Approach - - PowerPoint PPT Presentation

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Combining Speech and Speaker Recognition - A Joint Modeling


slide-1
SLIDE 1

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Combining Speech and Speaker Recognition

  • A Joint Modeling Approach

Hang Su

Supervised by: Prof. N. Morgan, Dr. S. Wegmann EECS, University of California, Berkeley, CA USA International Computer Science Institute, Berkeley, CA USA

August 16, 2018

Hang Su Dissertation Talk 1 / 71

slide-2
SLIDE 2

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Table of contents

1

Introduction and Motivation

2

Backgrounds on Speech and Speaker Recognition

3

Connecting Speech and Speaker Recognition

4

Joint Modeling of Speech and Speaker

5

Conclusion and Future Work

Hang Su Dissertation Talk 2 / 71

slide-3
SLIDE 3

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Introduction Motivation An ideal AI agent for speech

Table of contents

1

Introduction and Motivation

2

Backgrounds on Speech and Speaker Recognition

3

Connecting Speech and Speaker Recognition

4

Joint Modeling of Speech and Speaker

5

Conclusion and Future Work

Hang Su Dissertation Talk 3 / 71

slide-4
SLIDE 4

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Introduction Motivation An ideal AI agent for speech

Joint modeling of speech and speaker

The brief idea Automatic speech recognition (ASR)

translate speech to text automatically

Speaker recognition or speaker identification

identify speakers from characteristics of voice

Combining speech and speaker recognition

capture speech and speaker characteristics together

Hang Su Dissertation Talk 4 / 71

slide-5
SLIDE 5

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Introduction Motivation An ideal AI agent for speech

Why speech / speaker recognition

Application of speech & speaker recognition Human-Computer Interface Automatic speech recognition

In-car system, smart home, speech search...

Speaker recognition

Authentication, safety, personalization...

Hang Su Dissertation Talk 5 / 71

slide-6
SLIDE 6

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Introduction Motivation An ideal AI agent for speech

A problem

They are handled separately Different datasets / evaluations Different models / methods But they are closely related to each other Take speech as input Similar features / models

Hang Su Dissertation Talk 6 / 71

slide-7
SLIDE 7

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Introduction Motivation An ideal AI agent for speech

A problem

They are handled separately Different datasets / evaluations Different models / methods But they are closely related to each other Take speech as input Similar features / models (Same group of researchers :)

Hang Su Dissertation Talk 6 / 71

slide-8
SLIDE 8

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Introduction Motivation An ideal AI agent for speech

An ideal AI agent for speech

Hang Su Dissertation Talk 7 / 71

slide-9
SLIDE 9

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Introduction Motivation An ideal AI agent for speech

An ideal AI agent for speech

Hang Su Dissertation Talk 8 / 71

slide-10
SLIDE 10

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Automatic Speech Recognition Speaker Recognition

Table of contents

1

Introduction and Motivation

2

Backgrounds on Speech and Speaker Recognition

3

Connecting Speech and Speaker Recognition

4

Joint Modeling of Speech and Speaker

5

Conclusion and Future Work

Hang Su Dissertation Talk 9 / 71

slide-11
SLIDE 11

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Automatic Speech Recognition Speaker Recognition

Table of contents

1

Introduction and Motivation

2

Backgrounds on Speech and Speaker Recognition Automatic Speech Recognition Speaker Recognition

3

Connecting Speech and Speaker Recognition

4

Joint Modeling of Speech and Speaker

5

Conclusion and Future Work

Hang Su Dissertation Talk 10 / 71

slide-12
SLIDE 12

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Automatic Speech Recognition Speaker Recognition

Automatic Speech Recognition (ASR)

Transcribe speech into texts Frame-by-frame approach (10 ~30 ms) Components∗:

Feature extraction Acoustic modeling (GMM-HMM) Lexicon Language modeling (LM)

Or use end-to-end approach: discard HMM, optionally discard lexicon or language model

∗For a traditional ASR system.

Hang Su Dissertation Talk 11 / 71

slide-13
SLIDE 13

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Automatic Speech Recognition Speaker Recognition

Traditional ASR pipeline

Hang Su Dissertation Talk 12 / 71

slide-14
SLIDE 14

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Automatic Speech Recognition Speaker Recognition

Gaussian Mixture Model - HMM[9, 3]

Hang Su Dissertation Talk 13 / 71

slide-15
SLIDE 15

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Automatic Speech Recognition Speaker Recognition

Deep Neural Network - HMM[1, 11]

Hang Su Dissertation Talk 14 / 71

slide-16
SLIDE 16

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Automatic Speech Recognition Speaker Recognition

Long-Short Term Memory - HMM [8]

Hang Su Dissertation Talk 15 / 71

slide-17
SLIDE 17

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Automatic Speech Recognition Speaker Recognition

Table of contents

1

Introduction and Motivation

2

Backgrounds on Speech and Speaker Recognition Automatic Speech Recognition Speaker Recognition

3

Connecting Speech and Speaker Recognition

4

Joint Modeling of Speech and Speaker

5

Conclusion and Future Work

Hang Su Dissertation Talk 16 / 71

slide-18
SLIDE 18

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Automatic Speech Recognition Speaker Recognition

Speaker Recognition

Speaker Recognition: Identify speakers from speech Components:

Feature extraction Acoustic modeling Speaker modeling Scoring

Make utterance-level predictions

Hang Su Dissertation Talk 17 / 71

slide-19
SLIDE 19

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Automatic Speech Recognition Speaker Recognition

Text-independent speaker recognition

Hang Su Dissertation Talk 18 / 71

slide-20
SLIDE 20

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Automatic Speech Recognition Speaker Recognition

Factor analysis approach [2]

xt ∼

K

  • k

πk N(µk + Akzi, Σk) zi ∼ N(0, I)

K

  • k=1

πk = 1 (1) xt is p-dim speech feature for frame t πk is prior for mixture k zi : a q-dim speaker specific latent factor (i.e. i-vector) Ak : a p-by-q projection matrix for mixture c µk and Σk are Gaussian parameters

Hang Su Dissertation Talk 19 / 71

slide-21
SLIDE 21

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Automatic Speech Recognition Speaker Recognition

Post-processing of i-vectors

The factor-analysis model is an unsupervised model. Supervised methods could be used to improve i-vectors. Linear Discriminant Analysis [6] Probabilistic Linear Discriminant Analysis [6, 5]

Hang Su Dissertation Talk 20 / 71

slide-22
SLIDE 22

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Table of contents

1

Introduction and Motivation

2

Backgrounds on Speech and Speaker Recognition

3

Connecting Speech and Speaker Recognition

4

Joint Modeling of Speech and Speaker

5

Conclusion and Future Work

Hang Su Dissertation Talk 21 / 71

slide-23
SLIDE 23

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Table of contents

1

Introduction and Motivation

2

Backgrounds on Speech and Speaker Recognition

3

Connecting Speech and Speaker Recognition Speaker Recognition using ASR Speaker Adaptation Conclusion

4

Joint Modeling of Speech and Speaker

5

Conclusion and Future Work

Hang Su Dissertation Talk 22 / 71

slide-24
SLIDE 24

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Speaker recognition using ASR

Hang Su Dissertation Talk 23 / 71

slide-25
SLIDE 25

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Speaker recognition using ASR cont.

Substitute UBM with DNN model [7] Substitute UBM with Time-delay DNN [13] Use DNN initialized GMM acoustic model [13] Proposal: Use better DNN models for ASR †

Trained with raw MFCC feature Trained with LDA transformed feature Trained with LDA + fMLLR transformed feature Trained with Minimum Phone Error (MPE) method

†Factor Analysis Based Speaker Verification Using ASR.

Hang Su and Steven Wegmann. Interspeech 2016

Hang Su Dissertation Talk 24 / 71

slide-26
SLIDE 26

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Data description

Speaker recognition evaluation (SRE) data set Training data (SRE 2004-2008)

18,715 recordings from 3,009 speakers 1,000+ hours of data, 360,000,000 frame samples

Test data (SRE 2010)

387,112 trials (98% non-target) 11,983 enrollment speakers, 767 test speakers 2 ~3 mins per speaker

ASR data set Training data (Switchboard) Testing data (Eval2000)

Hang Su Dissertation Talk 25 / 71

slide-27
SLIDE 27

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Metric – DET curve and EER

Hang Su Dissertation Talk 26 / 71

slide-28
SLIDE 28

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Metric – Word Error Rate (WER)

WER = S + D + I R (2) S : number of substitutions D : number of deletions I : number of insertions R : number of words in references

Hang Su Dissertation Talk 27 / 71

slide-29
SLIDE 29

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Experimental results

Eval2000 WER SRE2010 EER UBM – 6.31 DNN-MFCC 19.4 6.39 + LDA + MLLT 16.3 4.84 + fMLLR∗ 14.9 4.55 + MPE∗ 13.5 4.38

Table 1: EER for speaker recognition systems in different settings

∗ASR decoding needed

Hang Su Dissertation Talk 28 / 71

slide-30
SLIDE 30

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Experimental results

Figure 1: DET curve for systems in different settings

Hang Su Dissertation Talk 29 / 71

slide-31
SLIDE 31

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Table of contents

1

Introduction and Motivation

2

Backgrounds on Speech and Speaker Recognition

3

Connecting Speech and Speaker Recognition Speaker Recognition using ASR Speaker Adaptation Conclusion

4

Joint Modeling of Speech and Speaker

5

Conclusion and Future Work

Hang Su Dissertation Talk 30 / 71

slide-32
SLIDE 32

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Speaker Adaptation

How to handle speaker-specific characteristics during recognition? Adapt speaker-independent systems to different speakers (model-space) Normalize speech features to compensate speaker characteristics (feature-space)

Hang Su Dissertation Talk 31 / 71

slide-33
SLIDE 33

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Speaker adaptation for DNN systems

Existing methods: Feature-space transformations (fMLLR) [4] Model-space transformations [15] Adapting model parameters via regularization [16] Learning hidden unit contributions (LHUC) [14]

Hang Su Dissertation Talk 32 / 71

slide-34
SLIDE 34

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Speaker adaptation using i-vectors[10]

h = Wax + Wsz

Hang Su Dissertation Talk 33 / 71

slide-35
SLIDE 35

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Speaker adaptation using i-vectors

Benefits of using i-vectors Does not require model re-training or ASR decoding Single DNN model for all speakers Potential drawback: Tend to overfit

Hang Su Dissertation Talk 34 / 71

slide-36
SLIDE 36

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Problem of speaker adaptation using i-vector

I-vectors are extracted for every recordings Frames 100 million, 4,800 recordings Acoustic feature dim ~440, i-vector dim 100~400 Better objective on training data does not translate into WER improvement Overfitting occurs

Hang Su Dissertation Talk 35 / 71

slide-37
SLIDE 37

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Treatment for overfitting

Mitigate overfitting by Reducing i-vector dimension[10] Using utterance-based i-vectors[12] Extract i-vectors using sliding window (in Kaldi) L2 regularization back to baseline DNN[12]

Hang Su Dissertation Talk 36 / 71

slide-38
SLIDE 38

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Regularization on i-vector sub-nnetwork

Lre = Lce + βwivec2

Hang Su Dissertation Talk 37 / 71

slide-39
SLIDE 39

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Data description

Switchboard data set Clean telephone speech, English ~300 hours transcribed data (~108,000,000 samples) ~4,800 recordings Eval2000 hub5 test set Switchboard portion + CallHome (family members) 40 + 40 speakers 2 hours + 1.6 hours

Hang Su Dissertation Talk 38 / 71

slide-40
SLIDE 40

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Metric – Word Error Rate (WER)

WER = S + D + I R (3) S : number of substitutions D : number of deletions I : number of insertions R : number of words in references

Hang Su Dissertation Talk 39 / 71

slide-41
SLIDE 41

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Experimental results

feature MFCC +fMLLR data Swbd Callhome Swbd Callhome acoustic feature 16.0 28.5 14.9 25.6 + i-vector 15.2 27.1 14.4 25.7 + regularization 14.6 26.3 14.3 24.9

Table 2: WER on i-vector adaptation using regularization

Hang Su Dissertation Talk 40 / 71

slide-42
SLIDE 42

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Speaker Recognition using ASR Speaker Adaptation Conclusion

Conclusion

A brief summarization: Speech and speaker recognition are two tasks that are closely related Speaker information can be used to improve speech recognition performance Acoustic models trained for ASR can be used to assist speaker recognition

Hang Su Dissertation Talk 41 / 71

slide-43
SLIDE 43

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

Table of contents

1

Introduction and Motivation

2

Backgrounds on Speech and Speaker Recognition

3

Connecting Speech and Speaker Recognition

4

Joint Modeling of Speech and Speaker

5

Conclusion and Future Work

Hang Su Dissertation Talk 42 / 71

slide-44
SLIDE 44

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

Table of contents

1

Introduction and Motivation

2

Backgrounds on Speech and Speaker Recognition

3

Connecting Speech and Speaker Recognition

4

Joint Modeling of Speech and Speaker TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

5

Conclusion and Future Work

Hang Su Dissertation Talk 43 / 71

slide-45
SLIDE 45

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

Existing Tools for Speech

Kaldi: Popular speech recognition tool Supports GMM, HMM, DNN, LSTM .... State-of-the-art recipes Tensorflow (TF) Flexible deep learning research framework Tensorflow Lite: esay to deploy on embedded devices Tensor Processing Unit (TPU)

Hang Su Dissertation Talk 44 / 71

slide-46
SLIDE 46

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

TIK

Bridge the gap between Tensorflow and Kaldi It supports acoustic modeling using Tensorflow It integrates with Kaldi decoder through a pipe It covers both speech and speaker recognition tasks

Hang Su Dissertation Talk 45 / 71

slide-47
SLIDE 47

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

System Design of TIK

Hang Su Dissertation Talk 46 / 71

slide-48
SLIDE 48

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

ASR performance using TIK

Swbd CallHome All Kaldi GMM 21.4 34.8 28.2 Kaldi DNN 14.9 25.6 20.3 TIK DNN 14.5 25.5 20.0 TIK BLSTM 13.6 24.3 19.0

Table 3: WER of ASR systems trained with Kaldi and TIK (Eval2000 test set)

Hang Su Dissertation Talk 47 / 71

slide-49
SLIDE 49

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

Speaker recognition performance using TIK

Cosine LDA PLDA Kaldi UBM 6.91 3.36 2.51 Kaldi DNN 4.00 1.83 1.27 TIK DNN 4.53 2.00 1.27

Table 4: EER of speaker recognition systems trained Kaldi and TIK (SRE2010 test set)

Hang Su Dissertation Talk 48 / 71

slide-50
SLIDE 50

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

Table of contents

1

Introduction and Motivation

2

Backgrounds on Speech and Speaker Recognition

3

Connecting Speech and Speaker Recognition

4

Joint Modeling of Speech and Speaker TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

5

Conclusion and Future Work

Hang Su Dissertation Talk 49 / 71

slide-51
SLIDE 51

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

X-vector approach

Figure 2: Structure of x-vector approach for speaker recognition

Hang Su Dissertation Talk 50 / 71

slide-52
SLIDE 52

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

JointDNN model

Figure 3: Structure of JointDNN model

Hang Su Dissertation Talk 51 / 71

slide-53
SLIDE 53

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

Loss function

L(θ) = −

S

  • s=1

sT

  • t=1

hs,t log P(hs,t|os,t) − β

S

  • s=1

xs log P(xs|os) (4) Interpolation of two cross-entropy losses β is the interpolation weight hs,t denotes the HMM state for frame t of segment s

  • s,t is the observed feature vector

xs is the correct speaker

  • s is speech features for segment s

Hang Su Dissertation Talk 52 / 71

slide-54
SLIDE 54

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

Data description

Training data Switchboard data set ~300 hours transcribed data (~108,000,000 samples) ~520 speakers Testing data Eval2000 hub5 test set for speech recognition SRE2010 test set for speaker recognition

Hang Su Dissertation Talk 53 / 71

slide-55
SLIDE 55

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

Performance of speaker recognition

EER Baseline i-vector 4.85 Kaldi x-vector 8.94 TIK x-vector 8.81 TIK jd-vector (beta0.01) 4.75

Table 5: EER of JointDNN model for speaker recognition (SRE2010 test set)

Hang Su Dissertation Talk 54 / 71

slide-56
SLIDE 56

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

Performance of speaker recognition

Figure 4: DET curve of JointDNN model for speaker recognition (SRE2010 test set)

Hang Su Dissertation Talk 55 / 71

slide-57
SLIDE 57

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

Performance of speech recognition

Swbd Callhome All Baseline DNN 16.1 28.4 22.3 JointDNN (beta 0.01) 16.8 29.0 22.9

Table 6: WER of JointDNN model for speech recognition

Hang Su Dissertation Talk 56 / 71

slide-58
SLIDE 58

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

Adjusting Interpolation Weight β

Development (%) Evaluation (%) Beta ASR acc Speaker acc SRE EER Swbd WER 0.1 39.07 97.22 5.10 16.7 0.01 39.20 94.10 4.75 16.8 0.001 38.60 85.36 9.19 17.2 0.0001 38.59 41.95 13.25 17.0

Table 7: EER of JointDNN model with different β

Hang Su Dissertation Talk 57 / 71

slide-59
SLIDE 59

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work TIK: An Open-source Tool JointDNN for speech and speaker recognition Conclusion

Conclusion

Summary of JointDNN model JointDNN can be used for ASR and SRE simultaneously ASR part helps guide speaker recognition sub-network Effective in using a limited amount of training data Uses less memory compared to i-vector approach (better for embeded device)

Hang Su Dissertation Talk 58 / 71

slide-60
SLIDE 60

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Table of contents

1

Introduction and Motivation

2

Backgrounds on Speech and Speaker Recognition

3

Connecting Speech and Speaker Recognition

4

Joint Modeling of Speech and Speaker

5

Conclusion and Future Work

Hang Su Dissertation Talk 59 / 71

slide-61
SLIDE 61

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Conclusion and Future Work

Conclusion of the talk Speech and speaker recognition are beneficial to each

  • ther

A joint model helps exploit both speech and speaker information Effective in using limited amount of training data

Hang Su Dissertation Talk 60 / 71

slide-62
SLIDE 62

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Future work

Future work on joint modeling Use a larger data set or data augmentation techniques Introduce recurrent structures into joint model End-to-end approaches for joint modeling Towards an all-around speech AI agent

Hang Su Dissertation Talk 61 / 71

slide-63
SLIDE 63

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work Hang Su Dissertation Talk 62 / 71

slide-64
SLIDE 64

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Reference I

Herve A. Bourlard and Nelson Morgan. Connectionist Speech Recognition: A Hybrid Approach. Kluwer Academic Publishers, Norwell, MA, USA, 1993. Najim Dehak, Patrick J Kenny, Réda Dehak, Pierre Dumouchel, and Pierre Ouellet. Front-end factor analysis for speaker verification. IEEE Transactions on Audio, Speech, and Language Processing, 19(4):788–798, 2011.

Hang Su Dissertation Talk 63 / 71

slide-65
SLIDE 65

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Reference II

Mark Gales and Steve Young. The application of hidden markov models in speech recognition. Foundations and Trends R in Signal Processing, 1(3):195–304, 2008. Mark JF Gales. Maximum likelihood linear transformations for hmm-based speech recognition. Computer speech & language, 12(2):75–98, 1998.

Hang Su Dissertation Talk 64 / 71

slide-66
SLIDE 66

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Reference III

Daniel Garcia-Romero and Carol Y Espy-Wilson. Analysis of i-vector length normalization in speaker recognition systems. In Twelfth Annual Conference of the International Speech Communication Association, 2011. Patrick Kenny, Themos Stafylakis, Pierre Ouellet, Md Jahangir Alam, and Pierre Dumouchel. Plda for speaker verification with utterances of arbitrary duration.

Hang Su Dissertation Talk 65 / 71

slide-67
SLIDE 67

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Reference IV

In Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on, pages 7649–7653. IEEE, 2013. Yun Lei, Scheffer Nicolas, Luciana Ferrer, and Mitchell McLaren. A novel scheme for speaker recognition using a phonetically-aware deep neural network. In ICASSP. IEEE, 2014.

Hang Su Dissertation Talk 66 / 71

slide-68
SLIDE 68

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Reference V

Abdel-rahman Mohamed, Frank Seide, Dong Yu, Jasha Droppo, Andreas Stoicke, Geoffrey Zweig, and Gerald Penn. Deep bi-directional recurrent networks over spectral windows. In Automatic Speech Recognition and Understanding (ASRU), 2015 IEEE Workshop on, pages 78–83. IEEE, 2015. Lawrence R Rabiner. A tutorial on hidden markov models and selected applications in speech recognition. Proceedings of the IEEE, 77(2):257–286, 1989.

Hang Su Dissertation Talk 67 / 71

slide-69
SLIDE 69

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Reference VI

George Saon, Hagen Soltau, David Nahamoo, and Michael Picheny. Speaker adaptation of neural network acoustic models using i-vectors. In ASRU, pages 55–59, 2013. Frank Seide, Gang Li, and Dong Yu. Conversational speech transcription using context-dependent deep neural networks. In Twelfth Annual Conference of the International Speech Communication Association, 2011.

Hang Su Dissertation Talk 68 / 71

slide-70
SLIDE 70

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Reference VII

Andrew Senior and Ignacio Lopez-Moreno. Improving dnn speaker independence with i-vector inputs. In Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on, pages 225–229. IEEE, 2014. David Snyder, Daniel Garcia-Romero, and Daniel Povey. Time delay deep neural network-based universal background models for speaker recognition. In ASRU. IEEE, 2015.

Hang Su Dissertation Talk 69 / 71

slide-71
SLIDE 71

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Reference VIII

Pawel Swietojanski and Steve Renals. Learning hidden unit contributions for unsupervised speaker adaptation of neural network acoustic models. In Spoken Language Technology Workshop (SLT), 2014 IEEE, pages 171–176. IEEE, 2014. Kaisheng Yao, Dong Yu, Frank Seide, Hang Su, Li Deng, and Yifan Gong. Adaptation of context-dependent deep neural networks for automatic speech recognition. In Spoken Language Technology Workshop (SLT), 2012 IEEE, pages 366–369. IEEE, 2012.

Hang Su Dissertation Talk 70 / 71

slide-72
SLIDE 72

Introduction and Motivation Backgrounds on Speech and Speaker Recognition Connecting Speech and Speaker Recognition Joint Modeling of Speech and Speaker Conclusion and Future Work

Reference IX

Dong Yu, Kaisheng Yao, Hang Su, Gang Li, and Frank Seide. Kl-divergence regularized deep neural network adaptation for improved large vocabulary speech recognition. In Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on, pages 7893–7897. IEEE, 2013.

Hang Su Dissertation Talk 71 / 71