{ Daniel Wilkey John Graham CS6998 Given speech, was the speaker - - PowerPoint PPT Presentation

daniel wilkey john graham cs6998 given speech was the
SMART_READER_LITE
LIVE PREVIEW

{ Daniel Wilkey John Graham CS6998 Given speech, was the speaker - - PowerPoint PPT Presentation

Detecting Intoxicated Speech { Daniel Wilkey John Graham CS6998 Given speech, was the speaker intoxicated? Interspeech 2011 Intoxication Challenge Application for field sobriety testing, ignition-guards Background ALC


slide-1
SLIDE 1

{

Detecting Intoxicated Speech

Daniel Wilkey John Graham CS6998

slide-2
SLIDE 2

 Given speech, was the speaker

intoxicated?

 Interspeech 2011 Intoxication Challenge  Application for field sobriety testing,

ignition-guards

Background

slide-3
SLIDE 3

 ALC – Alcohol Language Corpus  162 total participants: 84 male, 78 female  Participants reached a BAC .28 – 1.75  Read 15 minutes of intoxicated speech  Returned 2 weeks later  Read 30 minutes of sober speech

The Corpus

slide-4
SLIDE 4

 5400 samples in total, 75 per person  Divided into 3 sets:

 Development, Training, Test

 Development & Training are labeled with

4368 features

 Used cross validation to obtain results

The Corpus p2

slide-5
SLIDE 5

 Shrikanth Narayanan of UCLA

 Global speaker normalization  Normalizing by the sober class  Relative improvement of 7.04% overall

 Professor Hirchberg

 Phonotactic and phonetic cues

 Experiment tests un-weighted average recall… why?

 We chose f-measure  Includes recall and precision

Prior Research

slide-6
SLIDE 6

 Remove extraneous

features with WEKA

 Info-gain ratio algorithm  MFCC features

performed well

 No F0-based features

near the top

Experiment Preparation

slide-7
SLIDE 7

 Ignore test set

 unlabeled

 Down-sampling the training set

 Achieved 50/50 ratio of alcoholised to non-

alcoholised speech

Experiment Preparation

slide-8
SLIDE 8

 Global Speaker Normalization (Narayanan)

 Insignificant negative change

 Sober class normalization (Narayanan)

 Insignificant negative change

 Gender class normalization

 Insignificant positive change

 Combining global speaker with gender

normalization

 10.75% relative improvement in f-measure

 Poor performance potentially related to some F0

features being filtered out

Normalization Attempts

slide-9
SLIDE 9

 Tried retesting data with fringe cases omitted

 Fringe case BAC between .08% and .16%

proposed by Batliner

 We tried .02% to .08%

 Difference in data set and threshold

 Relative decrease of F-measure by 3.25%

On the Fringe

slide-10
SLIDE 10

Machine Learning Optimizations

slide-11
SLIDE 11

Optimizing the SVM

 Varied polynomial kernels  Radial basis function (RBF)

slide-12
SLIDE 12

 Varying number

 Folds  Iterations

Optimization Techniques

slide-13
SLIDE 13

 Configuration

 SVM kernel n=3  10-fold cross validation  Gender normaliation  Sober class normalization

Final Results

Difficult to compare!!

slide-14
SLIDE 14

 Difficult to compare results  Need better corpus  Extend with GMM super-vectors

Conclusions / Extensions