On the Correlation between Perceptual and Contextual Aspects of - - PowerPoint PPT Presentation

on the correlation between perceptual and contextual
SMART_READER_LITE
LIVE PREVIEW

On the Correlation between Perceptual and Contextual Aspects of - - PowerPoint PPT Presentation

Introduction Data Analysis Conclusions On the Correlation between Perceptual and Contextual Aspects of Laughter in Meetings Kornel Laskowski & Susanne Burger interACT, Carnegie Mellon University August 9, 2007 Kornel Laskowski &


slide-1
SLIDE 1

Introduction Data Analysis Conclusions

On the Correlation between Perceptual and Contextual Aspects of Laughter in Meetings

Kornel Laskowski & Susanne Burger

interACT, Carnegie Mellon University

August 9, 2007

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-2
SLIDE 2

Introduction Data Analysis Conclusions

Introduction

what we do:

data-driven, language-/text- independent modeling of multi-participant conversation for automatic conversation recognition and understanding

why?

who has the floor when? how many floors are there? who backchannels when? and towards whom? who interrupts who? who asks questions? who gives answers? how formal is the conversation? what is the social hierarchy of the participants? how do participants appear to feel?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-3
SLIDE 3

Introduction Data Analysis Conclusions

Introduction

what we do:

data-driven, language-/text- independent modeling of multi-participant conversation for automatic conversation recognition and understanding

why?

who has the floor when? how many floors are there? who backchannels when? and towards whom? who interrupts who? who asks questions? who gives answers? how formal is the conversation? what is the social hierarchy of the participants? how do participants appear to feel?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-4
SLIDE 4

Introduction Data Analysis Conclusions

Introduction

what we do:

data-driven, language-/text- independent modeling of multi-participant conversation for automatic conversation recognition and understanding

why?

who has the floor when? how many floors are there? who backchannels when? and towards whom? who interrupts who? who asks questions? who gives answers? how formal is the conversation? what is the social hierarchy of the participants? how do participants appear to feel?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-5
SLIDE 5

Introduction Data Analysis Conclusions

Introduction

what we do:

data-driven, language-/text- independent modeling of multi-participant conversation for automatic conversation recognition and understanding

why?

who has the floor when? how many floors are there? who backchannels when? and towards whom? who interrupts who? who asks questions? who gives answers? how formal is the conversation? what is the social hierarchy of the participants? how do participants appear to feel?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-6
SLIDE 6

Introduction Data Analysis Conclusions

Introduction

what we do:

data-driven, language-/text- independent modeling of multi-participant conversation for automatic conversation recognition and understanding

why?

who has the floor when? how many floors are there? who backchannels when? and towards whom? who interrupts who? who asks questions? who gives answers? how formal is the conversation? what is the social hierarchy of the participants? how do participants appear to feel?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-7
SLIDE 7

Introduction Data Analysis Conclusions

Introduction

what we do:

data-driven, language-/text- independent modeling of multi-participant conversation for automatic conversation recognition and understanding

why?

who has the floor when? how many floors are there? who backchannels when? and towards whom? who interrupts who? who asks questions? who gives answers? how formal is the conversation? what is the social hierarchy of the participants? how do participants appear to feel?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-8
SLIDE 8

Introduction Data Analysis Conclusions

Introduction

what we do:

data-driven, language-/text- independent modeling of multi-participant conversation for automatic conversation recognition and understanding

why?

who has the floor when? how many floors are there? who backchannels when? and towards whom? who interrupts who? who asks questions? who gives answers? how formal is the conversation? what is the social hierarchy of the participants? how do participants appear to feel?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-9
SLIDE 9

Introduction Data Analysis Conclusions

Introduction

what we do:

data-driven, language-/text- independent modeling of multi-participant conversation for automatic conversation recognition and understanding

why?

who has the floor when? how many floors are there? who backchannels when? and towards whom? who interrupts who? who asks questions? who gives answers? how formal is the conversation? what is the social hierarchy of the participants? how do participants appear to feel?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-10
SLIDE 10

Introduction Data Analysis Conclusions

A Text-Independent Representation of Multi-Participant Conversation

essentially monologue “multi-logue” heated “multi-logue” a mathematical artifact (the Haar wavelet basis) “multi-logue” with laughter

participants tend to wait their turn to speak participants do not wait to laugh

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-11
SLIDE 11

Introduction Data Analysis Conclusions

A Text-Independent Representation of Multi-Participant Conversation

essentially monologue “multi-logue” heated “multi-logue” a mathematical artifact (the Haar wavelet basis) “multi-logue” with laughter

participants tend to wait their turn to speak participants do not wait to laugh

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-12
SLIDE 12

Introduction Data Analysis Conclusions

A Text-Independent Representation of Multi-Participant Conversation

essentially monologue “multi-logue” heated “multi-logue” a mathematical artifact (the Haar wavelet basis) “multi-logue” with laughter

participants tend to wait their turn to speak participants do not wait to laugh

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-13
SLIDE 13

Introduction Data Analysis Conclusions

A Text-Independent Representation of Multi-Participant Conversation

essentially monologue “multi-logue” heated “multi-logue” a mathematical artifact (the Haar wavelet basis) “multi-logue” with laughter

participants tend to wait their turn to speak participants do not wait to laugh

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-14
SLIDE 14

Introduction Data Analysis Conclusions

A Text-Independent Representation of Multi-Participant Conversation

essentially monologue “multi-logue” heated “multi-logue” a mathematical artifact (the Haar wavelet basis) “multi-logue” with laughter

participants tend to wait their turn to speak participants do not wait to laugh

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-15
SLIDE 15

Introduction Data Analysis Conclusions

A Text-Independent Representation of Multi-Participant Conversation

essentially monologue “multi-logue” heated “multi-logue” a mathematical artifact (the Haar wavelet basis) “multi-logue” with laughter

participants tend to wait their turn to speak participants do not wait to laugh

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-16
SLIDE 16

Introduction Data Analysis Conclusions

Emotion and Laughter in Conversation

external observers of conversation appear to agree as to whether participants feel

neutral: 82% of utterances positive: 16% of utterances negative: 2% of utterances

transcribed laughter is strongly predictive of positive valence (92% classification accuracy) A FUTURE GOAL: to find laughter in continuous audio

acoustic features context states

context does discriminate between speech and laughter does context discriminate between voiced and unvoiced laughter?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-17
SLIDE 17

Introduction Data Analysis Conclusions

Emotion and Laughter in Conversation

external observers of conversation appear to agree as to whether participants feel

neutral: 82% of utterances positive: 16% of utterances negative: 2% of utterances

transcribed laughter is strongly predictive of positive valence (92% classification accuracy) A FUTURE GOAL: to find laughter in continuous audio

acoustic features context states

context does discriminate between speech and laughter does context discriminate between voiced and unvoiced laughter?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-18
SLIDE 18

Introduction Data Analysis Conclusions

Emotion and Laughter in Conversation

external observers of conversation appear to agree as to whether participants feel

neutral: 82% of utterances positive: 16% of utterances negative: 2% of utterances

transcribed laughter is strongly predictive of positive valence (92% classification accuracy) A FUTURE GOAL: to find laughter in continuous audio

acoustic features context states

context does discriminate between speech and laughter does context discriminate between voiced and unvoiced laughter?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-19
SLIDE 19

Introduction Data Analysis Conclusions

Emotion and Laughter in Conversation

external observers of conversation appear to agree as to whether participants feel

neutral: 82% of utterances positive: 16% of utterances negative: 2% of utterances

transcribed laughter is strongly predictive of positive valence (92% classification accuracy) A FUTURE GOAL: to find laughter in continuous audio

acoustic features context states

context does discriminate between speech and laughter does context discriminate between voiced and unvoiced laughter?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-20
SLIDE 20

Introduction Data Analysis Conclusions

Emotion and Laughter in Conversation

external observers of conversation appear to agree as to whether participants feel

neutral: 82% of utterances positive: 16% of utterances negative: 2% of utterances

transcribed laughter is strongly predictive of positive valence (92% classification accuracy) A FUTURE GOAL: to find laughter in continuous audio

acoustic features context states

context does discriminate between speech and laughter does context discriminate between voiced and unvoiced laughter?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-21
SLIDE 21

Introduction Data Analysis Conclusions

Emotion and Laughter in Conversation

external observers of conversation appear to agree as to whether participants feel

neutral: 82% of utterances positive: 16% of utterances negative: 2% of utterances

transcribed laughter is strongly predictive of positive valence (92% classification accuracy) A FUTURE GOAL: to find laughter in continuous audio

acoustic features context states

context does discriminate between speech and laughter does context discriminate between voiced and unvoiced laughter?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-22
SLIDE 22

Introduction Data Analysis Conclusions

The ICSI Meeting Corpus

naturally occurring project-oriented conversations for our purposes, 4 types of meetings: # of # of possible # of participants type meetings participants mod min max Bed 15 13 6 4 7 Bmr 29 15 7 3 9 Bro 23 10 6 4 8

  • ther

8 27 6 5 8 “other” contains types of which there are ≤3 meetings types represent longitudinal recordings rarely, meetings contain additional, uninstrumented participants

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-23
SLIDE 23

Introduction Data Analysis Conclusions

The ICSI Meeting Corpus

naturally occurring project-oriented conversations for our purposes, 4 types of meetings: # of # of possible # of participants type meetings participants mod min max Bed 15 13 6 4 7 Bmr 29 15 7 3 9 Bro 23 10 6 4 8

  • ther

8 27 6 5 8 “other” contains types of which there are ≤3 meetings types represent longitudinal recordings rarely, meetings contain additional, uninstrumented participants

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-24
SLIDE 24

Introduction Data Analysis Conclusions

The ICSI Meeting Corpus

naturally occurring project-oriented conversations for our purposes, 4 types of meetings: # of # of possible # of participants type meetings participants mod min max Bed 15 13 6 4 7 Bmr 29 15 7 3 9 Bro 23 10 6 4 8

  • ther

8 27 6 5 8 “other” contains types of which there are ≤3 meetings types represent longitudinal recordings rarely, meetings contain additional, uninstrumented participants

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-25
SLIDE 25

Introduction Data Analysis Conclusions

The ICSI Meeting Corpus

naturally occurring project-oriented conversations for our purposes, 4 types of meetings: # of # of possible # of participants type meetings participants mod min max Bed 15 13 6 4 7 Bmr 29 15 7 3 9 Bro 23 10 6 4 8

  • ther

8 27 6 5 8 “other” contains types of which there are ≤3 meetings types represent longitudinal recordings rarely, meetings contain additional, uninstrumented participants

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-26
SLIDE 26

Introduction Data Analysis Conclusions

The ICSI Meeting Corpus

naturally occurring project-oriented conversations for our purposes, 4 types of meetings: # of # of possible # of participants type meetings participants mod min max Bed 15 13 6 4 7 Bmr 29 15 7 3 9 Bro 23 10 6 4 8

  • ther

8 27 6 5 8 “other” contains types of which there are ≤3 meetings types represent longitudinal recordings rarely, meetings contain additional, uninstrumented participants

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-27
SLIDE 27

Introduction Data Analysis Conclusions

The ICSI Meeting Corpus: Amount of Audio

distribution of usable meeting durations over the 75 meetings:

0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 2 4 6 8 10 12 14 16 duration, hours number of meetings

a total of 66.3 hours of conversation the average participant vocalizes for 14.8% of the time

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-28
SLIDE 28

Introduction Data Analysis Conclusions

The ICSI Meeting Corpus: Amount of Audio

distribution of usable meeting durations over the 75 meetings:

0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 2 4 6 8 10 12 14 16 duration, hours number of meetings

a total of 66.3 hours of conversation the average participant vocalizes for 14.8% of the time

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-29
SLIDE 29

Introduction Data Analysis Conclusions

The ICSI Meeting Corpus: Amount of Audio

distribution of usable meeting durations over the 75 meetings:

0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 2 4 6 8 10 12 14 16 duration, hours number of meetings

a total of 66.3 hours of conversation the average participant vocalizes for 14.8% of the time

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-30
SLIDE 30

Introduction Data Analysis Conclusions

Laughter Annotation

the ICSI corpus (audio) is accompanied by orthographic transcription, which includes a relatively rich XML-style mark-up of laughter for our purposes, data preprocessing consisted of:

1

identifying laughter in the orthographic transcription

2

segmentation: specifying endpoints for identified laughter

3

classification: specifying voicing for segmented laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-31
SLIDE 31

Introduction Data Analysis Conclusions

Laughter Annotation

the ICSI corpus (audio) is accompanied by orthographic transcription, which includes a relatively rich XML-style mark-up of laughter for our purposes, data preprocessing consisted of:

1

identifying laughter in the orthographic transcription

2

segmentation: specifying endpoints for identified laughter

3

classification: specifying voicing for segmented laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-32
SLIDE 32

Introduction Data Analysis Conclusions

Laughter Annotation

the ICSI corpus (audio) is accompanied by orthographic transcription, which includes a relatively rich XML-style mark-up of laughter for our purposes, data preprocessing consisted of:

1

identifying laughter in the orthographic transcription

2

segmentation: specifying endpoints for identified laughter

3

classification: specifying voicing for segmented laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-33
SLIDE 33

Introduction Data Analysis Conclusions

Laughter Annotation

the ICSI corpus (audio) is accompanied by orthographic transcription, which includes a relatively rich XML-style mark-up of laughter for our purposes, data preprocessing consisted of:

1

identifying laughter in the orthographic transcription

2

segmentation: specifying endpoints for identified laughter

3

classification: specifying voicing for segmented laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-34
SLIDE 34

Introduction Data Analysis Conclusions

Identifying Laughter in the ICSI Corpus

  • rthographic, time-segmented transcription of speaker

contributions (.stm)

Bmr011 me013 chan1 3029.466 3029.911 Yeah. Bmr011 mn005 chan3 3030.230 3031.140 Film-maker. Bmr011 fe016 chan0 3030.783 3032.125 <Emphasis> colorful. </Emphasi... Bmr011 me011 chanB 3035.301 3036.964 Of beeps, yeah. Bmr011 fe008 chan8 3035.714 3037.314 <Pause/> of m- one hour of - <... Bmr011 mn014 chan2 3036.030 3036.640 Yeah. Bmr011 me013 chan1 3036.280 3037.600 <VocalSound Description="laugh"/> Bmr011 mn014 chan2 3036.640 3037.115 Yeah. Bmr011 mn005 chan3 3036.930 3037.335 Is - Bmr011 me011 chanB 3036.964 3038.573 <VocalSound Description="laugh"/>

laughter is identified using VocalSound and Comment tags

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-35
SLIDE 35

Introduction Data Analysis Conclusions

Identifying Laughter in the ICSI Corpus

  • rthographic, time-segmented transcription of speaker

contributions (.stm)

...9.911 Yeah. ...1.140 Film-maker. ...2.125 <Emphasis> colorful. </Emphasis> <Comment Description="while laughing"/> ...6.964 Of beeps, yeah. ...7.314 <Pause/> of m- one hour of - <Comment Description="while laughing"/> ...6.640 Yeah. ...7.600 <VocalSound Description="laugh"/> ...7.115 Yeah. ...7.335 Is - ...8.573 <VocalSound Description="laugh"/>

laughter is identified using VocalSound and Comment tags

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-36
SLIDE 36

Introduction Data Analysis Conclusions

Identifying Laughter in the ICSI Corpus

  • rthographic, time-segmented transcription of speaker

contributions (.stm)

...9.911 Yeah. ...1.140 Film-maker. ...2.125 <Emphasis> colorful. </Emphasis> <Comment Description="while laughing"/> ...6.964 Of beeps, yeah. ...7.314 <Pause/> of m- one hour of - <Comment Description="while laughing"/> ...6.640 Yeah. ...7.600 <VocalSound Description="laugh"/> ...7.115 Yeah. ...7.335 Is - ...8.573 <VocalSound Description="laugh"/>

laughter is identified using VocalSound and Comment tags

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-37
SLIDE 37

Introduction Data Analysis Conclusions

Identifying Laughter in the ICSI Corpus

  • rthographic, time-segmented transcription of speaker

contributions (.stm)

...9.911 Yeah. ...1.140 Film-maker. ...2.125 <Emphasis> colorful. </Emphasis> <Comment Description="while laughing"/> ...6.964 Of beeps, yeah. ...7.314 <Pause/> of m- one hour of - <Comment Description="while laughing"/> ...6.640 Yeah. ...7.600 <VocalSound Description="laugh"/> ...7.115 Yeah. ...7.335 Is - ...8.573 <VocalSound Description="laugh"/>

laughter is identified using VocalSound and Comment tags

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-38
SLIDE 38

Introduction Data Analysis Conclusions

Identifying Laughter in the ICSI Corpus

  • rthographic, time-segmented transcription of speaker

contributions (.stm)

...9.911 Yeah. ...1.140 Film-maker. ...2.125 <Emphasis> colorful. </Emphasis> <Comment Description="while laughing"/> ...6.964 Of beeps, yeah. ...7.314 <Pause/> of m- one hour of - <Comment Description="while laughing"/> ...6.640 Yeah. ...7.600 <VocalSound Description="laugh"/> ...7.115 Yeah. ...7.335 Is - ...8.573 <VocalSound Description="laugh"/>

laughter is identified using VocalSound and Comment tags

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-39
SLIDE 39

Introduction Data Analysis Conclusions

Sample VocalSound Instances

Freq Token Rank Count VocalSound Description Used 1 11515 laugh √ 2 7091 breath 3 4589 inbreath 4 2223 mouth 5 970 breath-laugh √ 11 97 laugh-breath √ 46 6 cough-laugh √ 63 3 laugh, "hmmph" √ 69 3 breath while smiling 75 2 very long laugh √

laughter is by far the most common non-verbal vocal sound annotated in this corpus

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-40
SLIDE 40

Introduction Data Analysis Conclusions

Sample VocalSound Instances

Freq Token Rank Count VocalSound Description Used 1 11515 laugh √ 2 7091 breath 3 4589 inbreath 4 2223 mouth 5 970 breath-laugh √ 11 97 laugh-breath √ 46 6 cough-laugh √ 63 3 laugh, "hmmph" √ 69 3 breath while smiling 75 2 very long laugh √

laughter is by far the most common non-verbal vocal sound annotated in this corpus

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-41
SLIDE 41

Introduction Data Analysis Conclusions

Sample Comment Instances

Freq Token Rank Count Comment Description 2 980 while laughing 16 59 while smiling 44 13 last two words while laughing 125 4 last word while laughing 145 3 vocal gesture, a mock laugh

the most frequent Comment is not related to conversation therefore, while laughing is the most frequent conversation-related Comment description Comment tags have an even richer description set than VocalSound tags

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-42
SLIDE 42

Introduction Data Analysis Conclusions

Sample Comment Instances

Freq Token Rank Count Comment Description 2 980 while laughing 16 59 while smiling 44 13 last two words while laughing 125 4 last word while laughing 145 3 vocal gesture, a mock laugh

the most frequent Comment is not related to conversation therefore, while laughing is the most frequent conversation-related Comment description Comment tags have an even richer description set than VocalSound tags

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-43
SLIDE 43

Introduction Data Analysis Conclusions

Segmenting Identified Laughter Instances

found 12570 non-farfield VocalSound instances

11845 were adjacent to a time-stamped utterance boundary or lexical item: endpoints were derived automatically 725 needed to be segmented manually

found 1108 non-farfield Comment instances

all needed to be segmented manually

manual segmententation performed by me, checked by at least

  • ne other annotator

merging immediately adjacent VocalSound and Comment instances, and removing transcribed instances for which we found counterevidence, resulted in 13259 segmented bouts of laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-44
SLIDE 44

Introduction Data Analysis Conclusions

Segmenting Identified Laughter Instances

found 12570 non-farfield VocalSound instances

11845 were adjacent to a time-stamped utterance boundary or lexical item: endpoints were derived automatically 725 needed to be segmented manually

found 1108 non-farfield Comment instances

all needed to be segmented manually

manual segmententation performed by me, checked by at least

  • ne other annotator

merging immediately adjacent VocalSound and Comment instances, and removing transcribed instances for which we found counterevidence, resulted in 13259 segmented bouts of laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-45
SLIDE 45

Introduction Data Analysis Conclusions

Segmenting Identified Laughter Instances

found 12570 non-farfield VocalSound instances

11845 were adjacent to a time-stamped utterance boundary or lexical item: endpoints were derived automatically 725 needed to be segmented manually

found 1108 non-farfield Comment instances

all needed to be segmented manually

manual segmententation performed by me, checked by at least

  • ne other annotator

merging immediately adjacent VocalSound and Comment instances, and removing transcribed instances for which we found counterevidence, resulted in 13259 segmented bouts of laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-46
SLIDE 46

Introduction Data Analysis Conclusions

Segmenting Identified Laughter Instances

found 12570 non-farfield VocalSound instances

11845 were adjacent to a time-stamped utterance boundary or lexical item: endpoints were derived automatically 725 needed to be segmented manually

found 1108 non-farfield Comment instances

all needed to be segmented manually

manual segmententation performed by me, checked by at least

  • ne other annotator

merging immediately adjacent VocalSound and Comment instances, and removing transcribed instances for which we found counterevidence, resulted in 13259 segmented bouts of laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-47
SLIDE 47

Introduction Data Analysis Conclusions

Segmenting Identified Laughter Instances

found 12570 non-farfield VocalSound instances

11845 were adjacent to a time-stamped utterance boundary or lexical item: endpoints were derived automatically 725 needed to be segmented manually

found 1108 non-farfield Comment instances

all needed to be segmented manually

manual segmententation performed by me, checked by at least

  • ne other annotator

merging immediately adjacent VocalSound and Comment instances, and removing transcribed instances for which we found counterevidence, resulted in 13259 segmented bouts of laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-48
SLIDE 48

Introduction Data Analysis Conclusions

Classifying Voicing of the Segmented Laughter Bouts

if any portion of the bout is voiced, the bout is voiced performed manually for all 13259 bouts by at least one annotator interlabeler kappa was 0.76-0.79 (we considered this low) all instances rechecked by Susi

not modified: 11961 bouts (90.2%) modified voicing: 942 bouts (7.1%) modified endpoints: 306 bouts (2.3%) removed: 50 bouts (0.4%)

total left: 13209 bouts

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-49
SLIDE 49

Introduction Data Analysis Conclusions

Classifying Voicing of the Segmented Laughter Bouts

if any portion of the bout is voiced, the bout is voiced performed manually for all 13259 bouts by at least one annotator interlabeler kappa was 0.76-0.79 (we considered this low) all instances rechecked by Susi

not modified: 11961 bouts (90.2%) modified voicing: 942 bouts (7.1%) modified endpoints: 306 bouts (2.3%) removed: 50 bouts (0.4%)

total left: 13209 bouts

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-50
SLIDE 50

Introduction Data Analysis Conclusions

Classifying Voicing of the Segmented Laughter Bouts

if any portion of the bout is voiced, the bout is voiced performed manually for all 13259 bouts by at least one annotator interlabeler kappa was 0.76-0.79 (we considered this low) all instances rechecked by Susi

not modified: 11961 bouts (90.2%) modified voicing: 942 bouts (7.1%) modified endpoints: 306 bouts (2.3%) removed: 50 bouts (0.4%)

total left: 13209 bouts

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-51
SLIDE 51

Introduction Data Analysis Conclusions

Classifying Voicing of the Segmented Laughter Bouts

if any portion of the bout is voiced, the bout is voiced performed manually for all 13259 bouts by at least one annotator interlabeler kappa was 0.76-0.79 (we considered this low) all instances rechecked by Susi

not modified: 11961 bouts (90.2%) modified voicing: 942 bouts (7.1%) modified endpoints: 306 bouts (2.3%) removed: 50 bouts (0.4%)

total left: 13209 bouts

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-52
SLIDE 52

Introduction Data Analysis Conclusions

Classifying Voicing of the Segmented Laughter Bouts

if any portion of the bout is voiced, the bout is voiced performed manually for all 13259 bouts by at least one annotator interlabeler kappa was 0.76-0.79 (we considered this low) all instances rechecked by Susi

not modified: 11961 bouts (90.2%) modified voicing: 942 bouts (7.1%) modified endpoints: 306 bouts (2.3%) removed: 50 bouts (0.4%)

total left: 13209 bouts

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-53
SLIDE 53

Introduction Data Analysis Conclusions

Classifying Voicing of the Segmented Laughter Bouts

if any portion of the bout is voiced, the bout is voiced performed manually for all 13259 bouts by at least one annotator interlabeler kappa was 0.76-0.79 (we considered this low) all instances rechecked by Susi

not modified: 11961 bouts (90.2%) modified voicing: 942 bouts (7.1%) modified endpoints: 306 bouts (2.3%) removed: 50 bouts (0.4%)

total left: 13209 bouts

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-54
SLIDE 54

Introduction Data Analysis Conclusions

Classifying Voicing of the Segmented Laughter Bouts

if any portion of the bout is voiced, the bout is voiced performed manually for all 13259 bouts by at least one annotator interlabeler kappa was 0.76-0.79 (we considered this low) all instances rechecked by Susi

not modified: 11961 bouts (90.2%) modified voicing: 942 bouts (7.1%) modified endpoints: 306 bouts (2.3%) removed: 50 bouts (0.4%)

total left: 13209 bouts

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-55
SLIDE 55

Introduction Data Analysis Conclusions

Classifying Voicing of the Segmented Laughter Bouts

if any portion of the bout is voiced, the bout is voiced performed manually for all 13259 bouts by at least one annotator interlabeler kappa was 0.76-0.79 (we considered this low) all instances rechecked by Susi

not modified: 11961 bouts (90.2%) modified voicing: 942 bouts (7.1%) modified endpoints: 306 bouts (2.3%) removed: 50 bouts (0.4%)

total left: 13209 bouts

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-56
SLIDE 56

Introduction Data Analysis Conclusions

Classifying Voicing of the Segmented Laughter Bouts

if any portion of the bout is voiced, the bout is voiced performed manually for all 13259 bouts by at least one annotator interlabeler kappa was 0.76-0.79 (we considered this low) all instances rechecked by Susi

not modified: 11961 bouts (90.2%) modified voicing: 942 bouts (7.1%) modified endpoints: 306 bouts (2.3%) removed: 50 bouts (0.4%)

total left: 13209 bouts

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-57
SLIDE 57

Introduction Data Analysis Conclusions

Voiced vs Unvoiced Laughter by Time

  • f 13209 bouts of laughter,

voiced: 8687 (65.8%) unvoiced: 4426 (33.5%) laughed speech: 96 (0.7%)

  • f 5.7 hours of laughter

voiced: 4.2 hours (73.7%) unvoiced: 1.5 hours (25.8%) laughed speech: <0.1 hours (0.5%)

since there is so little laughed speech, we ignore it in this work

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-58
SLIDE 58

Introduction Data Analysis Conclusions

Voiced vs Unvoiced Laughter by Time

  • f 13209 bouts of laughter,

voiced: 8687 (65.8%) unvoiced: 4426 (33.5%) laughed speech: 96 (0.7%)

  • f 5.7 hours of laughter

voiced: 4.2 hours (73.7%) unvoiced: 1.5 hours (25.8%) laughed speech: <0.1 hours (0.5%)

since there is so little laughed speech, we ignore it in this work

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-59
SLIDE 59

Introduction Data Analysis Conclusions

Voiced vs Unvoiced Laughter by Time

  • f 13209 bouts of laughter,

voiced: 8687 (65.8%) unvoiced: 4426 (33.5%) laughed speech: 96 (0.7%)

  • f 5.7 hours of laughter

voiced: 4.2 hours (73.7%) unvoiced: 1.5 hours (25.8%) laughed speech: <0.1 hours (0.5%)

since there is so little laughed speech, we ignore it in this work

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-60
SLIDE 60

Introduction Data Analysis Conclusions

Voiced vs Unvoiced Laughter by Time, by Participant

5 10 15 20 25 30 35 40 45 50 1 2 3 4 5 6 voiced laughter unvoiced laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-61
SLIDE 61

Introduction Data Analysis Conclusions

Voiced vs Unvoiced Bout Duration

0.1 0.2 0.5 1 2 5 0.05 0.1 0.15 voi laugh bouts unv laugh bouts talk spurts Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-62
SLIDE 62

Introduction Data Analysis Conclusions

Analysis of Laughter-in-Interaction

GOAL: characterize the correlation between voicing in laughter and the vocal interaction context in which laughter occurs

test for the statistical significance of association test for the strength of association (predictability)

1 discretize (in time) the voiced laughter, unvoiced laughter,

and talkspurt segmentations

allows for counting

2 for each discrete laugh frame, extract a set of

multi-participant, participant-independent features from the discretized context

3 characterize the association between context features and

voicing features

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-63
SLIDE 63

Introduction Data Analysis Conclusions

Analysis of Laughter-in-Interaction

GOAL: characterize the correlation between voicing in laughter and the vocal interaction context in which laughter occurs

test for the statistical significance of association test for the strength of association (predictability)

1 discretize (in time) the voiced laughter, unvoiced laughter,

and talkspurt segmentations

allows for counting

2 for each discrete laugh frame, extract a set of

multi-participant, participant-independent features from the discretized context

3 characterize the association between context features and

voicing features

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-64
SLIDE 64

Introduction Data Analysis Conclusions

Analysis of Laughter-in-Interaction

GOAL: characterize the correlation between voicing in laughter and the vocal interaction context in which laughter occurs

test for the statistical significance of association test for the strength of association (predictability)

1 discretize (in time) the voiced laughter, unvoiced laughter,

and talkspurt segmentations

allows for counting

2 for each discrete laugh frame, extract a set of

multi-participant, participant-independent features from the discretized context

3 characterize the association between context features and

voicing features

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-65
SLIDE 65

Introduction Data Analysis Conclusions

Analysis of Laughter-in-Interaction

GOAL: characterize the correlation between voicing in laughter and the vocal interaction context in which laughter occurs

test for the statistical significance of association test for the strength of association (predictability)

1 discretize (in time) the voiced laughter, unvoiced laughter,

and talkspurt segmentations

allows for counting

2 for each discrete laugh frame, extract a set of

multi-participant, participant-independent features from the discretized context

3 characterize the association between context features and

voicing features

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-66
SLIDE 66

Introduction Data Analysis Conclusions

Analysis of Laughter-in-Interaction

GOAL: characterize the correlation between voicing in laughter and the vocal interaction context in which laughter occurs

test for the statistical significance of association test for the strength of association (predictability)

1 discretize (in time) the voiced laughter, unvoiced laughter,

and talkspurt segmentations

allows for counting

2 for each discrete laugh frame, extract a set of

multi-participant, participant-independent features from the discretized context

3 characterize the association between context features and

voicing features

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-67
SLIDE 67

Introduction Data Analysis Conclusions

Analysis of Laughter-in-Interaction

GOAL: characterize the correlation between voicing in laughter and the vocal interaction context in which laughter occurs

test for the statistical significance of association test for the strength of association (predictability)

1 discretize (in time) the voiced laughter, unvoiced laughter,

and talkspurt segmentations

allows for counting

2 for each discrete laugh frame, extract a set of

multi-participant, participant-independent features from the discretized context

3 characterize the association between context features and

voicing features

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-68
SLIDE 68

Introduction Data Analysis Conclusions

Discretizing Segmentations

chop up each segmentation into non-overlapping 1 second frames for each participant k, declare a frame centered on time t as “on” when participant k vocalizes for at least 10% of that frame’s duration example:

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-69
SLIDE 69

Introduction Data Analysis Conclusions

Discretizing Segmentations

chop up each segmentation into non-overlapping 1 second frames for each participant k, declare a frame centered on time t as “on” when participant k vocalizes for at least 10% of that frame’s duration example:

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-70
SLIDE 70

Introduction Data Analysis Conclusions

Discretizing Segmentations

chop up each segmentation into non-overlapping 1 second frames for each participant k, declare a frame centered on time t as “on” when participant k vocalizes for at least 10% of that frame’s duration example:

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-71
SLIDE 71

Introduction Data Analysis Conclusions

Discretizing Segmentations

chop up each segmentation into non-overlapping 1 second frames for each participant k, declare a frame centered on time t as “on” when participant k vocalizes for at least 10% of that frame’s duration example:

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-72
SLIDE 72

Introduction Data Analysis Conclusions

Discretizing Segmentations

chop up each segmentation into non-overlapping 1 second frames for each participant k, declare a frame centered on time t as “on” when participant k vocalizes for at least 10% of that frame’s duration example:

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-73
SLIDE 73

Introduction Data Analysis Conclusions

Discretizing Segmentations

chop up each segmentation into non-overlapping 1 second frames for each participant k, declare a frame centered on time t as “on” when participant k vocalizes for at least 10% of that frame’s duration example:

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-74
SLIDE 74

Introduction Data Analysis Conclusions

Discretizing Segmentations

chop up each segmentation into non-overlapping 1 second frames for each participant k, declare a frame centered on time t as “on” when participant k vocalizes for at least 10% of that frame’s duration example:

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-75
SLIDE 75

Introduction Data Analysis Conclusions

Discretizing Segmentations

chop up each segmentation into non-overlapping 1 second frames for each participant k, declare a frame centered on time t as “on” when participant k vocalizes for at least 10% of that frame’s duration example:

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-76
SLIDE 76

Introduction Data Analysis Conclusions

Discretizing Segmentations

chop up each segmentation into non-overlapping 1 second frames for each participant k, declare a frame centered on time t as “on” when participant k vocalizes for at least 10% of that frame’s duration example:

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-77
SLIDE 77

Introduction Data Analysis Conclusions

Discretizing Segmentations

chop up each segmentation into non-overlapping 1 second frames for each participant k, declare a frame centered on time t as “on” when participant k vocalizes for at least 10% of that frame’s duration example:

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-78
SLIDE 78

Introduction Data Analysis Conclusions

Discretizing Segmentations

chop up each segmentation into non-overlapping 1 second frames for each participant k, declare a frame centered on time t as “on” when participant k vocalizes for at least 10% of that frame’s duration example:

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-79
SLIDE 79

Introduction Data Analysis Conclusions

Features Describing Conversational Context

for each frame t in which participant k laughs:

count how many other participants, at times t − 1, t, and t + 1, are producing a talk spurt count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which contains voicing count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which does not contain voicing determine whether participant k is speaking at times t − 1 and t + 1

in total, each frame of voiced or unvoiced laughter corresponds to a vocal interaction context defined by 11 features

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-80
SLIDE 80

Introduction Data Analysis Conclusions

Features Describing Conversational Context

for each frame t in which participant k laughs:

count how many other participants, at times t − 1, t, and t + 1, are producing a talk spurt count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which contains voicing count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which does not contain voicing determine whether participant k is speaking at times t − 1 and t + 1

in total, each frame of voiced or unvoiced laughter corresponds to a vocal interaction context defined by 11 features

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-81
SLIDE 81

Introduction Data Analysis Conclusions

Features Describing Conversational Context

for each frame t in which participant k laughs:

count how many other participants, at times t − 1, t, and t + 1, are producing a talk spurt count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which contains voicing count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which does not contain voicing determine whether participant k is speaking at times t − 1 and t + 1

in total, each frame of voiced or unvoiced laughter corresponds to a vocal interaction context defined by 11 features

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-82
SLIDE 82

Introduction Data Analysis Conclusions

Features Describing Conversational Context

for each frame t in which participant k laughs:

count how many other participants, at times t − 1, t, and t + 1, are producing a talk spurt count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which contains voicing count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which does not contain voicing determine whether participant k is speaking at times t − 1 and t + 1

in total, each frame of voiced or unvoiced laughter corresponds to a vocal interaction context defined by 11 features

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-83
SLIDE 83

Introduction Data Analysis Conclusions

Features Describing Conversational Context

for each frame t in which participant k laughs:

count how many other participants, at times t − 1, t, and t + 1, are producing a talk spurt count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which contains voicing count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which does not contain voicing determine whether participant k is speaking at times t − 1 and t + 1

in total, each frame of voiced or unvoiced laughter corresponds to a vocal interaction context defined by 11 features

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-84
SLIDE 84

Introduction Data Analysis Conclusions

Features Describing Conversational Context

for each frame t in which participant k laughs:

count how many other participants, at times t − 1, t, and t + 1, are producing a talk spurt count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which contains voicing count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which does not contain voicing determine whether participant k is speaking at times t − 1 and t + 1

in total, each frame of voiced or unvoiced laughter corresponds to a vocal interaction context defined by 11 features

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-85
SLIDE 85

Introduction Data Analysis Conclusions

Features Describing Conversational Context

for each frame t in which participant k laughs:

count how many other participants, at times t − 1, t, and t + 1, are producing a talk spurt count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which contains voicing count how many other participants, at times t − 1, t, and t + 1, are producing a laugh bout which does not contain voicing determine whether participant k is speaking at times t − 1 and t + 1

in total, each frame of voiced or unvoiced laughter corresponds to a vocal interaction context defined by 11 features

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-86
SLIDE 86

Introduction Data Analysis Conclusions

Summary of Context and Voicing Features

at this point, have:

context features z }| { # other participants in participant k in speech voiced laughter unvoiced laughter speech? Voicing? t − 1 t t + 1 t − 1 t t + 1 t − 1 t t + 1 t − 1 t + 1 1 1 1 1 2 N N Y 2 1 1 1 1 Y N Y 3 1 1 2 3 1 N Y N . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . | {z } integer features | {z } binary features

now, can proceed to analysis

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-87
SLIDE 87

Introduction Data Analysis Conclusions

Summary of Context and Voicing Features

at this point, have:

context features z }| { # other participants in participant k in speech voiced laughter unvoiced laughter speech? Voicing? t − 1 t t + 1 t − 1 t t + 1 t − 1 t t + 1 t − 1 t + 1 1 1 1 1 2 N N Y 2 1 1 1 1 Y N Y 3 1 1 2 3 1 N Y N . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . | {z } integer features | {z } binary features

now, can proceed to analysis

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-88
SLIDE 88

Introduction Data Analysis Conclusions

Summary of Context and Voicing Features

at this point, have:

context features z }| { # other participants in participant k in speech voiced laughter unvoiced laughter speech? Voicing? t − 1 t t + 1 t − 1 t t + 1 t − 1 t t + 1 t − 1 t + 1 1 1 1 1 2 N N Y 2 1 1 1 1 Y N Y 3 1 1 2 3 1 N Y N . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . | {z } integer features | {z } binary features

now, can proceed to analysis

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-89
SLIDE 89

Introduction Data Analysis Conclusions

Summary of Context and Voicing Features

at this point, have:

context features z }| { # other participants in participant k in speech voiced laughter unvoiced laughter speech? Voicing? t − 1 t t + 1 t − 1 t t + 1 t − 1 t t + 1 t − 1 t + 1 1 1 1 1 2 N N Y 2 1 1 1 1 Y N Y 3 1 1 2 3 1 N Y N . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . | {z } integer features | {z } binary features

now, can proceed to analysis

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-90
SLIDE 90

Introduction Data Analysis Conclusions

Testing Significance and Strength of Association

GOAL: correlate context features with the single voicing feature OPTION 1: standard, one-feature-at-a-time:

1

significance: a 2 × 2 χ2-test

2

strength: mutual information (or other entropy-related)

OPTION 2: optimal ordering of multiple-features-at-once:

1

strength: incremental, top-down mutual information

2

significance: bottom-up χ2-based pruning

latter is known as C4.5; developed for the inference of decision tree classifiers from data

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-91
SLIDE 91

Introduction Data Analysis Conclusions

Testing Significance and Strength of Association

GOAL: correlate context features with the single voicing feature OPTION 1: standard, one-feature-at-a-time:

1

significance: a 2 × 2 χ2-test

2

strength: mutual information (or other entropy-related)

OPTION 2: optimal ordering of multiple-features-at-once:

1

strength: incremental, top-down mutual information

2

significance: bottom-up χ2-based pruning

latter is known as C4.5; developed for the inference of decision tree classifiers from data

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-92
SLIDE 92

Introduction Data Analysis Conclusions

Testing Significance and Strength of Association

GOAL: correlate context features with the single voicing feature OPTION 1: standard, one-feature-at-a-time:

1

significance: a 2 × 2 χ2-test

2

strength: mutual information (or other entropy-related)

OPTION 2: optimal ordering of multiple-features-at-once:

1

strength: incremental, top-down mutual information

2

significance: bottom-up χ2-based pruning

latter is known as C4.5; developed for the inference of decision tree classifiers from data

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-93
SLIDE 93

Introduction Data Analysis Conclusions

Testing Significance and Strength of Association

GOAL: correlate context features with the single voicing feature OPTION 1: standard, one-feature-at-a-time:

1

significance: a 2 × 2 χ2-test

2

strength: mutual information (or other entropy-related)

OPTION 2: optimal ordering of multiple-features-at-once:

1

strength: incremental, top-down mutual information

2

significance: bottom-up χ2-based pruning

latter is known as C4.5; developed for the inference of decision tree classifiers from data

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-94
SLIDE 94

Introduction Data Analysis Conclusions

Testing Significance and Strength of Association

GOAL: correlate context features with the single voicing feature OPTION 1: standard, one-feature-at-a-time:

1

significance: a 2 × 2 χ2-test

2

strength: mutual information (or other entropy-related)

OPTION 2: optimal ordering of multiple-features-at-once:

1

strength: incremental, top-down mutual information

2

significance: bottom-up χ2-based pruning

latter is known as C4.5; developed for the inference of decision tree classifiers from data

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-95
SLIDE 95

Introduction Data Analysis Conclusions

Testing Significance and Strength of Association

GOAL: correlate context features with the single voicing feature OPTION 1: standard, one-feature-at-a-time:

1

significance: a 2 × 2 χ2-test

2

strength: mutual information (or other entropy-related)

OPTION 2: optimal ordering of multiple-features-at-once:

1

strength: incremental, top-down mutual information

2

significance: bottom-up χ2-based pruning

latter is known as C4.5; developed for the inference of decision tree classifiers from data

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-96
SLIDE 96

Introduction Data Analysis Conclusions

Inferred Decision Tree for Laughter Initiation

initiation of laughter: look at those laughter frames which are the first frames of each bout the inferred decision tree, χ2-pruned (p < 0.05) to retain

  • nly statistically significant nodes:

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-97
SLIDE 97

Introduction Data Analysis Conclusions

Inferred Decision Tree for Laughter Initiation

initiation of laughter: look at those laughter frames which are the first frames of each bout the inferred decision tree, χ2-pruned (p < 0.05) to retain

  • nly statistically significant nodes:

laughing with voicing at time t + 1 > 0 # of other participants NO YES at time t − 1? laugher speaking voiced voiced unvoiced

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-98
SLIDE 98

Introduction Data Analysis Conclusions

Understanding the Laughter Initiation Decision Tree

Case 1 when at least one other participant laughs with voicing just after − → voiced

?

laughing with voicing at time t + 1 > 0 # of other participants voiced voiced laugher speaking at time t − 1? YES NO unvoiced voiced laughter unvoiced laughter speech t − 1 t t + 1 k k′

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-99
SLIDE 99

Introduction Data Analysis Conclusions

Understanding the Laughter Initiation Decision Tree

Case 2 when no other participants laugh with voicing just after AND the laugher speaks just before − → voiced

X ?

t − 1 t t + 1 k′ k voiced laughter unvoiced laughter speech laughing with voicing at time t + 1 # of other participants voiced voiced NO unvoiced laugher speaking at time t − 1? > 0 YES

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-100
SLIDE 100

Introduction Data Analysis Conclusions

Understanding the Laughter Initiation Decision Tree

Case 3 when no other participants laugh with voicing just after AND the laugher does not speak just before − → unvoiced

X X ?

laughing with voicing at time t + 1 # of other participants voiced voiced laugher speaking at time t − 1? > 0 YES NO unvoiced voiced laughter unvoiced laughter speech t − 1 t t + 1 k k′

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-101
SLIDE 101

Introduction Data Analysis Conclusions

Inferred Decision Tree for Laughter Termination

termination of laughter: look at those laughter frames which are the last frames of each bout the inferred decision tree, χ2-pruned (p < 0.05) to retain

  • nly statistically significant nodes:

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-102
SLIDE 102

Introduction Data Analysis Conclusions

Inferred Decision Tree for Laughter Termination

termination of laughter: look at those laughter frames which are the last frames of each bout the inferred decision tree, χ2-pruned (p < 0.05) to retain

  • nly statistically significant nodes:

laughing with voicing at time t − 1 > 0 # of other participants YES NO at time t + 1? laugher speaking voiced voiced unvoiced

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-103
SLIDE 103

Introduction Data Analysis Conclusions

Understanding the Laughter Termination Decision Tree

Case 1 when at least one other participant laughs with voicing just before − → voiced

?

laughing with voicing > 0 # of other participants voiced voiced laugher speaking YES NO unvoiced voiced laughter unvoiced laughter speech t − 1 t t + 1 k k′ at time t − 1 at time t + 1?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-104
SLIDE 104

Introduction Data Analysis Conclusions

Understanding the Laughter Termination Decision Tree

Case 2 when no other participants laugh with voicing just before AND the laugher speaks just after − → voiced

X ?

laughing with voicing # of other participants voiced voiced NO unvoiced laugher speaking > 0 YES voiced laughter unvoiced laughter speech at time t − 1 at time t + 1? t − 1 t t + 1 k′ k

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-105
SLIDE 105

Introduction Data Analysis Conclusions

Understanding the Laughter Termination Decision Tree

Case 3 when no other participants laugh with voicing just before AND the laugher does not speak just after − → unvoiced

X X ?

laughing with voicing # of other participants voiced voiced laugher speaking > 0 YES NO unvoiced voiced laughter unvoiced laughter speech t − 1 t t + 1 k k′ at time t − 1 at time t + 1?

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-106
SLIDE 106

Introduction Data Analysis Conclusions

Some Interesting Observations

we found no statistically significant tree for laughter frames that were neither the first nor the last frame of a bout the initiation and termination tree are exactly symmetrical

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-107
SLIDE 107

Introduction Data Analysis Conclusions

Some Interesting Observations

we found no statistically significant tree for laughter frames that were neither the first nor the last frame of a bout the initiation and termination tree are exactly symmetrical

laughing with voicing at time t + 1 > 0 # of other participants laughing with voicing at time t − 1 > 0 # of other participants NO YES YES NO at time t − 1? laugher speaking at time t + 1? laugher speaking voiced voiced voiced voiced unvoiced unvoiced

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-108
SLIDE 108

Introduction Data Analysis Conclusions

Conclusions I

  • f 13209 studied bouts of laughter, 66.5% appear to be

voiced and 33.5% appear to be unvoiced

  • n average, each participant spends approximately 10% of

their vocalization effort on laughter (as opposed to speech) bout durations follow a log-normal distribution, as expected

the mode of voiced laugh bout durations is approximately twice as large as that of unvoiced laugh bout durations but bout duration does not discrimitate between voiced and unvoiced laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-109
SLIDE 109

Introduction Data Analysis Conclusions

Conclusions I

  • f 13209 studied bouts of laughter, 66.5% appear to be

voiced and 33.5% appear to be unvoiced

  • n average, each participant spends approximately 10% of

their vocalization effort on laughter (as opposed to speech) bout durations follow a log-normal distribution, as expected

the mode of voiced laugh bout durations is approximately twice as large as that of unvoiced laugh bout durations but bout duration does not discrimitate between voiced and unvoiced laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-110
SLIDE 110

Introduction Data Analysis Conclusions

Conclusions I

  • f 13209 studied bouts of laughter, 66.5% appear to be

voiced and 33.5% appear to be unvoiced

  • n average, each participant spends approximately 10% of

their vocalization effort on laughter (as opposed to speech) bout durations follow a log-normal distribution, as expected

the mode of voiced laugh bout durations is approximately twice as large as that of unvoiced laugh bout durations but bout duration does not discrimitate between voiced and unvoiced laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-111
SLIDE 111

Introduction Data Analysis Conclusions

Conclusions I

  • f 13209 studied bouts of laughter, 66.5% appear to be

voiced and 33.5% appear to be unvoiced

  • n average, each participant spends approximately 10% of

their vocalization effort on laughter (as opposed to speech) bout durations follow a log-normal distribution, as expected

the mode of voiced laugh bout durations is approximately twice as large as that of unvoiced laugh bout durations but bout duration does not discrimitate between voiced and unvoiced laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-112
SLIDE 112

Introduction Data Analysis Conclusions

Conclusions I

  • f 13209 studied bouts of laughter, 66.5% appear to be

voiced and 33.5% appear to be unvoiced

  • n average, each participant spends approximately 10% of

their vocalization effort on laughter (as opposed to speech) bout durations follow a log-normal distribution, as expected

the mode of voiced laugh bout durations is approximately twice as large as that of unvoiced laugh bout durations but bout duration does not discrimitate between voiced and unvoiced laughter

Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-113
SLIDE 113

Introduction Data Analysis Conclusions

Conclusions II

1 laughter which begins just before others laugh with voicing

and laughter which ends just after others laugh with voicing is likely to be voiced

2 when not (1), laughter which begins after the laugher speaks

and laughter which ends before the laugher speaks is likely to be voiced

3 when not (1) or (2), laughter is likely to be unvoiced Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-114
SLIDE 114

Introduction Data Analysis Conclusions

Conclusions II

1 laughter which begins just before others laugh with voicing

and laughter which ends just after others laugh with voicing is likely to be voiced

2 when not (1), laughter which begins after the laugher speaks

and laughter which ends before the laugher speaks is likely to be voiced

3 when not (1) or (2), laughter is likely to be unvoiced Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter

slide-115
SLIDE 115

Introduction Data Analysis Conclusions

Conclusions II

1 laughter which begins just before others laugh with voicing

and laughter which ends just after others laugh with voicing is likely to be voiced

2 when not (1), laughter which begins after the laugher speaks

and laughter which ends before the laugher speaks is likely to be voiced

3 when not (1) or (2), laughter is likely to be unvoiced Kornel Laskowski & Susanne Burger ICPhS 2007 Workshop on the Acoustics of Laughter