Analysis of the Occurrence of Laughter in Meetings

Similar documents
Detecting Attempts at Humor in Multiparty Meetings

Improving Frame Based Automatic Laughter Detection

Automatic Laughter Detection

Automatic Laughter Detection

Automatic Laughter Segmentation. Mary Tai Knox

Narrative Theme Navigation for Sitcoms Supported by Fan-generated Scripts

Music Segmentation Using Markov Chain Methods

Automatic discrimination between laughter and speech

PSYCHOLOGICAL AND CROSS-CULTURAL EFFECTS ON LAUGHTER SOUND PRODUCTION Marianna De Benedictis Università di Bari

Narrative Theme Navigation for Sitcoms Supported by Fan-generated Scripts

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

Multimodal databases at KTH

A Phonetic Analysis of Natural Laughter, for Use in Automatic Laughter Processing Systems

READING NOVEMBER, 2017 Part 5, 7 and 8

DiAML-Annotated Examples

Comparison of N-Gram 1 Rank Frequency Data from the Written Texts of the British National Corpus World Edition (BNC) and the author s Web Corpus

NEW QUERY-BY-HUMMING MUSIC RETRIEVAL SYSTEM CONCEPTION AND EVALUATION BASED ON A QUERY NATURE STUDY

Music Radar: A Web-based Query by Humming System

The MAMI Query-By-Voice Experiment Collecting and annotating vocal queries for music information retrieval

WASD PA Core Music Curriculum

A repetition-based framework for lyric alignment in popular songs

Humor: Prosody Analysis and Automatic Recognition for F * R * I * E * N * D * S *

Automatic transcription is not neutral. Wyke Stommel, Tom Koole, Tessa van Charldorp, Sandra van Dulmen en Antal van den Bosch ADVANT

Correlation to the Common Core State Standards

IMPROVED MELODIC SEQUENCE MATCHING FOR QUERY BASED SEARCHING IN INDIAN CLASSICAL MUSIC

AUTOMATIC RECOGNITION OF LAUGHTER

PulseCounter Neutron & Gamma Spectrometry Software Manual

Advertisement Detection and Replacement using Acoustic and Visual Repetition

BioGraph Infiniti Physiology Suite

ESS Questions administered by telephone or in person:

Time We Have Left. Episode 6 "First Day Back" Written By. Jason R. Harris

Phone-based Plosive Detection

A CONVERSATION ANALYSIS OF VERBAL BACKCHANNEL RESPONSE IN RADIO PROGRAM VALENTINE IN THE MORNING INTERVIEW WITH MICHAEL BUBLÉ.

PS User Guide Series Seismic-Data Display

Smile and Laughter in Human-Machine Interaction: a study of engagement

6.5 Percussion scalograms and musical rhythm

Project: IEEE P Working Group for Wireless Personal Area Networks (WPANs)

Work Package 9. Deliverable 32. Statistical Comparison of Islamic and Byzantine chant in the Worship Spaces

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

DETECTION OF SLOW-MOTION REPLAY SEGMENTS IN SPORTS VIDEO FOR HIGHLIGHTS GENERATION

BA in Acting, Fall 2018

Structured training for large-vocabulary chord recognition. Brian McFee* & Juan Pablo Bello

Exercise 1: Muscles in Face used for Smiling and Frowning Aim: To study the EMG activity in muscles of the face that work to smile or frown.

Real-time spectrum analyzer. Gianfranco Miele, Ph.D

Using the new psychoacoustic tonality analyses Tonality (Hearing Model) 1

arxiv: v1 [cs.sd] 8 Jun 2016

homework solutions for: Homework #4: Signal-to-Noise Ratio Estimation submitted to: Dr. Joseph Picone ECE 8993 Fundamentals of Speech Recognition

Acoustic synchronization: Rebuttal of Thomas reply to Linsker et al.

VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS. O. Javed, S. Khan, Z. Rasheed, M.Shah. {ojaved, khan, zrasheed,

Temporal data mining for root-cause analysis of machine faults in automotive assembly lines

Department of Electrical & Electronic Engineering Imperial College of Science, Technology and Medicine. Project: Real-Time Speech Enhancement

Fusion for Audio-Visual Laughter Detection

Analysis of the effects of signal distance on spectrograms

Transcription of the Singing Melody in Polyphonic Music

Topic 11. Score-Informed Source Separation. (chroma slides adapted from Meinard Mueller)

Assessment may include recording to be evaluated by students, teachers, and/or administrators in addition to live performance evaluation.

Revolutionizing the Transfer Die Industry for Maximum Production. STM Manufacturing, Inc.

LAB 1: Plotting a GM Plateau and Introduction to Statistical Distribution. A. Plotting a GM Plateau. This lab will have two sections, A and B.

Medium and High Voltage Circuit Breakers Characteristic Time Quantities of the Circuit Breaker with Applications

Audio Feature Extraction for Corpus Analysis

Automatic Rhythmic Notation from Single Voice Audio Sources

Experiments with Fisher Data

Finding Alternative Musical Scales

Browsing News and Talk Video on a Consumer Electronics Platform Using Face Detection

Critical Thinking 4.2 First steps in analysis Overcoming the natural attitude Acknowledging the limitations of perception

TRANSCRIBING GUIDELINES

Expressive Singing Synthesis based on Unit Selection for the Singing Synthesis Challenge 2016

EXTERNAL. A digital version of this document is available to download and submit online at ADDRESS POST CODE

USING MUSICAL STRUCTURE TO ENHANCE AUTOMATIC CHORD TRANSCRIPTION

Story Tracking in Video News Broadcasts. Ph.D. Dissertation Jedrzej Miadowicz June 4, 2004

CHAPTER I INTRODUCTION. language such as in a play or a film. Meanwhile the written dialogue is a dialogue

CS229 Project Report Polyphonic Piano Transcription

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University

Metonymy and Metaphor in Cross-media Semantic Interplay

Interface Practices Subcommittee SCTE STANDARD SCTE Measurement Procedure for Noise Power Ratio

Magnetic Rower. Manual Jetstream JMR-5000

Hidden Markov Model based dance recognition

Sampling Issues in Image and Video

VISSIM Tutorial. Starting VISSIM and Opening a File CE 474 8/31/06

Measurement of automatic brightness control in televisions critical for effective policy-making

Text Type Classification for the Historical DTA Corpus

Laughter Among Deaf Signers

... A Pseudo-Statistical Approach to Commercial Boundary Detection. Prasanna V Rangarajan Dept of Electrical Engineering Columbia University

Computer Coordination With Popular Music: A New Research Agenda 1

2014A Cappella Harmonv Academv Handout #2 Page 1. Sweet Adelines International Balance & Blend Joan Boutilier

TR 038 SUBJECTIVE EVALUATION OF HYBRID LOG GAMMA (HLG) FOR HDR AND SDR DISTRIBUTION

R&S FSW-B512R Real-Time Spectrum Analyzer 512 MHz Specifications

Music Information Retrieval Using Audio Input

Pre-processing of revolution speed data in ArtemiS SUITE 1

Knowledge-Based Systems

Composer Style Attribution

5. Analysing Casual Conversation Contents

C8491 C8000 1/17. digital audio modular processing system. 3G/HD/SD-SDI DSP 4/8/16 audio channels. features. block diagram

Laboratory Assignment 3. Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB

How to give a telling talk

jsymbolic 2: New Developments and Research Opportunities

Music 209 Advanced Topics in Computer Music Lecture 4 Time Warping

The MAHNOB Laughter Database. Stavros Petridis, Brais Martinez, Maja Pantic

ME 515 Mechatronics. Introduction to Digital Electronics

Reference. TDS7000 Series Digital Phosphor Oscilloscopes

Transcription:

Analysis of the Occurrence of Laughter in Meetings Kornel Laskowski 1,2 & Susanne Burger 2 1 interact, Universität Karlsruhe 2 interact, Carnegie Mellon University August 29, 2007

Introduction primary motivation: meeting understanding

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management emotion relevant

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management emotion relevant

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management emotion relevant laughter detection is particularly important for understanding both interaction and emotion if laughter occurs frequently

Introduction primary motivation: meeting understanding vocalization verbal non verbal words word fragments laughter other statements questions backchannel disruption floor grabbers interaction managing both emotion relevant other propositional content interaction management emotion relevant laughter detection is particularly important for understanding both interaction and emotion if laughter occurs frequently to date, for meetings, it is not known 1 how much laughter there actually is 2 when it tends to occur

Text-Independent Modeling of Multi-Participant Meetings To find interaction, model participants jointly.

Text-Independent Modeling of Multi-Participant Meetings To find interaction, model participants jointly. essentially monologue

Text-Independent Modeling of Multi-Participant Meetings To find interaction, model participants jointly. multi-logue

Text-Independent Modeling of Multi-Participant Meetings To find interaction, model participants jointly. multi-logue with more participant involvement

Text-Independent Modeling of Multi-Participant Meetings To find interaction, model participants jointly. a mathematical artifact (the Haar wavelet basis)

Text-Independent Modeling of Multi-Participant Meetings To find interaction, model participants jointly. multi-logue

Text-Independent Modeling of Multi-Participant Meetings To find interaction, model participants jointly. multi-logue with laughter participants tend to wait to speak participants do not wait to laugh

Three Questions of Interest 1 What is the quantity of laughter, relative to the quantity of speech?

Three Questions of Interest 1 What is the quantity of laughter, relative to the quantity of speech? 2 How does the durational distribution of episodes of laughter differ from that of episodes of speech?

Three Questions of Interest 1 What is the quantity of laughter, relative to the quantity of speech? 2 How does the durational distribution of episodes of laughter differ from that of episodes of speech? 3 How do meeting participants appear to affect each other in their use of laughter, relative to their use of speech?

Laugh Bouts vs Talk Spurts we will contrast the occurrence of laughter L with that of speech S

Laugh Bouts vs Talk Spurts we will contrast the occurrence of laughter L with that of speech S talk spurts contiguous per-participant intervals of speech (Shriberg et al, 2001), containing pauses no longer than 300 ms (as in NIST RT-06s SAD)

Laugh Bouts vs Talk Spurts we will contrast the occurrence of laughter L with that of speech S talk spurts contiguous per-participant intervals of speech (Shriberg et al, 2001), containing pauses no longer than 300 ms (as in NIST RT-06s SAD) laugh bouts contiguous per-participant intervals of laughter (Bachorowski et al, 2001), including recovery inhalation

Laugh Bouts vs Talk Spurts we will contrast the occurrence of laughter L with that of speech S talk spurts contiguous per-participant intervals of speech (Shriberg et al, 2001), containing pauses no longer than 300 ms (as in NIST RT-06s SAD) laugh bouts contiguous per-participant intervals of laughter (Bachorowski et al, 2001), including recovery inhalation S/L islands contiguous per-group intervals in which at least one participant talks/laughs

Laugh Bouts vs Talk Spurts we will contrast the occurrence of laughter L with that of speech S talk spurt laugh bout talk spurt islands laugh bout islands

The ICSI Meeting Corpus naturally occurring project-oriented conversations with varying number of participants

The ICSI Meeting Corpus naturally occurring project-oriented conversations with varying number of participants the largest such corpus available type # of # of participants meetings mod min max Bed 15 6 4 7 Bmr 29 7 3 9 Bro 23 6 4 8 other 8 6 5 8

The ICSI Meeting Corpus naturally occurring project-oriented conversations with varying number of participants the largest such corpus available type # of # of participants meetings mod min max Bed 15 6 4 7 Bmr 29 7 3 9 Bro 23 6 4 8 other 8 6 5 8 rarely, meetings contain additional, uninstrumented participants (we ignore them)

The ICSI Meeting Corpus naturally occurring project-oriented conversations with varying number of participants the largest such corpus available type # of # of participants meetings mod min max Bed 15 6 4 7 Bmr 29 7 3 9 Bro 23 6 4 8 other 8 6 5 8 rarely, meetings contain additional, uninstrumented participants (we ignore them) we use all 75 meetings: 66.3 hours of conversation

Identifying Laughter in the ICSI Corpus laughter is already annotated with rich XML-style mark-up

Identifying Laughter in the ICSI Corpus laughter is already annotated with rich XML-style mark-up therefore, for our purposes, data preprocessing consists of:

Identifying Laughter in the ICSI Corpus laughter is already annotated with rich XML-style mark-up therefore, for our purposes, data preprocessing consists of: 1 identifying laughter in the orthographic transcription

Identifying Laughter in the ICSI Corpus laughter is already annotated with rich XML-style mark-up therefore, for our purposes, data preprocessing consists of: 1 identifying laughter in the orthographic transcription 2 specifying endpoints for identified laughter

Identifying Laughter in the ICSI Corpus laughter is already annotated with rich XML-style mark-up therefore, for our purposes, data preprocessing consists of: 1 identifying laughter in the orthographic transcription 2 specifying endpoints for identified laughter 1 orthographic, time-segmented transcription of speaker contributions (.stm) Bmr011 me013 chan1 3029.466 3029.911 Yeah. Bmr011 mn005 chan3 3030.230 3031.140 Film-maker. Bmr011 fe016 chan0 3030.783 3032.125 <Emphasis> colorful. </Emphasi... Bmr011 me011 chanb 3035.301 3036.964 Of beeps, yeah. Bmr011 fe008 chan8 3035.714 3037.314 <Pause/> of m- one hour of - <... Bmr011 mn014 chan2 3036.030 3036.640 Yeah. Bmr011 me013 chan1 3036.280 3037.600 <VocalSound Description="laugh"/> Bmr011 mn014 chan2 3036.640 3037.115 Yeah. Bmr011 mn005 chan3 3036.930 3037.335 Is - Bmr011 me011 chanb 3036.964 3038.573 <VocalSound Description="laugh"/>

Identifying Laughter in the ICSI Corpus laughter is already annotated with rich XML-style mark-up therefore, for our purposes, data preprocessing consists of: 1 identifying laughter in the orthographic transcription 2 specifying endpoints for identified laughter 1 orthographic, time-segmented transcription of speaker contributions (.stm)...9.911 Yeah....1.140 Film-maker....2.125 <Emphasis> colorful. </Emphasis> <Comment Description="while laughing"/>...6.964 Of beeps, yeah....7.314 <Pause/> of m- one hour of - <Comment Description="while laughing"/>...6.640 Yeah....7.600 <VocalSound Description="laugh"/>...7.115 Yeah....7.335 Is -...8.573 <VocalSound Description="laugh"/>

Identifying Laughter in the ICSI Corpus laughter is already annotated with rich XML-style mark-up therefore, for our purposes, data preprocessing consists of: 1 identifying laughter in the orthographic transcription 2 specifying endpoints for identified laughter 1 orthographic, time-segmented transcription of speaker contributions (.stm)...9.911 Yeah....1.140 Film-maker....2.125 <Emphasis> colorful. </Emphasis> <Comment Description="while laughing"/>...6.964 Of beeps, yeah....7.314 <Pause/> of m- one hour of - <Comment Description="while laughing"/>...6.640 Yeah....7.600 <VocalSound Description="laugh"/>...7.115 Yeah....7.335 Is -...8.573 <VocalSound Description="laugh"/>

Identifying Laughter in the ICSI Corpus laughter is already annotated with rich XML-style mark-up therefore, for our purposes, data preprocessing consists of: 1 identifying laughter in the orthographic transcription 2 specifying endpoints for identified laughter 1 orthographic, time-segmented transcription of speaker contributions (.stm)...9.911 Yeah....1.140 Film-maker....2.125 <Emphasis> colorful. </Emphasis> <Comment Description="while laughing"/>...6.964 Of beeps, yeah....7.314 <Pause/> of m- one hour of - <Comment Description="while laughing"/>...6.640 Yeah....7.600 <VocalSound Description="laugh"/>...7.115 Yeah....7.335 Is -...8.573 <VocalSound Description="laugh"/>

Sample VocalSound Instances Freq Token Rank Count VocalSound Description 1 11515 laugh 2 7091 breath 3 4589 inbreath 4 2223 mouth 5 970 breath-laugh 11 97 laugh-breath 46 6 cough-laugh 63 3 laugh, "hmmph" 69 3 breath while smiling 75 2 very long laugh Used

Sample VocalSound Instances Freq Token Rank Count VocalSound Description 1 11515 laugh 2 7091 breath 3 4589 inbreath 4 2223 mouth 5 970 breath-laugh 11 97 laugh-breath 46 6 cough-laugh 63 3 laugh, "hmmph" 69 3 breath while smiling 75 2 very long laugh Used laughter is by far the most common non-verbal VocalSound idem for Comment instances

Segmenting Identified Laughter Instances found 12570 non-farfield VocalSound laughs

Segmenting Identified Laughter Instances found 12570 non-farfield VocalSound laughs 11845 were adjacent to a time-stamped utterance boundary or lexical item: endpoints were derived automatically 725 needed to be segmented manually

Segmenting Identified Laughter Instances found 12570 non-farfield VocalSound laughs 11845 were adjacent to a time-stamped utterance boundary or lexical item: endpoints were derived automatically 725 needed to be segmented manually found 1108 non-farfield Comment laughs all needed to be segmented manually

Segmenting Identified Laughter Instances found 12570 non-farfield VocalSound laughs 11845 were adjacent to a time-stamped utterance boundary or lexical item: endpoints were derived automatically 725 needed to be segmented manually found 1108 non-farfield Comment laughs all needed to be segmented manually manual segmententation performed by one annotator, checked by at least one other annotator

Segmenting Identified Laughter Instances found 12570 non-farfield VocalSound laughs 11845 were adjacent to a time-stamped utterance boundary or lexical item: endpoints were derived automatically 725 needed to be segmented manually found 1108 non-farfield Comment laughs all needed to be segmented manually manual segmententation performed by one annotator, checked by at least one other annotator merging immediately adjacent VocalSound and Comment instances, and removing transcribed instances for which we found counterevidence, resulted in 13259 bouts

Speech vs Laughter by Time 13259 laugh bouts

Speech vs Laughter by Time 13259 laugh bouts 110790 talk spurts

Speech vs Laughter by Time 13259 laugh bouts 110790 talk spurts by personal time:

Speech vs Laughter by Time 13259 laugh bouts 110790 talk spurts by personal time: 442.6 hours total recorded audio

Speech vs Laughter by Time 13259 laugh bouts 110790 talk spurts by personal time: 442.6 hours total recorded audio 55.2 hours spent in talk spurts (S), 12.47%

Speech vs Laughter by Time 13259 laugh bouts 110790 talk spurts by personal time: 442.6 hours total recorded audio 55.2 hours spent in talk spurts (S), 12.47% 5.6 hours spent in laugh bouts (L), 1.27%

Speech vs Laughter by Time, by Participant

Talk Spurt Duration vs Laugh Bout Duration

Vocalization Overlap Vocal Activity per part Vocalizing Time, hrs number of simultaneously per vocalizing participants meet 1 2 3 4 S 55.2 50.8 46.7 3.8 0.27 0.02 L 5.6 3.3 2.0 0.7 0.31 0.27 S L 0.2 0.2 0.2 0.0 0.0 0 S L 60.3 52.0 45.7 4.8 0.88 0.49

Vocalization Overlap Vocal Activity per part Vocalizing Time, hrs number of simultaneously per vocalizing participants meet 1 2 3 4 S 55.2 50.8 46.7 3.8 0.27 0.02 L 5.6 3.3 2.0 0.7 0.31 0.27 S L 0.2 0.2 0.2 0.0 0.0 0 S L 60.3 52.0 45.7 4.8 0.88 0.49 in S only, 84.6% of vocalization is not overlapped

Vocalization Overlap Vocal Activity per part Vocalizing Time, hrs number of simultaneously per vocalizing participants meet 1 2 3 4 S 55.2 50.8 46.7 3.8 0.27 0.02 L 5.6 3.3 2.0 0.7 0.31 0.27 S L 0.2 0.2 0.2 0.0 0.0 0 S L 60.3 52.0 45.7 4.8 0.88 0.49 in L only, 35.7% of vocalization is not overlapped

Vocalization Overlap Vocal Activity per part Vocalizing Time, hrs number of simultaneously per vocalizing participants meet 1 2 3 4 S 55.2 50.8 46.7 3.8 0.27 0.02 L 5.6 3.3 2.0 0.7 0.31 0.27 S L 0.2 0.2 0.2 0.0 0.0 0 S L 60.3 52.0 45.7 4.8 0.88 0.49 the proportion of laughed speech is negligible

Vocalization Overlap Vocal Activity per part Vocalizing Time, hrs number of simultaneously per vocalizing participants meet 1 2 3 4 S 55.2 50.8 46.7 3.8 0.27 0.02 L 5.6 3.3 2.0 0.7 0.31 0.27 S L 0.2 0.2 0.2 0.0 0.0 0 S L 60.3 52.0 45.7 4.8 0.88 0.49 there is 3 times as much 3-participant overlap when considering S L as opposed to S only

Vocalization Overlap Vocal Activity per part Vocalizing Time, hrs number of simultaneously per vocalizing participants meet 1 2 3 4 S 55.2 50.8 46.7 3.8 0.27 0.02 L 5.6 3.3 2.0 0.7 0.31 0.27 S L 0.2 0.2 0.2 0.0 0.0 0 S L 60.3 52.0 45.7 4.8 0.88 0.49 there is 25 times as much 4-participant overlap when considering S L as opposed to S only

Overlap Dynamics does laughter differ from speech in the way in which overlap arises and is resolved?

Overlap Dynamics does laughter differ from speech in the way in which overlap arises and is resolved? look at transition probabilities under a first-order Markov assumption

Overlap Dynamics does laughter differ from speech in the way in which overlap arises and is resolved? look at transition probabilities under a first-order Markov assumption 1 discretize L and S segmentations using non-overlapping analysis frames

Overlap Dynamics does laughter differ from speech in the way in which overlap arises and is resolved? look at transition probabilities under a first-order Markov assumption 1 discretize L and S segmentations using non-overlapping analysis frames 2 train an Extended Degree-of-Overlap (EDO) model on the discretized L and S segmentations P ({A} {A, B}) P ({A,B} {A}) P ({A} {B}) etc.

Overlap Dynamics does laughter differ from speech in the way in which overlap arises and is resolved? look at transition probabilities under a first-order Markov assumption 1 discretize L and S segmentations using non-overlapping analysis frames 2 train an Extended Degree-of-Overlap (EDO) model on the discretized L and S segmentations P ({A} {A, B}) P ({A,B} {A}) P ({A} {B}) etc. 3 compare inferred probabilities for L and S

Overlap Dynamics: Results Select EDO Transitions 500ms frames from (at t) to (at t + 1) S L {A} {A} 82.94 57.96 {A} {A, B} 6.21 8.43 {A} {A,B,C, } 0.39 2.39 {A, B} {A} 45.49 26.37 {A, B} {A, B} 40.88 46.93 {A,B} {A,B,C, } 4.46 13.65 {A,B,C, } {A} 19.24 6.69 {A,B,C, } {A,B} 40.94 17.45 {A,B,C, } {A,B,C, } 29.44 71.04

Overlap Dynamics: Results Select EDO Transitions 500ms frames from (at t) to (at t + 1) S L {A} {A} 82.94 57.96 {A} {A, B} 6.21 8.43 {A} {A,B,C, } 0.39 2.39 {A, B} {A} 45.49 26.37 {A, B} {A, B} 40.88 46.93 {A,B} {A,B,C, } 4.46 13.65 {A,B,C, } {A} 19.24 6.69 {A,B,C, } {A,B} 40.94 17.45 {A,B,C, } {A,B,C, } 29.44 71.04

Conclusions Based on the ICSI meetings, 1 approximately 9% of vocalizing time is spent on laughter

Conclusions Based on the ICSI meetings, 1 approximately 9% of vocalizing time is spent on laughter but participants vary widely (0% - 30%)

Conclusions Based on the ICSI meetings, 1 approximately 9% of vocalizing time is spent on laughter but participants vary widely (0% - 30%) 2 on average, laughter occurs once a minute

Conclusions Based on the ICSI meetings, 1 approximately 9% of vocalizing time is spent on laughter but participants vary widely (0% - 30%) 2 on average, laughter occurs once a minute 3 laughter accounts for the large majority of 3 participant overlap

Conclusions Based on the ICSI meetings, 1 approximately 9% of vocalizing time is spent on laughter but participants vary widely (0% - 30%) 2 on average, laughter occurs once a minute 3 laughter accounts for the large majority of 3 participant overlap 4 in contrast to speech, once laughter overlap is incurred, it is most likely to persist

Conclusions Based on the ICSI meetings, 1 approximately 9% of vocalizing time is spent on laughter but participants vary widely (0% - 30%) 2 on average, laughter occurs once a minute 3 laughter accounts for the large majority of 3 participant overlap 4 in contrast to speech, once laughter overlap is incurred, it is most likely to persist ie. 3-participant speech overlap is 2.5 times more likely than laughter to be resolved within 500 ms

We would like to thank: our annotators: Jörg Brunstein and Matthew Bell discussion: Alan Black and Liz Shriberg funding: EU CHIL