Laugh when you re winning

Similar documents
Seminar CHIST-ERA Istanbul : 4 March 2014 Kick-off meeting : 27 January 2014 (call IUI 2012)

Multimodal Analysis of laughter for an Interactive System

Laugh-aware Virtual Agent and its Impact on User Amusement

Rhythmic Body Movements of Laughter

The AV-LASYN Database : A synchronous corpus of audio and 3D facial marker data for audio-visual laughter synthesis

Real-time Laughter on Virtual Characters

Pitch-Synchronous Spectrogram: Principles and Applications

Digital audio and computer music. COS 116, Spring 2012 Guest lecture: Rebecca Fiebrink

AUTOMATIC RECOGNITION OF LAUGHTER

UNIVERSITY OF DUBLIN TRINITY COLLEGE

Laughter Animation Synthesis

Advanced Signal Processing 2

Multimodal databases at KTH

EE513 Audio Signals and Systems. Introduction Kevin D. Donohue Electrical and Computer Engineering University of Kentucky

Perception of Intensity Incongruence in Synthesized Multimodal Expressions of Laughter

Automatic Laughter Detection

Speech and Speaker Recognition for the Command of an Industrial Robot

The Belfast Storytelling Database

The Belfast Storytelling Database: A spontaneous social interaction database with laughter focused annotation

A Phonetic Analysis of Natural Laughter, for Use in Automatic Laughter Processing Systems

Music Radar: A Web-based Query by Humming System

LAUGHTER serves as an expressive social signal in human

Analysis, Synthesis, and Perception of Musical Sounds

Music Genre Classification and Variance Comparison on Number of Genres

Singing voice synthesis in Spanish by concatenation of syllables based on the TD-PSOLA algorithm

Deep learning for music data processing

Improving Frame Based Automatic Laughter Detection

Interacting with a Virtual Conductor

On Human Capability and Acoustic Cues for Discriminating Singing and Speaking Voices

A COMPARATIVE EVALUATION OF VOCODING TECHNIQUES FOR HMM-BASED LAUGHTER SYNTHESIS

2. AN INTROSPECTION OF THE MORPHING PROCESS

Narrative Theme Navigation for Sitcoms Supported by Fan-generated Scripts

Smile and Laughter in Human-Machine Interaction: a study of engagement

Automatic acoustic synthesis of human-like laughter

MAKING INTERACTIVE GUIDES MORE ATTRACTIVE

hit), and assume that longer incidental sounds (forest noise, water, wind noise) resemble a Gaussian noise distribution.

Joint bottom-up/top-down machine learning structures to simulate human audition and musical creativity

Automatic Laughter Detection

Acoustic Scene Classification

On human capability and acoustic cues for discriminating singing and speaking voices

This full text version, available on TeesRep, is the post-print (final version prior to publication) of:

Retrieval of textual song lyrics from sung inputs

A METHOD OF MORPHING SPECTRAL ENVELOPES OF THE SINGING VOICE FOR USE WITH BACKING VOCALS

1. Introduction NCMMSC2009

Semi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis

POLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING

Singer Traits Identification using Deep Neural Network

AUTOMASHUPPER: AN AUTOMATIC MULTI-SONG MASHUP SYSTEM

HBI Database. Version 2 (User Manual)

DERIVING A TIMBRE SPACE FOR THREE TYPES OF COMPLEX TONES VARYING IN SPECTRAL ROLL-OFF

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

VOCALISTENER: A SINGING-TO-SINGING SYNTHESIS SYSTEM BASED ON ITERATIVE PARAMETER ESTIMATION

Topics in Computer Music Instrument Identification. Ioanna Karydi

Audiovisual analysis of relations between laughter types and laughter motions

Appendix C ACCESSIBILITY SCALE CLOSED OPEN

Proposal for Application of Speech Techniques to Music Analysis

Singing voice synthesis based on deep neural networks

Speech Recognition and Signal Processing for Broadcast News Transcription

Laughter Among Deaf Signers

Tempo and Beat Analysis

ADX TRAX. Insights on Audio Source Separation. Audio Praxis. Rick Silva

Efficient Computer-Aided Pitch Track and Note Estimation for Scientific Applications. Matthias Mauch Chris Cannam György Fazekas

technical note flicker measurement display & lighting measurement

IEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS 1. Automated Laughter Detection from Full-Body Movements

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Supervised Musical Source Separation from Mono and Stereo Mixtures based on Sinusoidal Modeling

Implementing and Evaluating a Laughing Virtual Character

Expressive Singing Synthesis based on Unit Selection for the Singing Synthesis Challenge 2016

SOUND LABORATORY LING123: SOUND AND COMMUNICATION

Automatic music transcription

Pattern Recognition in Music

Automatic Labelling of tabla signals

Singing Pitch Extraction and Singing Voice Separation

The laughing brain - Do only humans laugh?

Speaking loud, speaking high: non-linearities in voice strength and vocal register variations. Christophe d Alessandro LIMSI-CNRS Orsay, France

Pitch. The perceptual correlate of frequency: the perceptual dimension along which sounds can be ordered from low to high.

SINGING EXPRESSION TRANSFER FROM ONE VOICE TO ANOTHER FOR A GIVEN SONG. Sangeon Yong, Juhan Nam

Thursday, April 28, 16

Expressive Multimodal Conversational Acts for SAIBA agents

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

The Language of Film and TV

Pitch Perception and Grouping. HST.723 Neural Coding and Perception of Sound

Acoustic Analysis of Beethoven Piano Sonata Op.110. Yan-bing DING and Qiu-hua HUANG

Pitch Analysis of Ukulele

MELODY EXTRACTION FROM POLYPHONIC AUDIO OF WESTERN OPERA: A METHOD BASED ON DETECTION OF THE SINGER S FORMANT

Lab 5 Linear Predictive Coding

Semi-supervised Musical Instrument Recognition

Towards automated full body detection of laughter driven by human expert annotation

SMS Composer and SMS Conductor: Applications for Spectral Modeling Synthesis Composition and Performance

Basic Operations App Guide

Effects of acoustic degradations on cover song recognition

Supervised Learning in Genre Classification

Automatic music transcription

CHAPTER 4 SEGMENTATION AND FEATURE EXTRACTION

Engineering in Recording

AN ALGORITHM FOR LOCATING FUNDAMENTAL FREQUENCY (F0) MARKERS IN SPEECH

A New Method for Calculating Music Similarity

A Matlab toolbox for. Characterisation Of Recorded Underwater Sound (CHORUS) USER S GUIDE

Multimodal Data Collection of Human-Robot Humorous Interactions in the JOKER Project

Transcription:

Laugh when you re winning Harry Griffin for the ILHAIRE Consortium 26 July, 2013

ILHAIRE Laughter databases Laugh when you re winning project Concept & Design Architecture Multimodal analysis Overview Audiovisual synthesis 2

Laughter? Ubiquitous Frequent In normal conversion ~ 1 laugh/minute Conveys various emotions Vital nonverbal social tool Multimodal - produces lots of signals Barely used in Human-Computer Interaction 3

Incorporating Laughter into Human-Avatar Interactions: Research and Evaluation 9 partners, for 3 years from September 2011 Range of expertise and approaches Psychology of laughter & humour Natural behaviour databases Multimodal signal analysis Machine learning Dialog management Visual and auditory synthesis 4

Natural behaviour databases Essential for understanding laughter ILHAIRE laughter database: http://www.qub.ac.uk/ilhairelaughter/ 5

Laugh when you re winning: Concept Games are a good way of making people laugh! Avatars as game companions Games are complex social situations and laughter could be important in ensuring that they flow smoothly Games require a face-to-face interaction 7

Laugh when you re winning: Design One or two users Simple social games Avatar: plays active part in game is a socially competent (laughing) game companion 8

System architecture Laugh when you re winning: Concept Laughter detection & intensity estimation Visual analysis Acoustic analysis Body movement analysis Dialog Manager Laughter planner Decision to AV synthesis Audiovisual Laughter Synthesis Game behaviour Respiration analysis Context (Game state) 9

Coefficients Frequency (Hz) Laughter Detection (Voice) 500 Some speech features: 1 Intensity (db) 500 0 Waveform 0 0 1.89825 Time (s) -0.8119 0 1.89825 5000 0 Time (s) 0 1.89825 Time (s) 83.45 Spectogram Pitch (Hz) Formants 0 0 1.89825 12 Time (s) MFCCs 34.05 0 1.89825 Time (s) 500 1 0 1.89825 Time (s) 0 0 1.89825 11

Smile Detection (Face) Features: Action units from Microsoft Kinect 12

Laughter movements (Body) Kinect depth mapping for more general measures e.g., contraction index Computer vision for extraction of key movements e.g., shoulder movement frequency and amplitude Motion capture with modified suit 13

Laughter respiration (Torso) Laughter-related respiratory actions generate characteristic non-rigid body movements 14

Acoustic Laughter synthesis Little past work Lack of naturalness UMONS: HMM-based synthesis with HTS No generation of laughter sequence for the moment: we play existing laughter phonetic transcriptions 15

Visual Laughter synthesis Active MQ (Message-oriented Middleware used in SEMAINE project) Generation of common BML scripts 2 different interpretations (Greta and Living Actor) BML Living Actor conversion selection of predefined animations + combination of morphing data matching BML parameters 16

Visual Laughter synthesis Living Actor Avatar A graph of animations corresponding to different types of laughter movements and different intensities Pseudo-phonemes associated with facial expressions and lips movements Combination of head, torso, and shoulder animations 17

Thank you 18