Laugh when you re winning
|
|
- Dwain Alexander
- 5 years ago
- Views:
Transcription
1 Laugh when you re winning Harry Griffin for the ILHAIRE Consortium 26 July, 2013
2 ILHAIRE Laughter databases Laugh when you re winning project Concept & Design Architecture Multimodal analysis Overview Audiovisual synthesis 2
3 Laughter? Ubiquitous Frequent In normal conversion ~ 1 laugh/minute Conveys various emotions Vital nonverbal social tool Multimodal - produces lots of signals Barely used in Human-Computer Interaction 3
4 Incorporating Laughter into Human-Avatar Interactions: Research and Evaluation 9 partners, for 3 years from September 2011 Range of expertise and approaches Psychology of laughter & humour Natural behaviour databases Multimodal signal analysis Machine learning Dialog management Visual and auditory synthesis 4
5 Natural behaviour databases Essential for understanding laughter ILHAIRE laughter database: 5
6 Laugh when you re winning: Concept Games are a good way of making people laugh! Avatars as game companions Games are complex social situations and laughter could be important in ensuring that they flow smoothly Games require a face-to-face interaction 7
7 Laugh when you re winning: Design One or two users Simple social games Avatar: plays active part in game is a socially competent (laughing) game companion 8
8 System architecture Laugh when you re winning: Concept Laughter detection & intensity estimation Visual analysis Acoustic analysis Body movement analysis Dialog Manager Laughter planner Decision to AV synthesis Audiovisual Laughter Synthesis Game behaviour Respiration analysis Context (Game state) 9
9 Coefficients Frequency (Hz) Laughter Detection (Voice) 500 Some speech features: 1 Intensity (db) Waveform Time (s) Time (s) Time (s) Spectogram Pitch (Hz) Formants Time (s) MFCCs Time (s) Time (s)
10 Smile Detection (Face) Features: Action units from Microsoft Kinect 12
11 Laughter movements (Body) Kinect depth mapping for more general measures e.g., contraction index Computer vision for extraction of key movements e.g., shoulder movement frequency and amplitude Motion capture with modified suit 13
12 Laughter respiration (Torso) Laughter-related respiratory actions generate characteristic non-rigid body movements 14
13 Acoustic Laughter synthesis Little past work Lack of naturalness UMONS: HMM-based synthesis with HTS No generation of laughter sequence for the moment: we play existing laughter phonetic transcriptions 15
14 Visual Laughter synthesis Active MQ (Message-oriented Middleware used in SEMAINE project) Generation of common BML scripts 2 different interpretations (Greta and Living Actor) BML Living Actor conversion selection of predefined animations + combination of morphing data matching BML parameters 16
15 Visual Laughter synthesis Living Actor Avatar A graph of animations corresponding to different types of laughter movements and different intensities Pseudo-phonemes associated with facial expressions and lips movements Combination of head, torso, and shoulder animations 17
16 Thank you 18
Seminar CHIST-ERA Istanbul : 4 March 2014 Kick-off meeting : 27 January 2014 (call IUI 2012)
project JOKER JOKe and Empathy of a Robot/ECA: Towards social and affective relations with a robot Seminar CHIST-ERA Istanbul : 4 March 2014 Kick-off meeting : 27 January 2014 (call IUI 2012) http://www.chistera.eu/projects/joker
More informationMultimodal Analysis of laughter for an Interactive System
Multimodal Analysis of laughter for an Interactive System Jérôme Urbain 1, Radoslaw Niewiadomski 2, Maurizio Mancini 3, Harry Griffin 4, Hüseyin Çakmak 1, Laurent Ach 5, Gualtiero Volpe 3 1 Université
More informationLaugh-aware Virtual Agent and its Impact on User Amusement
Laugh-aware Virtual Agent and its Impact on User Amusement Radosław Niewiadomski TELECOM ParisTech Rue Dareau, 37-39 75014 Paris, France niewiado@telecomparistech.fr Tracey Platt Universität Zürich Binzmuhlestrasse,
More informationRhythmic Body Movements of Laughter
Rhythmic Body Movements of Laughter Radoslaw Niewiadomski DIBRIS, University of Genoa Viale Causa 13 Genoa, Italy radek@infomus.org Catherine Pelachaud CNRS - Telecom ParisTech 37-39, rue Dareau Paris,
More informationThe AV-LASYN Database : A synchronous corpus of audio and 3D facial marker data for audio-visual laughter synthesis
The AV-LASYN Database : A synchronous corpus of audio and 3D facial marker data for audio-visual laughter synthesis Hüseyin Çakmak, Jérôme Urbain, Joëlle Tilmanne and Thierry Dutoit University of Mons,
More informationReal-time Laughter on Virtual Characters
Utrecht University Department of Computer Science Master Thesis Game & Media Technology Real-time Laughter on Virtual Characters Author: Jordi van Duijn (ICA-3344789) Supervisor: Dr. Ir. Arjan Egges September
More informationPitch-Synchronous Spectrogram: Principles and Applications
Pitch-Synchronous Spectrogram: Principles and Applications C. Julian Chen Department of Applied Physics and Applied Mathematics May 24, 2018 Outline The traditional spectrogram Observations with the electroglottograph
More informationDigital audio and computer music. COS 116, Spring 2012 Guest lecture: Rebecca Fiebrink
Digital audio and computer music COS 116, Spring 2012 Guest lecture: Rebecca Fiebrink Overview 1. Physics & perception of sound & music 2. Representations of music 3. Analyzing music with computers 4.
More informationAUTOMATIC RECOGNITION OF LAUGHTER
AUTOMATIC RECOGNITION OF LAUGHTER USING VERBAL AND NON-VERBAL ACOUSTIC FEATURES Tomasz Jacykiewicz 1 Dr. Fabien Ringeval 2 JANUARY, 2014 DEPARTMENT OF INFORMATICS - MASTER PROJECT REPORT Département d
More informationUNIVERSITY OF DUBLIN TRINITY COLLEGE
UNIVERSITY OF DUBLIN TRINITY COLLEGE FACULTY OF ENGINEERING & SYSTEMS SCIENCES School of Engineering and SCHOOL OF MUSIC Postgraduate Diploma in Music and Media Technologies Hilary Term 31 st January 2005
More informationLaughter Animation Synthesis
Laughter Animation Synthesis Yu Ding Institut Mines-Télécom Télécom Paristech CNRS LTCI Ken Prepin Institut Mines-Télécom Télécom Paristech CNRS LTCI Jing Huang Institut Mines-Télécom Télécom Paristech
More informationAdvanced Signal Processing 2
Advanced Signal Processing 2 Synthesis of Singing 1 Outline Features and requirements of signing synthesizers HMM based synthesis of singing Articulatory synthesis of singing Examples 2 Requirements of
More informationMultimodal databases at KTH
Multimodal databases at David House, Jens Edlund & Jonas Beskow Clarin Workshop The QSMT database (2002): Facial & Articulatory motion Clarin Workshop Purpose Obtain coherent data for modelling and animation
More informationEE513 Audio Signals and Systems. Introduction Kevin D. Donohue Electrical and Computer Engineering University of Kentucky
EE513 Audio Signals and Systems Introduction Kevin D. Donohue Electrical and Computer Engineering University of Kentucky Question! If a tree falls in the forest and nobody is there to hear it, will it
More informationPerception of Intensity Incongruence in Synthesized Multimodal Expressions of Laughter
2015 International Conference on Affective Computing and Intelligent Interaction (ACII) Perception of Intensity Incongruence in Synthesized Multimodal Expressions of Laughter Radoslaw Niewiadomski, Yu
More informationAutomatic Laughter Detection
Automatic Laughter Detection Mary Knox Final Project (EECS 94) knoxm@eecs.berkeley.edu December 1, 006 1 Introduction Laughter is a powerful cue in communication. It communicates to listeners the emotional
More informationSpeech and Speaker Recognition for the Command of an Industrial Robot
Speech and Speaker Recognition for the Command of an Industrial Robot CLAUDIA MOISA*, HELGA SILAGHI*, ANDREI SILAGHI** *Dept. of Electric Drives and Automation University of Oradea University Street, nr.
More informationThe Belfast Storytelling Database
2015 International Conference on Affective Computing and Intelligent Interaction (ACII) The Belfast Storytelling Database A spontaneous social interaction database with laughter focused annotation Gary
More informationThe Belfast Storytelling Database: A spontaneous social interaction database with laughter focused annotation
The Belfast Storytelling Database: A spontaneous social interaction database with laughter focused annotation McKeown, G., Curran, W., Wagner, J., Lingenfelser, F., & André, E. (2015). The Belfast Storytelling
More informationA Phonetic Analysis of Natural Laughter, for Use in Automatic Laughter Processing Systems
A Phonetic Analysis of Natural Laughter, for Use in Automatic Laughter Processing Systems Jérôme Urbain and Thierry Dutoit Université de Mons - UMONS, Faculté Polytechnique de Mons, TCTS Lab 20 Place du
More informationMusic Radar: A Web-based Query by Humming System
Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,
More informationLAUGHTER serves as an expressive social signal in human
Audio-Facial Laughter Detection in Naturalistic Dyadic Conversations Bekir Berker Turker, Yucel Yemez, Metin Sezgin, Engin Erzin 1 Abstract We address the problem of continuous laughter detection over
More informationAnalysis, Synthesis, and Perception of Musical Sounds
Analysis, Synthesis, and Perception of Musical Sounds The Sound of Music James W. Beauchamp Editor University of Illinois at Urbana, USA 4y Springer Contents Preface Acknowledgments vii xv 1. Analysis
More informationMusic Genre Classification and Variance Comparison on Number of Genres
Music Genre Classification and Variance Comparison on Number of Genres Miguel Francisco, miguelf@stanford.edu Dong Myung Kim, dmk8265@stanford.edu 1 Abstract In this project we apply machine learning techniques
More informationSinging voice synthesis in Spanish by concatenation of syllables based on the TD-PSOLA algorithm
Singing voice synthesis in Spanish by concatenation of syllables based on the TD-PSOLA algorithm ALEJANDRO RAMOS-AMÉZQUITA Computer Science Department Tecnológico de Monterrey (Campus Ciudad de México)
More informationDeep learning for music data processing
Deep learning for music data processing A personal (re)view of the state-of-the-art Jordi Pons www.jordipons.me Music Technology Group, DTIC, Universitat Pompeu Fabra, Barcelona. 31st January 2017 Jordi
More informationImproving Frame Based Automatic Laughter Detection
Improving Frame Based Automatic Laughter Detection Mary Knox EE225D Class Project knoxm@eecs.berkeley.edu December 13, 2007 Abstract Laughter recognition is an underexplored area of research. My goal for
More informationInteracting with a Virtual Conductor
Interacting with a Virtual Conductor Pieter Bos, Dennis Reidsma, Zsófia Ruttkay, Anton Nijholt HMI, Dept. of CS, University of Twente, PO Box 217, 7500AE Enschede, The Netherlands anijholt@ewi.utwente.nl
More informationOn Human Capability and Acoustic Cues for Discriminating Singing and Speaking Voices
On Human Capability and Acoustic Cues for Discriminating Singing and Speaking Voices Yasunori Ohishi 1 Masataka Goto 3 Katunobu Itou 2 Kazuya Takeda 1 1 Graduate School of Information Science, Nagoya University,
More informationA COMPARATIVE EVALUATION OF VOCODING TECHNIQUES FOR HMM-BASED LAUGHTER SYNTHESIS
A COMPARATIVE EVALUATION OF VOCODING TECHNIQUES FOR HMM-BASED LAUGHTER SYNTHESIS Bajibabu Bollepalli 1, Jérôme Urbain 2, Tuomo Raitio 3, Joakim Gustafson 1, Hüseyin Çakmak 2 1 Department of Speech, Music
More information2. AN INTROSPECTION OF THE MORPHING PROCESS
1. INTRODUCTION Voice morphing means the transition of one speech signal into another. Like image morphing, speech morphing aims to preserve the shared characteristics of the starting and final signals,
More informationNarrative Theme Navigation for Sitcoms Supported by Fan-generated Scripts
Narrative Theme Navigation for Sitcoms Supported by Fan-generated Scripts Gerald Friedland, Luke Gottlieb, Adam Janin International Computer Science Institute (ICSI) Presented by: Katya Gonina What? Novel
More informationSmile and Laughter in Human-Machine Interaction: a study of engagement
Smile and ter in Human-Machine Interaction: a study of engagement Mariette Soury 1,2, Laurence Devillers 1,3 1 LIMSI-CNRS, BP133, 91403 Orsay cedex, France 2 University Paris 11, 91400 Orsay, France 3
More informationAutomatic acoustic synthesis of human-like laughter
Automatic acoustic synthesis of human-like laughter Shiva Sundaram,, Shrikanth Narayanan, and, and Citation: The Journal of the Acoustical Society of America 121, 527 (2007); doi: 10.1121/1.2390679 View
More informationMAKING INTERACTIVE GUIDES MORE ATTRACTIVE
MAKING INTERACTIVE GUIDES MORE ATTRACTIVE Anton Nijholt Department of Computer Science University of Twente, Enschede, the Netherlands anijholt@cs.utwente.nl Abstract We investigate the different roads
More informationhit), and assume that longer incidental sounds (forest noise, water, wind noise) resemble a Gaussian noise distribution.
CS 229 FINAL PROJECT A SOUNDHOUND FOR THE SOUNDS OF HOUNDS WEAKLY SUPERVISED MODELING OF ANIMAL SOUNDS ROBERT COLCORD, ETHAN GELLER, MATTHEW HORTON Abstract: We propose a hybrid approach to generating
More informationJoint bottom-up/top-down machine learning structures to simulate human audition and musical creativity
Joint bottom-up/top-down machine learning structures to simulate human audition and musical creativity Jonas Braasch Director of Operations, Professor, School of Architecture Rensselaer Polytechnic Institute,
More informationAutomatic Laughter Detection
Automatic Laughter Detection Mary Knox 1803707 knoxm@eecs.berkeley.edu December 1, 006 Abstract We built a system to automatically detect laughter from acoustic features of audio. To implement the system,
More informationAcoustic Scene Classification
Acoustic Scene Classification Marc-Christoph Gerasch Seminar Topics in Computer Music - Acoustic Scene Classification 6/24/2015 1 Outline Acoustic Scene Classification - definition History and state of
More informationOn human capability and acoustic cues for discriminating singing and speaking voices
Alma Mater Studiorum University of Bologna, August 22-26 2006 On human capability and acoustic cues for discriminating singing and speaking voices Yasunori Ohishi Graduate School of Information Science,
More informationThis full text version, available on TeesRep, is the post-print (final version prior to publication) of:
This full text version, available on TeesRep, is the post-print (final version prior to publication) of: Charles, F. et. al. (2007) 'Affective interactive narrative in the CALLAS Project', 4th international
More informationRetrieval of textual song lyrics from sung inputs
INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Retrieval of textual song lyrics from sung inputs Anna M. Kruspe Fraunhofer IDMT, Ilmenau, Germany kpe@idmt.fraunhofer.de Abstract Retrieving the
More informationA METHOD OF MORPHING SPECTRAL ENVELOPES OF THE SINGING VOICE FOR USE WITH BACKING VOCALS
A METHOD OF MORPHING SPECTRAL ENVELOPES OF THE SINGING VOICE FOR USE WITH BACKING VOCALS Matthew Roddy Dept. of Computer Science and Information Systems, University of Limerick, Ireland Jacqueline Walker
More information1. Introduction NCMMSC2009
NCMMSC9 Speech-to-Singing Synthesis System: Vocal Conversion from Speaking Voices to Singing Voices by Controlling Acoustic Features Unique to Singing Voices * Takeshi SAITOU 1, Masataka GOTO 1, Masashi
More informationSemi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis
Semi-automated extraction of expressive performance information from acoustic recordings of piano music Andrew Earis Outline Parameters of expressive piano performance Scientific techniques: Fourier transform
More informationPOLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING
POLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING Luis Gustavo Martins Telecommunications and Multimedia Unit INESC Porto Porto, Portugal lmartins@inescporto.pt Juan José Burred Communication
More informationSinger Traits Identification using Deep Neural Network
Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic
More informationAUTOMASHUPPER: AN AUTOMATIC MULTI-SONG MASHUP SYSTEM
AUTOMASHUPPER: AN AUTOMATIC MULTI-SONG MASHUP SYSTEM Matthew E. P. Davies, Philippe Hamel, Kazuyoshi Yoshii and Masataka Goto National Institute of Advanced Industrial Science and Technology (AIST), Japan
More informationHBI Database. Version 2 (User Manual)
HBI Database Version 2 (User Manual) St-Petersburg, Russia 2007 2 1. INTRODUCTION...3 2. RECORDING CONDITIONS...6 2.1. EYE OPENED AND EYE CLOSED CONDITION....6 2.2. VISUAL CONTINUOUS PERFORMANCE TASK...6
More informationDERIVING A TIMBRE SPACE FOR THREE TYPES OF COMPLEX TONES VARYING IN SPECTRAL ROLL-OFF
DERIVING A TIMBRE SPACE FOR THREE TYPES OF COMPLEX TONES VARYING IN SPECTRAL ROLL-OFF William L. Martens 1, Mark Bassett 2 and Ella Manor 3 Faculty of Architecture, Design and Planning University of Sydney,
More informationAN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY
AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY Eugene Mikyung Kim Department of Music Technology, Korea National University of Arts eugene@u.northwestern.edu ABSTRACT
More informationVOCALISTENER: A SINGING-TO-SINGING SYNTHESIS SYSTEM BASED ON ITERATIVE PARAMETER ESTIMATION
VOCALISTENER: A SINGING-TO-SINGING SYNTHESIS SYSTEM BASED ON ITERATIVE PARAMETER ESTIMATION Tomoyasu Nakano Masataka Goto National Institute of Advanced Industrial Science and Technology (AIST), Japan
More informationTopics in Computer Music Instrument Identification. Ioanna Karydi
Topics in Computer Music Instrument Identification Ioanna Karydi Presentation overview What is instrument identification? Sound attributes & Timbre Human performance The ideal algorithm Selected approaches
More informationAudiovisual analysis of relations between laughter types and laughter motions
Speech Prosody 16 31 May - 3 Jun 216, Boston, USA Audiovisual analysis of relations between laughter types and laughter motions Carlos Ishi 1, Hiroaki Hata 1, Hiroshi Ishiguro 1 1 ATR Hiroshi Ishiguro
More informationAppendix C ACCESSIBILITY SCALE CLOSED OPEN
Appendix C ACCESSIBILITY SCALE CLOSED OPEN Scale Point: 1. Closed: Jaw clenched Blank facial expression No smiles Tears if present, are choked Nearly silent Eyes cast down or eyes closed Body and face,
More informationProposal for Application of Speech Techniques to Music Analysis
Proposal for Application of Speech Techniques to Music Analysis 1. Research on Speech and Music Lin Zhong Dept. of Electronic Engineering Tsinghua University 1. Goal Speech research from the very beginning
More informationSinging voice synthesis based on deep neural networks
INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Singing voice synthesis based on deep neural networks Masanari Nishimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda
More informationSpeech Recognition and Signal Processing for Broadcast News Transcription
2.2.1 Speech Recognition and Signal Processing for Broadcast News Transcription Continued research and development of a broadcast news speech transcription system has been promoted. Universities and researchers
More informationLaughter Among Deaf Signers
Laughter Among Deaf Signers Robert R. Provine University of Maryland, Baltimore County Karen Emmorey San Diego State University The placement of laughter in the speech of hearing individuals is not random
More informationTempo and Beat Analysis
Advanced Course Computer Science Music Processing Summer Term 2010 Meinard Müller, Peter Grosche Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Tempo and Beat Analysis Musical Properties:
More informationADX TRAX. Insights on Audio Source Separation. Audio Praxis. Rick Silva
ADX TRAX Insights on Audio Source Separation Audionamix has served the professional audio community for years, lending expertise and creative solutions to individuals and companies at the forefront of
More informationEfficient Computer-Aided Pitch Track and Note Estimation for Scientific Applications. Matthias Mauch Chris Cannam György Fazekas
Efficient Computer-Aided Pitch Track and Note Estimation for Scientific Applications Matthias Mauch Chris Cannam György Fazekas! 1 Matthias Mauch, Chris Cannam, George Fazekas Problem Intonation in Unaccompanied
More informationtechnical note flicker measurement display & lighting measurement
technical note flicker measurement display & lighting measurement Contents 1 Introduction... 3 1.1 Flicker... 3 1.2 Flicker images for LCD displays... 3 1.3 Causes of flicker... 3 2 Measuring high and
More informationIEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS 1. Automated Laughter Detection from Full-Body Movements
IEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS 1 Automated Laughter Detection from Full-Body Movements Radoslaw Niewiadomski, Maurizio Mancini, Giovanna Varni, Gualtiero Volpe, and Antonio Camurri Abstract
More informationInstrument Recognition in Polyphonic Mixtures Using Spectral Envelopes
Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu
More informationSupervised Musical Source Separation from Mono and Stereo Mixtures based on Sinusoidal Modeling
Supervised Musical Source Separation from Mono and Stereo Mixtures based on Sinusoidal Modeling Juan José Burred Équipe Analyse/Synthèse, IRCAM burred@ircam.fr Communication Systems Group Technische Universität
More informationImplementing and Evaluating a Laughing Virtual Character
Implementing and Evaluating a Laughing Virtual Character MAURIZIO MANCINI, DIBRIS, University of Genoa, Italy BEATRICE BIANCARDI and FLORIAN PECUNE, CNRS-LTCI, Télécom-ParisTech, France GIOVANNA VARNI,
More informationExpressive Singing Synthesis based on Unit Selection for the Singing Synthesis Challenge 2016
Expressive Singing Synthesis based on Unit Selection for the Singing Synthesis Challenge 2016 Jordi Bonada, Martí Umbert, Merlijn Blaauw Music Technology Group, Universitat Pompeu Fabra, Spain jordi.bonada@upf.edu,
More informationSOUND LABORATORY LING123: SOUND AND COMMUNICATION
SOUND LABORATORY LING123: SOUND AND COMMUNICATION In this assignment you will be using the Praat program to analyze two recordings: (1) the advertisement call of the North American bullfrog; and (2) the
More informationAutomatic music transcription
Music transcription 1 Music transcription 2 Automatic music transcription Sources: * Klapuri, Introduction to music transcription, 2006. www.cs.tut.fi/sgn/arg/klap/amt-intro.pdf * Klapuri, Eronen, Astola:
More informationPattern Recognition in Music
Pattern Recognition in Music SAMBA/07/02 Line Eikvil Ragnar Bang Huseby February 2002 Copyright Norsk Regnesentral NR-notat/NR Note Tittel/Title: Pattern Recognition in Music Dato/Date: February År/Year:
More informationAutomatic Labelling of tabla signals
ISMIR 2003 Oct. 27th 30th 2003 Baltimore (USA) Automatic Labelling of tabla signals Olivier K. GILLET, Gaël RICHARD Introduction Exponential growth of available digital information need for Indexing and
More informationSinging Pitch Extraction and Singing Voice Separation
Singing Pitch Extraction and Singing Voice Separation Advisor: Jyh-Shing Roger Jang Presenter: Chao-Ling Hsu Multimedia Information Retrieval Lab (MIR) Department of Computer Science National Tsing Hua
More informationThe laughing brain - Do only humans laugh?
The laughing brain - Do only humans laugh? Martin Meyer Institute of Neuroradiology University Hospital of Zurich Aspects of laughter Humour, sarcasm, irony privilege to adolescents and adults children
More informationSpeaking loud, speaking high: non-linearities in voice strength and vocal register variations. Christophe d Alessandro LIMSI-CNRS Orsay, France
Speaking loud, speaking high: non-linearities in voice strength and vocal register variations Christophe d Alessandro LIMSI-CNRS Orsay, France 1 Content of the talk Introduction: voice quality 1. Voice
More informationPitch. The perceptual correlate of frequency: the perceptual dimension along which sounds can be ordered from low to high.
Pitch The perceptual correlate of frequency: the perceptual dimension along which sounds can be ordered from low to high. 1 The bottom line Pitch perception involves the integration of spectral (place)
More informationSINGING EXPRESSION TRANSFER FROM ONE VOICE TO ANOTHER FOR A GIVEN SONG. Sangeon Yong, Juhan Nam
SINGING EXPRESSION TRANSFER FROM ONE VOICE TO ANOTHER FOR A GIVEN SONG Sangeon Yong, Juhan Nam Graduate School of Culture Technology, KAIST {koragon2, juhannam}@kaist.ac.kr ABSTRACT We present a vocal
More informationThursday, April 28, 16
Drama Unit Learning Targets I can analyze the development of a theme over the course of a text. I can analyze how a drama s form or structure contributes to its meaning. I can compare and contrast a written
More informationExpressive Multimodal Conversational Acts for SAIBA agents
Expressive Multimodal Conversational Acts for SAIBA agents Jeremy Riviere 1, Carole Adam 1, Sylvie Pesty 1, Catherine Pelachaud 2, Nadine Guiraud 3, Dominique Longin 3, and Emiliano Lorini 3 1 Grenoble
More informationComposer Identification of Digital Audio Modeling Content Specific Features Through Markov Models
Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Aric Bartle (abartle@stanford.edu) December 14, 2012 1 Background The field of composer recognition has
More informationTOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC
TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu
More informationThe Language of Film and TV
The Language of Film and TV Summary - 1. Cinematographic Language -What cinema is -1.1 Parts of the Cinematographic language -1.2 Camera Movements -1.3 Camera Angles -1.4 Narrative structure of a film
More informationPitch Perception and Grouping. HST.723 Neural Coding and Perception of Sound
Pitch Perception and Grouping HST.723 Neural Coding and Perception of Sound Pitch Perception. I. Pure Tones The pitch of a pure tone is strongly related to the tone s frequency, although there are small
More informationAcoustic Analysis of Beethoven Piano Sonata Op.110. Yan-bing DING and Qiu-hua HUANG
2016 International Conference on Advanced Materials Science and Technology (AMST 2016) ISBN: 978-1-60595-397-7 Acoustic Analysis of Beethoven Piano Sonata Op.110 Yan-bing DING and Qiu-hua HUANG Key Lab
More informationPitch Analysis of Ukulele
American Journal of Applied Sciences 9 (8): 1219-1224, 2012 ISSN 1546-9239 2012 Science Publications Pitch Analysis of Ukulele 1, 2 Suphattharachai Chomphan 1 Department of Electrical Engineering, Faculty
More informationMELODY EXTRACTION FROM POLYPHONIC AUDIO OF WESTERN OPERA: A METHOD BASED ON DETECTION OF THE SINGER S FORMANT
MELODY EXTRACTION FROM POLYPHONIC AUDIO OF WESTERN OPERA: A METHOD BASED ON DETECTION OF THE SINGER S FORMANT Zheng Tang University of Washington, Department of Electrical Engineering zhtang@uw.edu Dawn
More informationLab 5 Linear Predictive Coding
Lab 5 Linear Predictive Coding 1 of 1 Idea When plain speech audio is recorded and needs to be transmitted over a channel with limited bandwidth it is often necessary to either compress or encode the audio
More informationSemi-supervised Musical Instrument Recognition
Semi-supervised Musical Instrument Recognition Master s Thesis Presentation Aleksandr Diment 1 1 Tampere niversity of Technology, Finland Supervisors: Adj.Prof. Tuomas Virtanen, MSc Toni Heittola 17 May
More informationTowards automated full body detection of laughter driven by human expert annotation
2013 Humaine Association Conference on Affective Computing and Intelligent Interaction Towards automated full body detection of laughter driven by human expert annotation Maurizio Mancini, Jennifer Hofmann,
More informationSMS Composer and SMS Conductor: Applications for Spectral Modeling Synthesis Composition and Performance
SMS Composer and SMS Conductor: Applications for Spectral Modeling Synthesis Composition and Performance Eduard Resina Audiovisual Institute, Pompeu Fabra University Rambla 31, 08002 Barcelona, Spain eduard@iua.upf.es
More informationBasic Operations App Guide
Basic Operations App Guide Table of Contents 1. Outline 2. Items to Prepare 3. User Registration 4. Login 5. Connect Camera 6. Change or Delete Camera Name 7. Customer Analysis 7.1 Customer Analysis Main
More informationEffects of acoustic degradations on cover song recognition
Signal Processing in Acoustics: Paper 68 Effects of acoustic degradations on cover song recognition Julien Osmalskyj (a), Jean-Jacques Embrechts (b) (a) University of Liège, Belgium, josmalsky@ulg.ac.be
More informationSupervised Learning in Genre Classification
Supervised Learning in Genre Classification Introduction & Motivation Mohit Rajani and Luke Ekkizogloy {i.mohit,luke.ekkizogloy}@gmail.com Stanford University, CS229: Machine Learning, 2009 Now that music
More informationAutomatic music transcription
Educational Multimedia Application- Specific Music Transcription for Tutoring An applicationspecific, musictranscription approach uses a customized human computer interface to combine the strengths of
More informationCHAPTER 4 SEGMENTATION AND FEATURE EXTRACTION
69 CHAPTER 4 SEGMENTATION AND FEATURE EXTRACTION According to the overall architecture of the system discussed in Chapter 3, we need to carry out pre-processing, segmentation and feature extraction. This
More informationEngineering in Recording
Engineering in Recording Jez Wells Audio Lab - Department of Electronics University of York, Heslington, York, UK ISEE, University of Sheffield, July 2012 1 Overview Is Recording Engineering? Some definitions
More informationAN ALGORITHM FOR LOCATING FUNDAMENTAL FREQUENCY (F0) MARKERS IN SPEECH
AN ALGORITHM FOR LOCATING FUNDAMENTAL FREQUENCY (F0) MARKERS IN SPEECH by Princy Dikshit B.E (C.S) July 2000, Mangalore University, India A Thesis Submitted to the Faculty of Old Dominion University in
More informationA New Method for Calculating Music Similarity
A New Method for Calculating Music Similarity Eric Battenberg and Vijay Ullal December 12, 2006 Abstract We introduce a new technique for calculating the perceived similarity of two songs based on their
More informationA Matlab toolbox for. Characterisation Of Recorded Underwater Sound (CHORUS) USER S GUIDE
Centre for Marine Science and Technology A Matlab toolbox for Characterisation Of Recorded Underwater Sound (CHORUS) USER S GUIDE Version 5.0b Prepared for: Centre for Marine Science and Technology Prepared
More informationMultimodal Data Collection of Human-Robot Humorous Interactions in the JOKER Project
Multimodal Data Collection of Human-Robot Humorous Interactions in the JOKER Project Laurence Devillers, Sophie Rosset, Guillaume Dubuisson Duplessis, Mohamed A. Sehili, Lucile Béchade, Agnès Delaborde,
More information