An action based metaphor for description of expression in music performance

Similar documents
Action and expression in music performance

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

Expressive information

Nature Neuroscience: doi: /nn Supplementary Figure 1. Emergence of dmpfc and BLA 4-Hz oscillations during freezing behavior.

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

Modeling and Control of Expressiveness in Music Performance

MOTIVATION AGENDA MUSIC, EMOTION, AND TIMBRE CHARACTERIZING THE EMOTION OF INDIVIDUAL PIANO AND OTHER MUSICAL INSTRUMENT SOUNDS

THE EFFECT OF EXPERTISE IN EVALUATING EMOTIONS IN MUSIC

The Role of Time in Music Emotion Recognition: Modeling Musical Emotions from Time-Varying Music Features

Analysis, Synthesis, and Perception of Musical Sounds

Perceptual Evaluation of Automatically Extracted Musical Motives

Supervised Learning in Genre Classification

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

Toward a Computationally-Enhanced Acoustic Grand Piano

Towards a multi-layer architecture for multi-modal rendering of expressive actions

Topic 10. Multi-pitch Analysis

Problem Weight Score Total 100

Digital music synthesis using DSP

Music Mood Classification - an SVM based approach. Sebastian Napiorkowski

MELODY ANALYSIS FOR PREDICTION OF THE EMOTIONS CONVEYED BY SINHALA SONGS

HOW COOL IS BEBOP JAZZ? SPONTANEOUS

Embodied music cognition and mediation technology

2. AN INTROSPECTION OF THE MORPHING PROCESS

Lecture 10 Harmonic/Percussive Separation

CSC475 Music Information Retrieval

Automatic Music Clustering using Audio Attributes

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

Transcription An Historical Overview

Automatic Piano Music Transcription

MUSI-6201 Computational Music Analysis

Implementation of 24P, 25P and 30P Segmented Frames for Production Format

Classification of Timbre Similarity

A User-Oriented Approach to Music Information Retrieval.

A Categorical Approach for Recognizing Emotional Effects of Music

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

LEARNING SPECTRAL FILTERS FOR SINGLE- AND MULTI-LABEL CLASSIFICATION OF MUSICAL INSTRUMENTS. Patrick Joseph Donnelly

Acoustic Prosodic Features In Sarcastic Utterances

GCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam

Modeling expressiveness in music performance

Chapter 5 Sequential Circuits

CTP431- Music and Audio Computing Musical Acoustics. Graduate School of Culture Technology KAIST Juhan Nam

An interdisciplinary approach to audio effect classification

Department of Electrical & Electronic Engineering Imperial College of Science, Technology and Medicine. Project: Real-Time Speech Enhancement

MODELING MUSICAL MOOD FROM AUDIO FEATURES AND LISTENING CONTEXT ON AN IN-SITU DATA SET

Semi-supervised Musical Instrument Recognition

MUSICAL MOODS: A MASS PARTICIPATION EXPERIMENT FOR AFFECTIVE CLASSIFICATION OF MUSIC

AUDIO/VISUAL INDEPENDENT COMPONENTS

A DISCRETE FILTER BANK APPROACH TO AUDIO TO SCORE MATCHING FOR POLYPHONIC MUSIC

Perceptual Mixing for Musical Production

From quantitative empirï to musical performology: Experience in performance measurements and analyses

Problem Weight Total 100

Simple Harmonic Motion: What is a Sound Spectrum?

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

Experiments on musical instrument separation using multiplecause

Research & Development. White Paper WHP 228. Musical Moods: A Mass Participation Experiment for the Affective Classification of Music

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

Perceptual dimensions of short audio clips and corresponding timbre features

into a Cognitive Architecture

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

Expressive performance in music: Mapping acoustic cues onto facial expressions

Physical Modelling of Musical Instruments Using Digital Waveguides: History, Theory, Practice

Sequential Circuits. Output depends only and immediately on the inputs Have no memory (dependence on past values of the inputs)

A NON LINEAR APPROACH TOWARDS AUTOMATED EMOTION ANALYSIS IN HINDUSTANI MUSIC

Music, Timbre and Time

Bunch-by-bunch feedback and LLRF at ELSA

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

ANALYSIS of MUSIC PERFORMED IN DIFFERENT ACOUSTIC SETTINGS in STAVANGER CONCERT HOUSE

Application of Pattern Recognition Method in a Linguistic Experiment with Unsupervised Classification

THE importance of music content analysis for musical

The Human Features of Music.

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

A CLASSIFICATION-BASED POLYPHONIC PIANO TRANSCRIPTION APPROACH USING LEARNED FEATURE REPRESENTATIONS

Automatic Construction of Synthetic Musical Instruments and Performers

Interface Practices Subcommittee SCTE STANDARD SCTE Measurement Procedure for Noise Power Ratio

Colour Reproduction Performance of JPEG and JPEG2000 Codecs

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

Single Channel Speech Enhancement Using Spectral Subtraction Based on Minimum Statistics

Perception and Sound Design

A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES

TongArk: a Human-Machine Ensemble

Analysis and Clustering of Musical Compositions using Melody-based Features

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

TYING SEMANTIC LABELS TO COMPUTATIONAL DESCRIPTORS OF SIMILAR TIMBRES

Hong Kong University of Science and Technology 2 The Information Systems Technology and Design Pillar,

ON FINDING MELODIC LINES IN AUDIO RECORDINGS. Matija Marolt

Sofia Dahl Cognitive and Systematic Musicology Lab, School of Music. Looking at movement gesture Examples from drumming and percussion Sofia Dahl

Music Information Retrieval with Temporal Features and Timbre

CONCATENATIVE SYNTHESIS FOR NOVEL TIMBRAL CREATION. A Thesis. presented to. the Faculty of California Polytechnic State University, San Luis Obispo

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

EIGHTH GRADE RELIGION

CHAPTER 4 SEGMENTATION AND FEATURE EXTRACTION

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

Module 8 : Numerical Relaying I : Fundamentals

MPEG-7 AUDIO SPECTRUM BASIS AS A SIGNATURE OF VIOLIN SOUND

Learning Joint Statistical Models for Audio-Visual Fusion and Segregation

Audio-Based Video Editing with Two-Channel Microphone

Music Segmentation Using Markov Chain Methods

A Recipe for Emotion in Music (Music & Meaning Part II)

Transcription:

An action based metaphor for description of expression in music performance Luca Mion CSC-SMC, Centro di Sonologia Computazionale Department of Information Engineering University of Padova Workshop Toni Mian, October 26, 2007

Why modelling expressiveness Understanding human communication strategies Non-verbal communication Expressiveness tells how to take the explicit message Disambiguate language expressions (e.g. in a movie) To embody expressive knowledge in machines To adapt HCI to the basic forms of human behaviour Expression through the auditory channel Artistic applications, games and entertainment Auditory Display Music Information Retrieval Education Luca Mion, Workshop Toni Mian October 26, 2007 2 / 15

Why modelling expressiveness Understanding human communication strategies Non-verbal communication Expressiveness tells how to take the explicit message Disambiguate language expressions (e.g. in a movie) To embody expressive knowledge in machines To adapt HCI to the basic forms of human behaviour Expression through the auditory channel Artistic applications, games and entertainment Auditory Display Music Information Retrieval Education Luca Mion, Workshop Toni Mian October 26, 2007 2 / 15

Why modelling expressiveness Understanding human communication strategies Non-verbal communication Expressiveness tells how to take the explicit message Disambiguate language expressions (e.g. in a movie) To embody expressive knowledge in machines To adapt HCI to the basic forms of human behaviour Expression through the auditory channel Artistic applications, games and entertainment Auditory Display Music Information Retrieval Education Luca Mion, Workshop Toni Mian October 26, 2007 2 / 15

Why modelling expressiveness Understanding human communication strategies Non-verbal communication Expressiveness tells how to take the explicit message Disambiguate language expressions (e.g. in a movie) To embody expressive knowledge in machines To adapt HCI to the basic forms of human behaviour Expression through the auditory channel Artistic applications, games and entertainment Auditory Display Music Information Retrieval Education Luca Mion, Workshop Toni Mian October 26, 2007 2 / 15

Expression in music performance Expressive intention is a broad concept: emotions affective domain metaphoric aspects sensory domain We want to describe expression at an intermediate level between sound as waveform and music as language yet another metaphor? we propose an interpretation of the expressive intentions based on action metaphor and on ideal physical systems Luca Mion, Workshop Toni Mian October 26, 2007 3 / 15

Expression in music performance Expressive intention is a broad concept: emotions affective domain metaphoric aspects sensory domain We want to describe expression at an intermediate level between sound as waveform and music as language yet another metaphor? we propose an interpretation of the expressive intentions based on action metaphor and on ideal physical systems Luca Mion, Workshop Toni Mian October 26, 2007 3 / 15

Objectives Exploring similarities of affective and sensorial expressive intentions Organization of the feature space Interpretation of affective and sensorial expressions clustering Perceptual experiment to study whether/how the listeners cluster expressive intentions Luca Mion, Workshop Toni Mian October 26, 2007 4 / 15

Methodology Feature selection and validation: Sequential Forward Selection (SFS) and Minimum Distance Classifier Projection on a 2D space by Principal Component Analysis (PCA) Interpretation of the expression clusters (k-means clustering) Perceptual experiment, Pearson s χ 2 test, Correspondence Analysis Audio data: performances from Flute, Violin and Guitar (repeated notes, scales, excerpts) Luca Mion, Workshop Toni Mian October 26, 2007 5 / 15

The affective and sensorial domains Affective Domain: the Valence/Arousal space Happy, Sad, Angry, Calm Sensorial Domain: the Kinetics/Energy space Hard, Soft, Heavy, Light Neutral performance: with no artistic aims Angry AROUSAL Happy Heavy Hard ENERGY Sad VALENCE KINETICS Light Calm Soft Luca Mion, Workshop Toni Mian October 26, 2007 6 / 15

The affective and sensorial domains Affective Domain: the Valence/Arousal space Happy, Sad, Angry, Calm Sensorial Domain: the Kinetics/Energy space Hard, Soft, Heavy, Light Neutral performance: with no artistic aims Angry AROUSAL Happy Heavy Hard ENERGY Sad VALENCE KINETICS Light Calm Soft Luca Mion, Workshop Toni Mian October 26, 2007 6 / 15

The affective and sensorial domains Affective Domain: the Valence/Arousal space Happy, Sad, Angry, Calm Sensorial Domain: the Kinetics/Energy space Hard, Soft, Heavy, Light Neutral performance: with no artistic aims Angry AROUSAL Happy Heavy Hard ENERGY Sad VALENCE KINETICS Light Calm Soft Luca Mion, Workshop Toni Mian October 26, 2007 6 / 15

Audio features for sound description Local audio features (on 46 ms-length non overlapping frames) Roughness R (cochlear filter-bank, texture perception) Spectral Ratio SR a = j LB X(j) 2 / N/2 1 k=1 X(k) 2 indicates the amount of energy in the low frequency band LB (f < 1 khz) Residual Energy ratio RE h = j HB X R(j) 2 / N/2 1 k=1 X(k) 2 describes the stochastic energy in the high frequency band HB (f > 1.8 khz) Event audio features (on 4s-duration and 3.5s-overlapping frames) Peak Sound Level PSL Notes per Second NPS Attack time A Luca Mion, Workshop Toni Mian October 26, 2007 7 / 15

Representing expressions in the feature space k-means algorithm for unsupervised clustering (cosine distance metric) d = 1 cos(x,y) = 1 < x,y > / x y 0.5 0.4 Cluster analysis on violin performances hard heavy Second Principal Component 0.3 0.2 0.1 0-0.1-0.2-0.3 angry Elasticity Friction Inertia light happy neutral soft -0.4-0.8-0.6-0.4-0.2 0 0.2 0.4 0.6 0.8 First Principal Component calm sad Luca Mion, Workshop Toni Mian October 26, 2007 8 / 15

Interpretation of the joint space Using the physical analogy: Force is considered as the cause, movement (velocity or position) as the effect Cause-effect relationship: represented by the admittance operator Y: v(t) = t f (τ) y(t τ)dτ It describes the dynamic mapping and the qualitative behavior from force to velocity by an integral-differential equation. Luca Mion, Workshop Toni Mian October 26, 2007 9 / 15

Admittance as metaphor f(t) cause force Y velocity effect t Damping f(t) v(t) v(t) t Friction Mass f(t) v(t) m v(t) t Inertia Spring f(t) k v(t) t Elasticity correspondences (low) Friction: Angry - Hard, Heavy Elasticity: Happy - Light Inertia: Sad, Calm - Soft Luca Mion, Workshop Toni Mian October 26, 2007 10 / 15

Perceptual test We want to find how the listeners cluster expressive intentions. 16 participants (musicians and not musicians) 9 expressive intentions (4 emotional, 4 sensory plus neutral) 3 performances representative for our clusters Representative adjectives (minimizing the cosine distance): hard for friction, happy for elasticity, calm for inertia. Task: associate to one representative the expressive intention that the performer wants to convey. Luca Mion, Workshop Toni Mian October 26, 2007 11 / 15

Perceptual test We want to find how the listeners cluster expressive intentions. 16 participants (musicians and not musicians) 9 expressive intentions (4 emotional, 4 sensory plus neutral) 3 performances representative for our clusters Representative adjectives (minimizing the cosine distance): hard for friction, happy for elasticity, calm for inertia. Task: associate to one representative the expressive intention that the performer wants to convey. Luca Mion, Workshop Toni Mian October 26, 2007 11 / 15

Results from perceptual test Pearson s χ 2 test: compared observed frequencies with expected frequencies strong dependence between clusters and adjectives χ 2 = 401.99 (critical value = 36.12, α=0.001, df=14) neutral performance not categorized, as expected by the performer Luca Mion, Workshop Toni Mian October 26, 2007 12 / 15

Results from perceptual test (contd.) Simple Correspondence Analysis on the contingency table: 2 factors (74.19% and the 25.81% of the total inertia) agreement with k-means clustering on the feature space three stable groups hard, happy, calm near to the centroid of respective cluster (distances 0.088, 0.214, 0.199 respectively) intrinsic similarity of affective and sensory expressive intentions Correspondence analysis plot (violin) Dimension 2 2.0 1.5 1.0 0.5 0.0 hard FRICTION angry heavy soft neutral -0.5-1.0 ELASTICITY light happy INERTIA calm sad -2-1 0 1 2 Dimension 1 Luca Mion, Workshop Toni Mian October 26, 2007 13 / 15

Conclusions Three clusters emerged, both in the feature space and in the subjects evaluation Agreement between clusters (k-means clustering) listener evaluation in agreement with perfomer s intention Correspondence of affective and sensorial expressive intentions Perceptual tests with visual and haptic representatives Can action metaphor be extended to other gesture based arts, such as dance, drawing etc.? Luca Mion, Workshop Toni Mian October 26, 2007 14 / 15

Thanks for your attention. Sound and Music Computing Group http://smc.dei.unipd.it luca.mion@dei.unipd.it