Measuring & Modeling Musical Expression

Similar documents
Director Musices: The KTH Performance Rules System

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

A prototype system for rule-based expressive modifications of audio recordings

On time: the influence of tempo, structure and style on the timing of grace notes in skilled musical performance

Temporal coordination in string quartet performance

A Computational Model for Discriminating Music Performers

Tempo and Beat Analysis

Quarterly Progress and Status Report. Perception of just noticeable time displacement of a tone presented in a metrical sequence at different tempos

Machine Learning Term Project Write-up Creating Models of Performers of Chopin Mazurkas

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis

From quantitative empirï to musical performology: Experience in performance measurements and analyses

Computer Coordination With Popular Music: A New Research Agenda 1

Machine Learning of Expressive Microtiming in Brazilian and Reggae Drumming Matt Wright (Music) and Edgar Berdahl (EE), CS229, 16 December 2005

A Case Based Approach to the Generation of Musical Expression

Research Projects. Measuring music similarity and recommending music. Douglas Eck Research Statement 2

The influence of musical context on tempo rubato. Renee Timmers, Richard Ashley, Peter Desain, Hank Heijink

Introductions to Music Information Retrieval

Computational Modelling of Harmony

Music Representations

Multidimensional analysis of interdependence in a string quartet

Making music with voice. Distinguished lecture, CIRMMT Jan 2009, Copyright Johan Sundberg

ESP: Expression Synthesis Project

Importance of Note-Level Control in Automatic Music Performance

A Bayesian Network for Real-Time Musical Accompaniment

Structural Communication

LESSON 1 PITCH NOTATION AND INTERVALS

Playing Mozart by Analogy: Learning Multi-level Timing and Dynamics Strategies

Finger motion in piano performance: Touch and tempo

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

On the contextual appropriateness of performance rules

Human Preferences for Tempo Smoothness

Quarterly Progress and Status Report. Matching the rule parameters of PHRASE ARCH to performances of Träumerei : a preliminary study

Quarterly Progress and Status Report. Musicians and nonmusicians sensitivity to differences in music performance

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

Robert Alexandru Dobre, Cristian Negrescu

Music Radar: A Web-based Query by Humming System

Data Driven Music Understanding

A PRELIMINARY COMPUTATIONAL MODEL OF IMMANENT ACCENT SALIENCE IN TONAL MUSIC

WHO IS WHO IN THE END? RECOGNIZING PIANISTS BY THEIR FINAL RITARDANDI

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm

Week 14 Music Understanding and Classification

Classification of Dance Music by Periodicity Patterns

MUSI-6201 Computational Music Analysis

HYBRID NUMERIC/RANK SIMILARITY METRICS FOR MUSICAL PERFORMANCE ANALYSIS

Experiments on gestures: walking, running, and hitting

Widmer et al.: YQX Plays Chopin 12/03/2012. Contents. IntroducAon Expressive Music Performance How YQX Works Results

Musical Entrainment Subsumes Bodily Gestures Its Definition Needs a Spatiotemporal Dimension

A Beat Tracking System for Audio Signals

In Search of the Horowitz Factor

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

TEMPO AND BEAT are well-defined concepts in the PERCEPTUAL SMOOTHNESS OF TEMPO IN EXPRESSIVELY PERFORMED MUSIC

CS229 Project Report Polyphonic Piano Transcription

Lab P-6: Synthesis of Sinusoidal Signals A Music Illusion. A k cos.! k t C k / (1)

A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS

Acoustic and musical foundations of the speech/song illusion

String Orchestra - 7th Grade

jsymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada

Grade HS Band (1) Basic

A cross-cultural comparison study of the production of simple rhythmic patterns

Artificial Social Composition: A Multi-Agent System for Composing Music Performances by Emotional Communication

Investigations of Between-Hand Synchronization in Magaloff s Chopin

COMPUTATIONAL INVESTIGATIONS INTO BETWEEN-HAND SYNCHRONIZATION IN PIANO PLAYING: MAGALOFF S COMPLETE CHOPIN

Unobtrusive practice tools for pianists

PRESCOTT UNIFIED SCHOOL DISTRICT District Instructional Guide January 2016

Connecticut State Department of Education Music Standards Middle School Grades 6-8

ESTIMATING THE ERROR DISTRIBUTION OF A TAP SEQUENCE WITHOUT GROUND TRUTH 1

Instrumental Performance Band 7. Fine Arts Curriculum Framework

Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies

Temporal Coordination and Adaptation to Rate Change in Music Performance

Music Performance Panel: NICI / MMM Position Statement

Perceiving temporal regularity in music

MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC

Estimating the Time to Reach a Target Frequency in Singing

Elements of Music. How can we tell music from other sounds?

Music Curriculum Glossary

Copyright 2009 Pearson Education, Inc. or its affiliate(s). All rights reserved. NES, the NES logo, Pearson, the Pearson logo, and National

TOWARDS AUTOMATED EXTRACTION OF TEMPO PARAMETERS FROM EXPRESSIVE MUSIC RECORDINGS

The Relationship Between Auditory Imagery and Musical Synchronization Abilities in Musicians

Music Representations. Beethoven, Bach, and Billions of Bytes. Music. Research Goals. Piano Roll Representation. Player Piano (1900)

Evaluation of the Audio Beat Tracking System BeatRoot

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception

SOA PIANO ENTRANCE AUDITIONS FOR 6 TH - 12 TH GRADE

A Case Based Approach to Expressivity-aware Tempo Transformation

Computational Models of Expressive Music Performance: The State of the Art

Goebl, Pampalk, Widmer: Exploring Expressive Performance Trajectories. Werner Goebl, Elias Pampalk and Gerhard Widmer (2004) Introduction

Introduction. Figure 1: A training example and a new problem.

Perceptual Smoothness of Tempo in Expressively Performed Music

Notes on David Temperley s What s Key for Key? The Krumhansl-Schmuckler Key-Finding Algorithm Reconsidered By Carley Tanoue

CS 591 S1 Computational Audio

Modeling and Control of Expressiveness in Music Performance

Measurement of overtone frequencies of a toy piano and perception of its pitch

CTP431- Music and Audio Computing Music Information Retrieval. Graduate School of Culture Technology KAIST Juhan Nam

Audio Feature Extraction for Corpus Analysis

Timing variations in music performance: Musical communication, perceptual compensation, and/or motor control?

Woodlynne School District Curriculum Guide. General Music Grades 3-4

Analysis of Musical Content in Digital Audio

SPECTRAL LEARNING FOR EXPRESSIVE INTERACTIVE ENSEMBLE MUSIC PERFORMANCE

Transcription:

Measuring & Modeling Musical Expression Douglas Eck University of Montreal Department of Computer Science BRAMS Brain Music and Sound International Laboratory for Brain, Music and Sound Research

Overview Why care about timing and dynamics in music? Previous approaches to measuring timing and dynamics Models which predict something about expression Working without musical scores A correlation-based approach for constructing metrical trees 2

Note-level measures (MIDI) Pitch Velocity Duration IOI (inter-onset interval) KOT (key overlap time) Pedaling (piano)!!"#$%&' ()' *+,' -&."/"0"1/' 1.' "/0&%21/3&0' "/0&%4+5' *676 /,8' 9$%+0"1/'*-: /,'+/9';&<'14&%5+='0">&'*?7@ /,'.1%'@7AB / '.1551C&9'D<'+/'14&%5+=="/#'@7AB /E( )' '*D,'-&."/"0"1/'1.' "/0&%21/3&0' "/0&%4+5' *676 /,8' 9$%+0"1/' *-: /,' +/9' ;&<' 9&0+FG&9' 0">&' *?-@ /,'.1%' @7AB / '.1551C&9' D<' +' /1/2 14&%5+=="/#'@7AB /E( )' From R. Bresin Articulation Rules for Automatic Music Performance 3

Example: Chopin Etude Opus 10 No 3

Example: Chopin Etude Opus 10 No 3 Deadpan (no expressive timing or dynamics) Human performance (Recorded on Boesendorfer ZEUS) Differences limited to: timing (onset, length) velocity (seen as red) pedaling (blue shading)

What can we measure? Repp (1989) measured note IOIs in 19 famous recordings of a Beethoven minuet (Sonata op 31 no 3) 1ooo MINUET?00 600 500.! I I I [ I I [ I I I I I I I I E: 3 q 5 6 7 8 9 18 t! 1 13,tq t5 16 BRR NO. Grand average timing patterns of performances with repeats plotted separately. (From B. Repp Patterns of expressive timing in performances of a Beethoven minuet by nineteen famous pianists,1990) 7

What can we measure? L00-000- FACTOR I PCA analysis yields 2 major components 600- B00- Phrase final lengthening Phrase internal variation 700-600- 500- Simply taking mean IOIs yields can yield pleasing performance Reconstructing using principal component(s) can yield pleasing performance 900-088- I I I I I I I I I E 3 q 5 6 7 8 FACTOR 3 BRR I I I I I I I 9 10 11 1E 13 lq 15 16 NO. Adapted from Repp (1990) Concluded that timing underlies musical structure 708-608- 500- I [ I I I I I I I I I I I I I I I E 3 q 5 6 7 8 9 10 11 IE 13 tq 15 16 BRR NO, 8

Timing versus expressive dynamics Repp (1997; experiment 2): generated MIDI from audio for 15 famous performances of Chopin s op. 10 No 3; Added 9 graduate student performances Retained only timing (no expressive dynamics) Judges ranked the average timing profile of the expert pianists (EA) highest, followed by E11, S1, S3, S9, S2, and SA. Conclusions: EA, SA sound better than average but lack individuality (Repp) Something is lost in discarding non-temporal expressive dynamics. Timing and expressive dynamics may be inter-dependent However, interesting that EA, SA sound good at all 9

KTH Model Johan Sundberg, Anders Friberg, many others Models performance of Western music Rule-based system built using analysis-by-synthesis: assess impact of individual rules by listening analysis-by-measurement: fit rules to performance data Incorporates wide range of music perception research (e.g. meter perception, pitch perception, motor control constraints) 10

Table 1. An overview of the rule system Phrasing Phrase arch Create arch-like tempo and sound level changes over phrases Final ritardando Apply a ritardando in the end of the piece High loud Increase sound level in proportion to pitch height Micro-level timing Duration contrast Shorten relatively short notes and lengthen relatively long notes Faster uphill Increase tempo in rising pitch sequences Metrical patterns and grooves Double duration Decrease duration ratio for two notes with a nominal value of 2:1 Inégales Introduce long-short patterns for equal note values (swing) Articulation Punctuation Find short melodic fragments and mark them with a final micropause Score legato/staccato Articulate legato/staccato when marked in the score Repetition articulation Add articulation for repeated notes. Overall articulation Add articulation for all notes except very short ones Tonal tension Melodic charge Emphasize the melodic tension of notes relatively the current chord Harmonic charge Emphasize the harmonic tension of chords relatively the key Chromatic charge Emphasize regions of small pitch changes Intonation High sharp Stretch all intervals in proportion to size Melodic intonation Intonate according to melodic context Harmonic intonation Intonate according to harmonic context Mixed intonation Intonate using a combination of melodic and harmonic intonation Ensemble timing Melodic sync Synchronize using a new voice containing all relevant onsets Ensemble swing Introduce metrical timing patterns for the instruments in a jazz ensemble Performance noise Noise control Simulate inaccuracies in motor From: A. Friberg, R. Bresin & J. Sundberg (2006). Overview of the KTH rule system for musical performance. Advances in Cognitive Psychology, 2(2-3):145-161.

Figure 2. The resulting IOI deviations by applying Phrase arch, Duration contrast, Melodic charge, and Punctuation to the Swedish nursery tune Ekorr n satt i granen. All rules were applied with the rule quantity k=1 except the Melodic charge rule that was applied with k=2. From: A. Friberg, R. Bresin & J. Sundberg (2006). Overview of the KTH rule system for musical performance. Advances in Cognitive Psychology, 2(2-3): 145-161.

Widmer et al. performance model Automatic deduction of rules for music performance Rich feature set (29 attributes including local melodic contour, scale degree, duration, etc) Performance is matched to score (metrical position). PLCG: Partition Learn Cluster Generalize (Widmer, 2003) Discovery of simple partial rules-based models Inspired by ensemble learning PLCG compares favorably to rule learning algorithm RIPPER Rules learned by PLCG similar to some KTH rules (Widmer 13

RULE TL2: abstract_duration_context = equal-longer & metr_strength 1 ritardando Given two notes of equal duration followed by a longer note, lengthen the note (i.e., play it more slowly) that precedes the final, longer one, if this note is in a metrically weak position ( metrical strength 1). Fig. 5. Mozart Sonata K.331, 1st movement, 1st part, as played by pianist and learner. The curve plots the relative tempo at each note notes above the 1.0 line are shortened relative to the tempo of the piece, notes below 1.0 are lengthened. A perfectly regular performance with no timing deviations would correspond to a straight line at y = 1.0. From: G. Widmer (2003). Discovering simple rules in complex data: A metalearning algorithm and some surprising musical discoveries. Artificial Intelligence 146:129-148.

Music Plus One (C. Raphael) Task I : Listen Inputs: sampled acoustic signal musical score Output: Time at which notes occur Task 2 : Play Inputs: output from Listen module musical score rehearsal data from musician performances of accompaniment Output: Music accompaniment in real time t1 t2 t3 t4 t5 t6 t7 t8 t9 Solo 4 Accompaniment 4 t1 t2 t3 t4 t5 Text and graphics on following pages from slide presentation by Chris Raphael.Thanks Chris! 15

35 30 Five performances of same musical phrase 25 Intuition: there are regularities to be learned 20 seconds 15 10 5 0 0 1 2 3 4 5 6 7 8 measures

Graphical model for Play component tn = time in secs of nth note sn = rate (secs/meas) at nth note ( tn+1 s n+1 ) = ( 1 lengthn 0 1 ) ( ) tn s n + ( τn σ n ) Listen and Accomp modeled as noisy observations of true note time Listen Update Composite Accomp 17

Inference and generation in Play component Inference: Model trained using EM, first on accompaniment data then solo data. Listen Update Composite Accomp Real time accompaniment: Each time new info observed recompute marginal for next accomp. note and schedule. Listen Update Composite Accomp 18

KCCA (Dorard, Hardoon & Shawe-Taylor) Yesterday s talk so I ll keep it short... Want to fit specific performer style (unlike, e.g., Widmer et.al.) Correlate musical score to performance Score features: melody and chords projected into vector using Paiement et.al. Figure 3: First two bars of Etude 3 Opus 10 by Chopin Beat Melody Chord 1 B3 B3 2 E3 [E2 B2 G#3 B3 E4] 3 D#3 [E2 B2 G#3 D#3]......... Figure 4: Feature representation of the score in Figure 3 19

KCCA (Dorard, Hardoon & Shawe-Taylor) Yesterday s talk so I ll keep it short... Audio performance features: instantaneous tempo and loudness of onsets ( worm of Dixon et al) Use KCCA (a kernel version of Canonical Correlation Analysis) to correlate these two views. Required kernel for score features and kernel for audio (worm) features Currently only preliminary results. Figure 1: Smoothed graphical view of a worm Beat Tempo (bpm) Loudness (sone) 1 22.3881 3.2264 2 22.3881 2.3668 3 21.4286 6.7167 4 19.0597 4.2105 5 28.1426 8.3444 6 30.0000 10.2206 7 26.7857 14.1084 8 25.8621 14.0037 9 35.7143 7.8521......... Figure 2: Machine representation of a worm 20

Summary Important information in timing and dynamics. Artificial expressive performances can be pleasing We saw four approaches to automatic performance: classic AI rules-based system (KTH) rules induction (Widmer) generative model (Raphael) kernel approach (Dorard et. al.) But: these all make use of a musical score. (Some less than others...) Can we get away from that? 21

Challenges in score-free expressive performance Local information is not sufficient for modeling music expression Score contains long-timescale information about phrasing and metrical organization Automatic methods exist for estimating deep hierarchical structure in music from a score Without score, this task is more difficult Graphic from AITEC Department of Future Technologies (ftp.icot.or.jp) 22

Focus: musical meter Meter provides long-timescale framework for music Meter and performance are closely related Mean Error Proportions 0.6 0.5 0.4 0.3 0.2 0.1 0 1 2 3 4 5 6 7 8 4/4 performances Model Example: performance errors correlate with meter. Palmer & Pfordresher (2003) Rest of the talk: use meter as proxy for musical score to gain access to nonlocal information Event Strength S x * M x (i) Absolute Distance * Slow 23

Audio pre-processing (not necessary for MIDI) Waveform at original sampling rate Log spectrogram with ~10ms frames Sum of gradient yields ~100hz signal

Computing Autocorrelation 100Hz signal Autocorrelation Signal Signal shifted by k Autocorrelation value a(k) for a single lag k is the sum of dot-product between signal and signal shifted k points.

Preserving phase (example: lag 380) Signal Signal shifted by k Dot-product points from 0 to 379 points from 380 to 759... points from k * lag to (k+1)* lag -1 Σ lag-380 autocorrelation energy stored mod-380 Store autocorrelation information for a single lag K in a vector of length K. Phase of autocorrelation energy is preserved spatially in the vector.

The Autocorrelation Phase Matrix (APM) The autocorrelation phase matrix (APM) has a row for each lag. Rows are ordered by lag. Phase is stored in milliseconds. Thus the matrix is triangular (long lags take more time before they cycle around). lag (ms)

The Autocorrelation Phase Matrix (APM) The APM provides a local representation for tempo variations and rhythmical variations Small horizontal changes on APM reach near-neighbors in frequency Small vertical changes on APM reach near-neighbors in phase

Metrical Interpretation A metrical tree can be specified as a set of metrically related points on the APM Search is thus done in space of meter and tempo lag (ms)

Finding beat and meter Time Search is done through the space of metrical trees using Viterbi alignment. Resulting metrical tree contracts and expands with changing tempo. Details in Eck (2007). 30

Expressive performance dynamics Use the APM to identify meter as it changes in time. Measure expressive dynamics and timing with respect to the APM Measurements made in milliseconds (time) but stored in radians (phase) Allows us to generalize to new pieces of music with different tempi and meter Integrate over time the winning metrical tree. 31

Modest example Morph the Chopin etude to sound a bit like me playing Heart and Soul after a couple of beers. Use hill climbing to find nearest maxima in target vector. Provides rudimentary measurelevel perturbation only (preliminary and unrealistic). Timing, velocity, chord spread. 32

Collecting performance stats for the piano For piano, identify hands using clustering Easier than finding leading melodic voice. No melodic analysis required Once hands are identified, identify chords Measure duration, velocity, legato,chord spread 33

Hand-specific statistics for piano Hands are somewhat rhythmically independent Measurements with respect to single hand are different than those for both hands (here: duration) 34

Conclusions Expressive timing and dynamics are important part of music Short overview of approaches Discussed task of score-free expressive performance Suggest using metrical structure as proxy for musical score Related this to APM model Future work: There remains more future work than completed work! So this list would be too long... Thank you for your patience. 35

Bibliography B. Repp. (1990). Patterns of expressive timing in performances of a Beethoven minuet by nineteen famous pianists. Journal of the Acoustical Society of America, 88: 622-641. B. Repp. (1997). The aesthetic quality of a quantitatively average music performance: Two preliminary experiments. Music Perception, 14: 419-444. A. Friberg, R. Bresin & J. Sundberg (2006). Overview of the KTH rule system for musical performance. Advances in Cognitive Psychology, 2(2-3):145-161. G. Widmer (2003). Discovering simple rules in complex data: A meta-learning algorithm and some surprising musical discoveries. Artificial Intelligence 146:129-148. C. Raphael (2004). A Bayesian Network for Real-Time Musical Accompaniment, Neural Information Processing Systems (NIPS) 14. L. Dorard, D. Hardoon & J. Shawe-Taylor (2007). Can Style be Learned? A Machine Learning Approach Towards Performing as Famous Pianists. NIPS Music Brain Cognition Workshop, Whistler. J.F. Paiement, D. Eck, S. Bengio & D. Barber (2005). A graphical model for chord progressions embedded in a psychoacoustic space. In Proceedings of the 22nd International Conference on Machine Learning (ICML), Bonn, Germany. S. Dixon, W. Goebl & G. Widmer (2002). The performance worm: Real time visualisation of expression based on Langner s tempo-loudness animation. In Proceedings of the International Computer Music Conference (ICMC). C. Palmer & P. Pfordresher (2003). Incremental planning in sequence production. Psychological Review 110:683-712. D. Eck. (2007). Beat tracking using an autocorrelation phase matrix. In Proceedings of the 2007 International Conference on Acoustics, Speech and Signal Processing (ICASSP), 1313-1316. 36

Following are deleted slides 37

Example: Chopin Etude Opus 10 No 3 Deadpan (no expressive timing or dynamics) Human performance (Recorded on Boesendorfer ZEUS) Differences limited to: timing (onset, length) velocity (seen as red) pedaling (blue shading) Flat timing Flat velocity Expressive timing Flat velocity Expressive timing Expressive velocity

Example: Chopin Etude Opus 10 No 3 Deadpan (no expressive timing or dynamics) Human performance (Recorded on Boesendorfer ZEUS) Differences limited to: timing (onset, length) velocity (seen as red) pedaling (blue shading) Flat timing Flat velocity Expressive timing Flat velocity Expressive timing Expressive velocity

Example: Chopin Etude Opus 10 No 3 Deadpan (no expressive timing or dynamics) Human performance (Recorded on Boesendorfer ZEUS) Differences limited to: timing (onset, length) velocity (seen as red) pedaling (blue shading) Flat timing Flat velocity Expressive timing Flat velocity Expressive timing Expressive velocity

Example: Chopin Etude Opus 10 No 3 Deadpan (no expressive timing or dynamics) Human performance (Recorded on Boesendorfer ZEUS) Differences limited to: timing (onset, length) velocity (seen as red) pedaling (blue shading) Flat timing Flat velocity Expressive timing Flat velocity Expressive timing Expressive velocity

Focus: musical meter Meter is the measurement of a musical line into measures of stressed and unstressed "beats", indicated in Western music notation by the time signature. Many methods for (imperfectly) estimating metrical structure in audio and MIDI 39