COMPUTATIONAL INVESTIGATIONS INTO BETWEEN-HAND SYNCHRONIZATION IN PIANO PLAYING: MAGALOFF S COMPLETE CHOPIN

Similar documents
Investigations of Between-Hand Synchronization in Magaloff s Chopin

THE MAGALOFF CORPUS: AN EMPIRICAL ERROR STUDY

Maintaining skill across the life span: Magaloff s entire Chopin at age 77

Maintaining skill across the life span: Magaloff s entire Chopin at age 77

The Magaloff Project: An Interim Report

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

Finger motion in piano performance: Touch and tempo

The Magaloff Project: An Interim Report

Analysis of local and global timing and pitch change in ordinary

Human Preferences for Tempo Smoothness

On time: the influence of tempo, structure and style on the timing of grace notes in skilled musical performance

Semi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis

Unobtrusive practice tools for pianists

Goebl, Pampalk, Widmer: Exploring Expressive Performance Trajectories. Werner Goebl, Elias Pampalk and Gerhard Widmer (2004) Introduction

Computer Coordination With Popular Music: A New Research Agenda 1

Unit Outcome Assessment Standards 1.1 & 1.3

Table of Contents. Contents of this CD-ROM. (click on a category to go to that section of the Table of Contents) MISCELLANEOUS SOLO WORKS

Widmer et al.: YQX Plays Chopin 12/03/2012. Contents. IntroducAon Expressive Music Performance How YQX Works Results

From quantitative empirï to musical performology: Experience in performance measurements and analyses

Machine Learning Term Project Write-up Creating Models of Performers of Chopin Mazurkas

An Integrated Music Chromaticism Model

A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS

Pre-College Levels. PC 1 PC 2 Diploma. Pre-College 1 TECHNIQUE:

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis

Towards a Complete Classical Music Companion

Tempo and Beat Analysis

Improving Piano Sight-Reading Skills of College Student. Chian yi Ang. Penn State University

2 The Tonal Properties of Pitch-Class Sets: Tonal Implication, Tonal Ambiguity, and Tonalness

Romantic is a term used to describe the music and art that was created from about 1810 to 1900.

Experiment on adjustment of piano performance to room acoustics: Analysis of performance coded into MIDI data.

Temporal coordination in string quartet performance

WHO IS WHO IN THE END? RECOGNIZING PIANISTS BY THEIR FINAL RITARDANDI

Zooming into saxophone performance: Tongue and finger coordination

CS229 Project Report Polyphonic Piano Transcription

Nocturnes And Polonaises By Ignacy Jan Paderewski, Frederic Chopin

Extracting Significant Patterns from Musical Strings: Some Interesting Problems.

TEXAS MUSIC TEACHERS ASSOCIATION Student Affiliate World of Music

Trevor de Clercq. Music Informatics Interest Group Meeting Society for Music Theory November 3, 2018 San Antonio, TX

2016 HSC Music 1 Aural Skills Marking Guidelines Written Examination

Harding University Department of Music. PIANO PRINCIPAL HANDBOOK (rev )

L van Beethoven: 1st Movement from Piano Sonata no. 8 in C minor Pathétique (for component 3: Appraising)

SAMPLE ASSESSMENT TASKS MUSIC CONTEMPORARY ATAR YEAR 11

EXPLORING EXPRESSIVE PERFORMANCE TRAJECTORIES: SIX FAMOUS PIANISTS PLAY SIX CHOPIN PIECES

TEMPO AND BEAT are well-defined concepts in the PERCEPTUAL SMOOTHNESS OF TEMPO IN EXPRESSIVELY PERFORMED MUSIC

How to Obtain a Good Stereo Sound Stage in Cars

chopin preludes 0D9D8BBF0D49128FF1FB738265B82467 Chopin Preludes 1 / 6

rhinegold education: subject to endorsement by ocr Mozart: Clarinet Concerto in A, K. 622, first movement Context Scores AS PRESCRIBED WORK 2017

Introduction. Figure 1: A training example and a new problem.

Music Representations

MARK SCHEME for the June 2005 question paper 0410 MUSIC

Calculating Dissonance in Chopin s Étude Op. 10 No. 1

MTO 18.1 Examples: Ohriner, Grouping Hierarchy and Trajectories of Pacing

17. Beethoven. Septet in E flat, Op. 20: movement I

Marc I. Johnson, Texture Technologies Corp. 6 Patton Drive, Hamilton, MA Tel

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University

A PRELIMINARY COMPUTATIONAL MODEL OF IMMANENT ACCENT SALIENCE IN TONAL MUSIC

Student Performance Q&A:

ASD JHS CHOIR ADVANCED TERMS & SYMBOLS ADVANCED STUDY GUIDE Level 1 Be Able To Hear And Sing:

An Interpretive Analysis Of Mozart's Sonata #6

Introduction to Music

Tempo Fluctuation in Two Versions of Body and Soul by John Coltrane and Dexter Gordon

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

\it J 11 ;JJ' JJJ J J J J J i JJJ JIH JITTTB^

Eligibility / Application Requirements / Repertoire

ESP: Expression Synthesis Project

jsymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada

Measuring & Modeling Musical Expression

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

Lesson One. New Terms. Cambiata: a non-harmonic note reached by skip of (usually a third) and resolved by a step.

jsymbolic 2: New Developments and Research Opportunities

MATCH: A MUSIC ALIGNMENT TOOL CHEST

University of West Florida Department of Music Levels of Attainment piano

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm

Example 1 (W.A. Mozart, Piano Trio, K. 542/iii, mm ):

The Relationship Between Auditory Imagery and Musical Synchronization Abilities in Musicians

MUSIC FOR THE PIANO SESSION TWO: FROM FORTEPIANO TO PIANOFORTE,

> f. > œœœœ >œ œ œ œ œ œ œ

Handbook for Applied Piano Students

Laboratory Assignment 3. Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB

ST. JOHN S EVANGELICAL LUTHERAN SCHOOL Curriculum in Music. Ephesians 5:19-20

Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment

A Recipe for Emotion in Music (Music & Meaning Part II)

Monday 23 May 2016 Morning

Music Complexity Descriptors. Matt Stabile June 6 th, 2008

Lesson One. New Terms. a note between two chords, dissonant to the first and consonant to the second. example

The Pines of the Appian Way from Respighi s Pines of Rome. Ottorino Respighi was an Italian composer from the early 20 th century who wrote

GRAAD 12 NATIONAL SENIOR CERTIFICATE GRADE 12

Choir Scope and Sequence Grade 6-12

Music Performance Solo

MPATC-GE 2042: Psychology of Music. Citation and Reference Style Rhythm and Meter

Haydn: Symphony No. 97 in C major, Hob. I:97. the Esterhazy court. This meant that the wonderful composer was stuck in one area for a large

Instrumental Performance Band 7. Fine Arts Curriculum Framework

Flute Studio University of Idaho GUIDELINES for ADMISSION TO MusA 115

Study Guide. Solutions to Selected Exercises. Foundations of Music and Musicianship with CD-ROM. 2nd Edition. David Damschroder

Quarterly Progress and Status Report. Perception of just noticeable time displacement of a tone presented in a metrical sequence at different tempos

Chapter 40: MIDI Tool

LEVEL 11. Complete Requirements

An Empirical Comparison of Tempo Trackers

Music Representations

2014 Music Performance GA 3: Aural and written examination

Transcription:

COMPUTATIONAL INVESTIGATIONS INTO BETWEEN-HAND SYNCHRONIZATION IN PIANO PLAYING: MAGALOFF S COMPLETE CHOPIN Werner Goebl, Sebastian Flossmann, and Gerhard Widmer Department of Computational Perception Johannes Kepler University Linz, Austria ABSTRACT The paper reports on first steps towards automated computational analysis of a unique and unprecedented corpus of symbolic performance data. In particular, we focus on between-hand asynchronies an expressive device that plays an important role particularly in Romantic music, but has not been analyzed quantitatively in any substantial way. The historic data were derived from performances by the renowned pianist Nikita Magaloff, who played the complete work of Chopin live on stage, on a computer-controlled grand piano. The mere size of this corpus (over 32, performed notes or almost 1 hours of continuous performance) challenges existing analysis approaches. The computational steps include score extraction, score-performance matching, definition and measurement of the analyzed features, and a computational visualization tool. We then present preliminary data to demonstrate the potential of our approach for future computational modeling and its application in computational musicology. 1 INTRODUCTION This paper presents research towards automated computational analysis of large corpora of music performance data. In particular, we give a first report on a computational approach to analyzing a unique corpus of historic performance data: basically the complete work of Chopin, performed by the renowned pianist Nikita Magaloff. Corpora of that size hundreds of thousands of played notes require a level of automation of analysis that has not been accomplished so far. We describe the required processing steps, from converting scanned scores into symbolic notation, to score-performance matching, definition and automatic measurement of between-hand asynchronies, and a computational visualization tool for exploring and understanding the extracted information. SMC 29, July 23 25, Porto, Portugal Copyrights remain with the authors As the two hands of a pianist have the possibility to produce different musical parts independently, the between-hand asynchronies yield a spectrum of artistic expression ranging from the fairly constrained melody lead effect [1] to bass anticipations [5], or the earlier type of tempo rubato [3]. Towards the end of the paper, we present preliminary results on the between-hand asynchronies in Magaloff s complete Chopin to demonstrate the scope of insights that such large corpora can offer. Finally, we discuss the future pathways of this research endeavor and its potential for computational modeling and musicological investigation. 2 THE CHOPIN CORPUS The analyzed Chopin corpus comprises live concert performances by the Georgian-Russian pianist Nikita Magaloff (1912 1992), who played almost the entire solo repertoire of Chopin in a series of 6 recitals between January and May 1989 at the Mozart-Saal of the Wiener Konzerthaus 1 in Vienna on a Bösendorfer computer-controlled grand piano. In this unprecedented project, Magaloff, by that time already 77 years old, performed all works of Chopin for solo piano that appeared in print during Chopin s life time, keeping a strict ascending order by opus number, starting with the Rondo, Op. 1 up to the three Waltzes Op. 64, including the three Sonatas, 41 Mazurkas, 25 Préludes, 24 Études, 18 Nocturnes, 8 Waltzes, 6 Polonaises, 4 Scherzos, 4 Ballades, 3 Impromptus, 3 Rondos, and other works (Variations brillantes, Bolero, Tarantelle, Allegro de Concert, Fantaisie, Berceuse, Barcarole, and Polonaise-Fantaisie). 2 Magaloff played these recitals on a Bösendorfer SE computer-controlled grand piano that recorded his performances onto computer hard disk. The SE format stores the performance information in a symbolic format with high precision, providing detailed information on the onset and off- 1 This concert hall provides about 7 seats, http://www. konzerthaus.at. 2 The works not played were either piano works with orchestra accompaniment (Op. 2, 11, 13, 14, 21, and 22), works with other instruments (Op. 3, 8 and 65), or works with higher (op. posth., starting from Op. 66, the Fantaisie-Impromptu) or no opus numbers. Page 291

set timing of each played tone, the dynamics, and the pedalling. The entire corpus comprises more than 15 individual pieces, over 323, performed notes or almost 1 hours of continuous performance. 3 COMPUTATIONAL ANALYSIS OF PERFORMANCE DATA 3.1 Score Extraction In order to analyze symbolic performance data automatically, the performances have to be connected to the corresponding musical scores (score-performance matching). As symbolic scores were not available for the complete work of Chopin, the first step was to extract this information from the printed music scores. We used a music recognition software ( SharpEye ) to convert the 946 pages of scanned music into a musicxml representation. Extensive manual verification of the conversion process was necessary to eliminate a large number of conversion errors, as well as postcorrection of conversion incapabilities of the used software (e.g., ottava lines, parts crossing staffs, etc.). 3.2 Score Performance Matching The symbolic scores were then matched to Magaloff s performances employing a matching algorithm based on an edit distance metric [4]. Also the matching results had to be inspected and corrected manually with an interactive graphical user interface that displays the note-by-note match between the score information and the performance. All uncorrectly played notes or performed variants were identified and labeled. (This, by the way, will also make it possible to perform large-scale, in-depth analyses of the kinds of errors accomplished pianists make.) 3.3 Defining and Measuring Asynchronies Our aim was to analyze the between-hand asynchronies of notes that are notated as nominally simultaneous in the score (that is, all tones belonging to the same score event ). To that end, we first needed to compute these asynchronies automatically from the corpus. The staff information of the musical notation (upper versus lower staff) was used to calculate the between-hand asynchronies. As the performance data does not contain information about what hand played what parts of the music, we assumed that overall the right hand played the upper staff tones and the left hand the lower. 3 We define a between-hand asynchrony as follows: For all notes belonging to the same score event, subtract the on- 3 Certainly, there are numerous passages where this simple assumption is wrong or not likely to be true, but given the sheer size of the data set, the potential bias may be tolerable. set timing of each upper-staff note from the onset timing of each lower-staff note ( lower minus upper ). Thus, positive asynchrony values indicate that the upper staff (right hand) is early, while negative numbers denote bass anticipations (left hand early). Multiple asynchronies within one score event were averaged for subsequent data analyses. All notated arpeggios, ornaments, trills, or grace notes were excluded from our preliminary data analysis (about 1% of the entire data), as these cases feature special and usually larger asynchronies than regular score events. These special cases deserve a separate detailed analysis that would exceed the scope of the present paper. 3.4 Computational Visualization For a first intuitive analysis and understanding of this huge amount of measurement data, adequate visualization methods are needed. Thus, we developed a dedicated computational visualization tool. A screenshot is presented in Figure 1. It comprises three panels on top of each other, sharing the same time axis. The upper panel shows the individual tempo curves of the two hands; the middle panel shows the average asynchronies for each score event that contained simultaneous notes in each staff; and the lower panel features a piano-roll representation of the performances with the relevant notes connected by vertical lines. The color of these lines is either red (indicating a right-hand lead) or green (indicating a left-hand lead). The grey area in the middle panel marks a range of +/ 3 ms within which asynchronies are not likely to be perceived as such [2]. Furthermore, the tool indicates occurrences of bass anticipations ( B.A., lower panel) and out-of-sync regions (horizontal bars, middle panel; for description see below). 4 PRELIMINARY RESULTS In the following, we give a first glimpse of the scope of results that such large-scale analyses may yield. 4.1 Overall Asynchronies The distribution of all asynchronies between the two hands is shown in Figure 2, including the mean and the mode value. The positive mode value reflects an overall tendency for the right hand to be early, which is most likely attributable to the well-known melody lead effect [1]. Moreover, the mean value is slightly below the mode value reflecting a slightly skewed histogram towards the left side. Particularly in the region of 1 to 3 ms there is a slight increase of values, most likely due to frequent bass anticipations (red line, see below). The asynchrony distributions of the individual pieces vary considerably and may depend on the specifics of the pieces. The pieces played most synchronously by Magaloff are those Page 292

TEMPO (bpm) 24 16 12 8 6 4 3 Op. 27 No. 2 played by Nikita Magaloff 522 524 526 528 53 532 534 536 538 Right hand Left hand ASYNCH (ms) 2 1 1 2 11 522 524 526 528 53 532 534 536 538 ) 1 9 8 PITCH 7 6 265. 226. 224. 255. 264. 157. 228. B.A. 5 4 B.A. B.A. 3 2 522 524 526 528 53 532 534 536 538 TIME (s) Figure 1. Screenshot of a computational visualization tool showing bars 5 54 of Chopin s Nocturne Op. 27 No. 2 performed by Magaloff. The tempo curves of the two hands are plotted in the upper panel, the asynchronies in middle (positive values indicate an early right hand; negative an early left hand; the grey area sketches the +/ 3-ms region around zero), and a piano roll representation is shown in the lower panel. All tones of computed asynchronies are connected by (almost) vertical lines that are plotted in red when the melody (right hand) was ahead, green when it lagged. Event wise Count 12 1 8 6 4 2 Mean=4.388 Mode=13. 5 25 25 5 Between hand Asynchronies (ms) Figure 2. Histogram of the signed between-hand asynchronies per event over the entire Chopin corpus (displaying a total of 163,28 asynchronies). The distribution of bass anticipations (see Section 4.2) is drawn by a red line on the left part of the histogram. that feature predominantly chordal textures (Op. 4-1, 28-9, 28-2, 1-2, see Figure 3); the least synchronous pieces are those that have strong melodic textures that leave room for artistic interpretation. There is a significant tempo effect within the investigated pieces. Figure 3 shows the mean absolute (unsigned) asynchronies per piece (a) and the standard error of the asynchronies (b) against the average event rate (in events per second). An event rate value was computed for each score event by counting the performed events (chords) within a time window of 3 seconds around it. The average event rate is the piece-wise mean of those values. We found that the faster the piece, the lower is the absolute asynchrony and also the lower is the variability of the asynchronies, which suggests that Magaloff takes more room to employ expressive asynchronies in slower pieces than faster pieces. 4.2 Bass Anticipations A bass anticipation is labeled as such when the lowest tone of a lower-staff score event is more than 5 ms ahead of the mean onsets of the upper-staff tones of that event. The Page 293

a) b) mean Unsigned Asynchrony (ms) SE of Asnchronies (ms) 8 7 6 5 4 3 2 1 28 2 1 6 45 62 1 48 2 58 3 28 13 27 2 55 2 15 2 9 1 27 1 9 3 37 1 17 4 63 2 r=.28*** n=15 p<.1 28 4 9 232 124 1 1 12 25 2 55 1 4 3 5 2 57 1 9 28 24 32 2 62 2 28 7 15 1 25 1 7 2 3 3 34 2 51 35 1 54 17 2 25 7 59 2 6 59 1 56 3 64 2 58 1 1 1 58 2 63 3 33 3 3 148 1 6 2 17 3 24 4 24 3 7 5 1 3 6 4 2 2319 29 28 19 28 6 47 3652 31 25 11 3 4 26 15 3 59 3 38 28 21 5 28 23 34 1 25 3 35 4 35 3 41 1 28 15 28 17 33 4 4 261 5 1 37 2 6 1 1 7 341 3 56 2 64 3 7 1 4 1 12 25 8 43 58 4 28 3 33 1 15 3 17 1 28 1 24 235 25 1 49 28 5 16 4 2 28 11 39 42 64 1 56 1 46 28 16 63 1 41 4 44 53 34 3 1 7 25 6 1 4 1 5 7 4 28 14 4 4 1 8 3 2 41 2 18 28 22 33 2 6 3 26 2 25 4 28 2 25 9 28 12 28 18 1 2 28 9 4 1 2 4 6 8 1 12 2 15 1 5 28 2 r=.417*** n=15 p<.1 28 7 1 6 45 63 2 28 1 33 3 37 1 28 13 28 4 15 2 28 6 58 3 7 2 6 4 62 17 4 28 23 9 1 9 217 2 48 2 3 1 28 24 24 1 32 1 4 3 27 1 3 3 55 2 5 2 62 2 27 2 24 3 28 21 7 5 55 1 56 2 25 7 9 3 28 11 1 12 33 1 32 2 17 1 15 1 9 28 3 17 3 6 1 1 3 4 2 6 2 7 3 24 4 7 4 15 3 28 9 28 15 41 1 56 3 63 3 59 2 58 2 3 4 34 2 59 1 35 3 5 1 41 3 28 2 26 1 37 2 33 4 48 1 35 2 3635 1 28 22 24 2 25 419 1 7 6 3 7 1 25 1 4 112 223 25 8 28 12 28 5 1525 9 28 14 25 3 1 1 25 1 1 2 25 2 1 525 11 28 173 2 5 3 41 4 4 2 51 57 63 1 64 2 59 3 64 1 6 56 1 47 38 34 1 34 3 1 4 1 8 28 16 28 18 41 2 58 1 61 64 3 54 29 1833 2 31 16 42 25 6 28 19 35 4 4 1 26 2 44 53494652 39 4 4 43 58 4 2 4 6 8 1 12 Figure 3. Absolute (unsigned) asynchronies (a) and standard error (SE) of the mean asynchronies (b) against the mean event rate per piece. The numbers refer to the opus numbering of the pieces. overall distribution of the bass leads is shown in Figure 2 (red line) on the left side of the histogram, and the invididual pieces are shown in Figure 4. The proportion of bass anticipations is lowest on average for the Etudes, the Preludes, and the Rondos (well below an average of 1% of the events), and highest in the Mazurkas and the Nocturnes (almost 2%). No bass anticipations were found particularly in the Preludes (16 out of 25 did not contain bass anticipations) and the Etudes (7 of 24). An exception is the Prelude Op. 28 No. 2, which exhibits both the highest mean asynchronies and the largest proportion of bass anticipations among all pieces (clearly visible in Figures 3 and 4). This very slow and short piece features a constant 1/8-note accompaniment with a single-note melody above it. The sad character and the slow tempo may be the reason for the high temporal independence of the melody in Proportion of Bass Anticipations.2.15.1.5 28 2 63 2 1 6 17 4 r=.349*** n=115 p<.1 9 2 4 2 58 2 37 1 28 13 41 124 1 62 1 15 2 58 3 59 2 27 2 17 3 33 4 6 28 23 7 2 26 1 17 2 15 1 17 124 4 28 21 3 3 5 248 1 57 56 3 25 1 4 1 38 15 3 4 3 34 2 36 4 2 6 1 7 5 5 1 2 9 1 9 3 2 23 25 8 48 2 35 1 25 7 24 2 28 11 25 4 18 28 22 31 3 1 1 3 6 26 3 12 19 28 129 1 16 4 4 1 12 1 4 1 5 25 11 32 1 28 17 32 2 41 4 4447 46 5 3 51 61 55 1 59 1 64 2 63 3 34 3 35 3 37 2 59 3 56 2 56 1 3 2 35 2 49 27 1 41 3 534 1 52 54 42 41 2 58 1 64 3 64 1 1 7 25 6 6 4 39 35 4 28 2 28 7 28 4 28 9 28 6 1 11 28 15 33 1 3 4 5 1 33 3 55 2 45 28 1 7 3 62 2 7 1 4 1 63 1 26 2 24 3 7 4 28 18 33 2 1 925 9 28 5 28 24 28 14 25 3 1 1 43 25 1 58 4 28 19 28 3 25 2 1 1 1 8 28 16 2 4 6 8 1 12 Figure 4. Proportion of bass anticipations against mean event rate per piece. Zero proportions (34 pieces) were excluded from the calculation of the regression line. Magaloff s performance. There is also an effect of event rate, suggesting that bass leads become less frequent as the tempo of the pieces increases (Figure 4). Again, slower pieces leave more room for expressive freedom than do faster pieces. To further analyze the occurrences of bass anticipations, we categorized all score events bar-wise into first beats, on-beats (all beat events except the first beat), and off-beats. It turns out that metrical position has a significant effect: the highest number of bass anticipations fall on the first beat (1.8%); other on-beat events receive 1.48% bass anticipations, and.66% are found on off-beat events. This suggests that bass anticipations are used by Magaloff to emphasize predominantly strong beats. 4.3 The Earlier Type of Tempo Rubato An expressive means that has a long performance tradition is the tempo rubato in the earlier meaning [3]. It refers to expressive temporal deviations of the melody line, while the accompaniment, offering the temporal reference frame, remains strictly in time. Chopin in particular often recommended his pupils to keep the accompaniment undisturbed like a conductor, and give the right hand the freedom of expression with fluctuations of speed [3, p. 193]. In contrast, the later meaning of tempo rubato was used more and more to refer to the parallel slowing down and speeding up of the entire music (today more generally referred to as expressive timing). In expressive performance, both forms of rubato are present simultaneously. We aim at identifying sequences of earlier tempo rubato automatically from the entire corpus. To extract overall information about sequences where Magaloff apparently employed an earlier tempo rubato, we count the out-of-sync re- Page 294

Number of O o S Regions 1 8 6 4 2 48 2 32 2 r=.349*** n=89 p<. 58 317 4 9 2 62 1 55 2 6 27 2 48 1 28 17 56 3 26 1 5 2 9 1 35 2 519 3 15 1 28 432 1 55 1 1 3 3 3 63 2 34 2 57 3835 1 2 54 1 37 1 1 6 4 3 24 1 37 2 6 27 1 59 1 49 3652 23 28 2 28 15 25 7 63 3 3 1 33 4 62 2 15 2 24 4 26 2 41 4 47 6146 58 1 19 29 25 1 25 2 35 3 15 3 28 13 33 1 17 1 45 7 3 3 2 24 3 5 3 4 2 25 1 41 2 64 3 1 9 31 539 28 5 28 1 1 12 58 4 25 11 35 4 28 2 28 7 28 9 28 6 1 11 41 1 17 217 3 3 4 5 1 33 3 28 1 59 259 3 24 2 6 3 6 1 4 1 7 1 63 1 6 4 41 3 28 21 44 56 2 56 1 28 11 53 7 4 28 18 34 1 25 4 7 5 64 2 12 28 228 23 33 2 25 9 34 3 25 8 28 12 28 24 28 14 16 4 4 25 3 1 7 42 1 1 25 6 58 2 64 1 43 1 21 1 28 19 28 3 1 4 1 5 1 8 28 16 2 4 6 8 1 12 Figure 5. The number of out-of-sync regions (earlier tempo rubato) per piece is plotted against the event rate. gions of each piece. An out-of-sync region is defined as a sequence of consecutive asynchronies that are larger than the typical perceptual threshold (3 ms) and that contain more elements (events) than the average event rate of that piece (2 13 tones). Usually faster pieces contain many shorter runs that are out-of-sync, but due to the fast tempo, these regions extend only to some fraction of a second. To take this into account, the search for out-of-sync regions is coupled to the piece-wise average event rate. On average, a piece contains 1.8 such regions. The pieces with the lowest numbers are generally the Mazurkas, Preludes, and Etudes (below 1), the pieces with the highest counts are by far the Nocturnes (on average well over 5), suggesting that particularly this genre within Chopin s music leaves most room for letting the melody move freely above the accompaniment. Figure 5 shows the number of out-of-sync regions per piece against the average event rate of the piece. It demonstrates that faster pieces contain fewer such regions, suggesting that this form of tempo rubato is bound to slower and medium tempi (such as the Nocturnes, the slowest piece category in the Chopin corpus). This overall finding is not surprising; the earlier tempo rubato is expected to be more often found in melodic contexts rather than in virtuoso pieces, as the historic origins of the earlier tempo rubato go back to vocal music. To illustrate, the example of the visualization tool presented in Figure 1 is briefly discussed. It shows an excerpt (bars 5 54) of the Nocturne Op. 27 No. 2. This example contains two runs of tempo rubato as determined by the algorithm (indicated by horizontal bars in the middle panel). The first starts on the downbeat of bar 5, where Magaloff delayed the melody note by 265 ms, only to be early over the next few notes of the descending triplet passage. The beginning of the 48-tuplet figure (which is interpreted as 1/16-note triplets as well) also leads the accompaniment. Towards its end, the second run of tempo rubato as determined by our algorithm begins, just when Magaloff starts to lag behind the accompaniment. This lag coincides with a downward motion and a notated decrescendo. The following embellishment of the b-flat (notated as 1/32 notes and 1/32 triplets) is again clearly ahead of the accompaniment. The first note of the next phrase is also ahead, potentially to underline the notated anticipation of the upcoming harmony change towards e-flat minor. 5 SUMMARY AND FUTURE WORK This paper presented a computational approach to making large performance corpora accessible for detailed analysis. It defined and automatically measured between-hand synchronization in over 15 pieces by Frédéric Chopin. Working with data sets of that size, i.e., complete sets of performances of a single composer or several hundred thousand played notes requires, among other things, effective scoreperformance matching algorithms and interactive graphical user interfaces for post-hoc data inspection and correction. Preliminary data analysis of the between-hand synchronization attempted to demonstrate the rich use of asynchronies in Magaloff s complete Chopin corpus, a historic document of a unique performance project that offers unequaled opportunities for performance analysis. We sketched overall trends of asynchronicity with respect to pieces, tempo, and metrical constraints, as well as specific cases of bass anticipations and occurrences of tempo rubato in its earlier meaning. This research endeavor is preliminary as it stands. It is meant to lead to further modeling efforts to be able to predict asynchronies from Romantic scores following Nikita Magaloff s instrinsic style. Piece category, overall musical texture, as well as local and global aspects from the scores may be promising features for training machine learning algorithms in order to derive predictive computational models of between-hand asynchronization. Such models may enhance existing performance rendering systems by adding an important expressive feature. To be able to automatically assess performance corpora of this scale offers completely new pathways for computational musicology. Historic documents such as the present corpus are in managable reach for detailed analysis. Other large corpora, such as piano rolls of historic reproducing pianos or the performance database of the Yamaha ecompetition 4, may be other sources for future large-scale performance investigation. Finally, detailed knowledge derived from performances by established musicians may help us develop real-time visualization tools that give intelligent feedback to practicing piano students to enhance their awareness of what they are doing, and potentially help them to improve their play instantly. 4 http://www.piano-e-competition.com Page 295

6 ACKNOWLEDGEMENTS This research was supported by the Austrian Research Fund (FWF, grant P19349-N15). We are indebted to Mme. Irène Magaloff for her generous permission to use her late husband s performance data for our research. 7 REFERENCES [1] GOEBL, W. Melody lead in piano performance: Expressive device or artifact? Journal of the Acoustical Society of America 11, 1 (21), 563 572. [2] GOEBL, W., AND PARNCUTT, R. The influence of relative intensity on the perception of onset asynchronies. In Proceedings of the 7th International Conference on Music Perception and Cognition, Sydney (ICMPC7), Aug. 17 21, 22 (Adelaide, 22), C. Stevens, D. Burnham, G. McPherson, E. Schubert, and J. Renwick, Eds., Causal Productions, pp. 613 616. [3] HUDSON, R. Stolen Time: The History of Tempo Rubato. Clarendon Press, Oxford, 1994. [4] MONGEAU, M., AND SANKOFF, D. Comparison of musical sequences. Computers and the Humanities 24 (199), 161 175. [5] VERNON, L. N. Synchronization of chords in artistic piano music. In Objective Analysis of Musical Performance, C. E. Seashore, Ed., vol. IV of Studies in the Psychology of Music. University Press, Iowa City, 1936, pp. 36 345. Page 296