Unobtrusive practice tools for pianists

Similar documents
However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

EXPLORING EXPRESSIVE PERFORMANCE TRAJECTORIES: SIX FAMOUS PIANISTS PLAY SIX CHOPIN PIECES

COMPUTATIONAL INVESTIGATIONS INTO BETWEEN-HAND SYNCHRONIZATION IN PIANO PLAYING: MAGALOFF S COMPLETE CHOPIN

ESP: Expression Synthesis Project

Finger motion in piano performance: Touch and tempo

Zooming into saxophone performance: Tongue and finger coordination

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

Investigations of Between-Hand Synchronization in Magaloff s Chopin

Robert Alexandru Dobre, Cristian Negrescu

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis

Analysis of Musical Content in Digital Audio

Measuring & Modeling Musical Expression

Maintaining skill across the life span: Magaloff s entire Chopin at age 77

THE MAGALOFF CORPUS: AN EMPIRICAL ERROR STUDY

6.UAP Project. FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System. Daryl Neubieser. May 12, 2016

Analysis of local and global timing and pitch change in ordinary

Visual and Aural: Visualization of Harmony in Music with Colour. Bojan Klemenc, Peter Ciuha, Lovro Šubelj and Marko Bajec

Introductions to Music Information Retrieval


Goebl, Pampalk, Widmer: Exploring Expressive Performance Trajectories. Werner Goebl, Elias Pampalk and Gerhard Widmer (2004) Introduction

Toward a Computationally-Enhanced Acoustic Grand Piano

An Empirical Comparison of Tempo Trackers

Computer Coordination With Popular Music: A New Research Agenda 1

A prototype system for rule-based expressive modifications of audio recordings

Semi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis

A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS

TEMPO AND BEAT are well-defined concepts in the PERCEPTUAL SMOOTHNESS OF TEMPO IN EXPRESSIVELY PERFORMED MUSIC

Assigning and Visualizing Music Genres by Web-based Co-Occurrence Analysis

Maintaining skill across the life span: Magaloff s entire Chopin at age 77

A Beat Tracking System for Audio Signals

Evaluation of the Audio Beat Tracking System BeatRoot

ANNOTATING MUSICAL SCORES IN ENP

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

PLAYSOM AND POCKETSOMPLAYER, ALTERNATIVE INTERFACES TO LARGE MUSIC COLLECTIONS

Subjective evaluation of common singing skills using the rank ordering method

Tempo and Beat Analysis

Temporal coordination in string quartet performance

Extracting Significant Patterns from Musical Strings: Some Interesting Problems.

Human Preferences for Tempo Smoothness

Multidimensional analysis of interdependence in a string quartet

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm

Music Representations

Sentiment Extraction in Music

Artificial Social Composition: A Multi-Agent System for Composing Music Performances by Emotional Communication

THE SOUND OF SADNESS: THE EFFECT OF PERFORMERS EMOTIONS ON AUDIENCE RATINGS

Evaluation of the Audio Beat Tracking System BeatRoot

Enhancing Music Maps

Automatic meter extraction from MIDI files (Extraction automatique de mètres à partir de fichiers MIDI)

Subjective Similarity of Music: Data Collection for Individuality Analysis

Perceptual Smoothness of Tempo in Expressively Performed Music

ESTIMATING THE ERROR DISTRIBUTION OF A TAP SEQUENCE WITHOUT GROUND TRUTH 1

Automatic Rhythmic Notation from Single Voice Audio Sources

Music Radar: A Web-based Query by Humming System

Rechnergestützte Methoden für die Musikethnologie: Tool time!

WHO IS WHO IN THE END? RECOGNIZING PIANISTS BY THEIR FINAL RITARDANDI

A Computational Model for Discriminating Music Performers

Expressive information

A Case Based Approach to the Generation of Musical Expression

Director Musices: The KTH Performance Rules System

OBSERVED DIFFERENCES IN RHYTHM BETWEEN PERFORMANCES OF CLASSICAL AND JAZZ VIOLIN STUDENTS

Chapter 40: MIDI Tool

Classification of Dance Music by Periodicity Patterns

Towards a Complete Classical Music Companion

QUALITY OF COMPUTER MUSIC USING MIDI LANGUAGE FOR DIGITAL MUSIC ARRANGEMENT

Quantifying the Benefits of Using an Interactive Decision Support Tool for Creating Musical Accompaniment in a Particular Style

Scoregram: Displaying Gross Timbre Information from a Score

SMS Composer and SMS Conductor: Applications for Spectral Modeling Synthesis Composition and Performance

MATCH: A MUSIC ALIGNMENT TOOL CHEST

Automatic Reduction of MIDI Files Preserving Relevant Musical Content

Secrets To Better Composing & Improvising

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

Application Note AN-708 Vibration Measurements with the Vibration Synchronization Module

On time: the influence of tempo, structure and style on the timing of grace notes in skilled musical performance

MusCat: A Music Browser Featuring Abstract Pictures and Zooming User Interface

Measurement of overtone frequencies of a toy piano and perception of its pitch

Automatic Music Clustering using Audio Attributes

Good playing practice when drumming: Influence of tempo on timing and preparatory movements for healthy and dystonic players

Analysing Musical Pieces Using harmony-analyser.org Tools

Sound design strategy for enhancing subjective preference of EV interior sound

Guide to Computing for Expressive Music Performance

DEVELOPMENT OF MIDI ENCODER "Auto-F" FOR CREATING MIDI CONTROLLABLE GENERAL AUDIO CONTENTS

CS 591 S1 Computational Audio

IMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM

How to Obtain a Good Stereo Sound Stage in Cars

APP USE USER MANUAL 2017 VERSION BASED ON WAVE TRACKING TECHNIQUE

Judgments of distance between trichords

The ubiquity of digital music is a characteristic

Automatic music transcription

The Human Features of Music.

PULSE-DEPENDENT ANALYSES OF PERCUSSIVE MUSIC

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t

Outline. Why do we classify? Audio Classification

Igaluk To Scare the Moon with its own Shadow Technical requirements

METRICAL STRENGTH AND CONTRADICTION IN TURKISH MAKAM MUSIC

The effect of exposure and expertise on timing judgments in music: Preliminary results*

Analytic Comparison of Audio Feature Sets using Self-Organising Maps

Music Complexity Descriptors. Matt Stabile June 6 th, 2008

Acoustic and musical foundations of the speech/song illusion

Tempo and Beat Tracking

Adaptive Key Frame Selection for Efficient Video Coding

Transcription:

To appear in: Proceedings of the 9 th International Conference on Music Perception and Cognition (ICMPC9), Bologna, August 2006 Unobtrusive practice tools for pianists ABSTRACT Werner Goebl (1) (1) Austrian Research Institute for Artificial Intelligence Vienna, Austria http://www.ofai.at/music This paper proposes novel computer-based interfaces for piano practicing. They are designed to display in real time certain well-defined sub-aspects of piano playing. They are intelligent and unobtrusive in that they adjust automatically to the needs of the practitioner so that no other interaction is needed than moving the piano keys. They include 1) a pattern display, finding recurring pitch patterns and displaying expressive timing and dynamics thereof, 2) a chord display, showing timing asynchronies and tone intensity variations of chords tones, and 3) an acoustic piano roll display that visually models the acoustic piano tone from MIDI data. Keywords Piano Practice, Performance Visualization, Music Education Proceedings of the 9th International Conference on Music Perception & Cognition (ICMPC9). 2006 The Society for Music Perception & Cognition (SMPC) and European Society for the Cognitive Sciences of Music (ESCOM). Copyright of the content of an individual paper is held by the primary (first-named) author of that paper. All rights reserved. No paper from this book may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photocopying, recording, or by any information retrieval systems, without written permission from the paper s primary author. No other part of this book may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photocopying, recording, or by any information retrieval system, with written permission from SMPC and ESCOM. Gerhard Widmer (1&2) (2) Department of Computational Perception University of Linz Linz, Austria http://www.cp.jku.at INTRODUCTION There is an increasing interest in the possible roles and uses of computers in music education (Webster, 2002), and the technical facilities needed are easily available to virtually everyone. However, there are still only few approaches for developing computer-based tools that are successfully used in every-day music performance practice. As an example, we mention refined spectrogram representations with fundamental frequency trackers that are used in singing lessons (see e.g., the VOXed system Welch, Himonides, Howard, & Brereton, 2004). In piano performance, visual feedback from conventional piano roll displays has proved to be a helpful means in piano instruction (Riley-Butler, 2001, 2002). A computational implementation of a MIDI-based performance visualization system has been presented by Smoliar, Waterworth, & Kellock (1995). The displayed performance parameter include onset and offset timing, dynamics, articulation (referring to tone length), and tone onset asynchronies. However, the paper lacks a technical description of how these parameters were computed and implemented, but it becomes evident that this system works in an offline manner (computing the displayed parameters after recording). The elaborated MIDI-based visualization system comp-i (Hiraga, Mizaki, & Fujishiro, 2002) depicts performance information (onset and offset timing, dynamics, pitch) in a 3-dimensional space. Another approach aimed to explicitly exemplify higher-level performance information was provided by Hiraga & Matsuda (2004). They calculate and display local tempo change, local articulation, and local dynamics change. Both of the just mentioned systems work offline (Hiraga, Mizaki, & Fujishiro, 2002, Hiraga & Matsuda, 2004). More recent developments aim at developing computer software that allows relatively fast access to performance information, such as tone intensity or timing (MIDIator software from the Piano Pedagogy Research Lab in Ottawa 1 ). High-level approaches aim to impose specific models of emotions and emotional playing on music students via feedback about the goodness of fit of the human performance to the model (Juslin, Friberg, Schoonderwaldt, & Karlsson, 2004). However, such an approach seems problematic because it makes an aesthetic judgment by predefining a target performance on the basis of an artificial cognitive model. AIMS In this paper, we propose piano practice interfaces that support the pianist in enhancing the effectiveness of his or her daily practice load. They are not meant to be THE solution to all practice problems; instead, they focus deliberately on certain sub-tasks of piano playing. They provide in real time immediate visual feedback via the computer screen to the performer and require a MIDI-compatible piano. They are intelligent in that they listen to the pianist s playing and decide themselves what to display. 1 http://www.piano.uottawa.ca, seen in May 2006 1

Figure 1 Pattern Display. The left panel shows a 6-tone pattern in a pitch time space just after its last tone has been played. Circle size and color indicate dynamics (MIDI velocity), horizontal placement the timing deviation relative to the timing of the corresponding pattern tone of the previous pattern cycle (displayed as shaded disks). On the top-right, the auto-correlation pattern is shown (correlation coefficients against shift amount). The bottom-right panel depicts a screen-shot of the frugal beattracker (against time in s), with the present moment marked by now. Circles denote played tones, (red) lines tracked beats, and shaded lines expected beat onsets. On the right, the future expected beat is outlined. They are unobtrusive in that they do not require any direct interaction of the user (the pianist) via computer keyboard or mouse. We argue that any change of haptic modality (from piano keyboard to computer touch pad) would disrupt the flow of practice considerably. The proposed practice interfaces could run permanently in a practice session, getting sometimes the pianist s attention (e.g., when she is deliberately scrutinizing her own chord play), and sometimes none at all. The proposed practice interfaces currently consist of three parts which we will describe in the following: 1. A pattern display that identifies recurring pitch sequences and shows timing and dynamics deviations of the found pattern; 2. A chord display that displays timing asynchrony and tone intensity balance of chords (sequences of tones within a certain time window); and 3. An acoustic piano roll representation that visually models properties of the piano sound including decay and pedal interaction. The current version is implemented in JAVA 1.5 programming language and therefore runs on multiple computer platforms. PATTERN DISPLAY Let s assume the following practice situation: a piano student wants to improve his or her technical ability in performing fast accompaniment passages, such as Alberti bass figures frequently found in the Classical piano literature. For example, take a bass tone followed by two mediumpitched tones, all three alternated by a higher-pitched tone therefore a six-tone pattern. It repeats several times with slightly varying pitches, but with the same pitch contour. The student wants to know what it is that makes these figures sound irregular, dull, or the opposite swinging and vibrant. Our interface automatically identifies the above pattern after typically 2 or 3 repetitions and displays the timing and dynamics of its six tones in real time while the student is practicing. The student can therefore react on his or her irregularities and shape the performance of that pattern according to her taste. We describe in the following the three processing levels of this pattern interface. Pattern Finding Algorithm The pattern finding algorithm uses autocorrelation to detect a pattern. Sequences of relative pitch (pitch differences) are correlated to each other repeatedly, each time shifted one element further apart. 2 The smallest shift at which the correlation coefficient has a peak and is beyond a certain threshold is taken to be the cycle length of the 2 In case of a chord (more than one tone within 70 ms), the most prominent tone (in terms of MIDI velocity) is considered for calculation. 2

Figure 2 The Chord Display (voice number against time in ms) shows onset asynchronies of chords and the dynamics of each tone. Mean first-note lead and maximum chord spread are printed on top of the panel. pattern. In order to provide a more stable behavior of the pattern finding algorithm, cycle length is kept in a buffer of which the most frequent value (mode) is considered. The phase of the pattern is determined by the relative tone intensity within the pattern, the loudest tone being the first. Through this definition, the performer can re-adjust the pattern start as desired simply by emphasizing the first note of a pattern. Frugal Beat Tracking Parallel to the pattern finding algorithm, a simple (frugal) beattracker determines a fundamental tempo, from which estimates for future beats can be inferred. For programming efficiency, a very simple approach was followed (Pardo, 2004), although we assume that more elaborate beat tracking algorithms (such as Dixon, 2001) are more apt for this purpose. They will be integrated into the present system in the near future. Display As soon as a pattern has been established, its constituent tones are shown in a pitch time space. Individual tones are displayed as colored disks varying in size and intensity of color (reddishness) with tone intensity. The higher the MIDI velocity value, the more the yellow of soft tones turns into red and the larger the disk gets. Timing deviations are displayed against the beat estimate from the beattracker. A tone occurring too early is displaced leftwards with the corresponding previous tone of the pattern shown grey in the background. A more common way to display timing would be to show inter-onset interval (IOI) timing. However, we strongly prefer the present display over plotting IOIs because every IOI value stems from two tone onsets, so it would be hard for the performer to disentangle these two pieces of information. A screenshot of the components of the pattern display is shown in Figure 1. The left panel shows a 6-tone pattern just after its last tone has been played with pitch on the y axis (the middle-c key is marked grey) and time on the x axis. Circle size and color correspond to dynamics (MIDI velocity), horizontal placement to timing deviation relative to the timing of the corresponding pattern tone of the previous pattern cycle (displayed as shaded disks). On the top-right, the auto-correlation pattern is shown (correlation coefficients against shift amount), indicating a pattern period of six tones. The bottom-right panel depicts a screen-shot of the frugal beattracker (against time in s), with the present moment marked by now. Circles denote played tones, red (in case of black and white print: dark) lines successfully tracked beats, grey lines expected beat onsets. On the right, the future expected beat is outlined. Chord Display The second aspect of piano performance visualized by the proposed system concerns timing asynchronies and tone intensity balance of chords. Pianists almost never play nominally synchronous score notes entirely simultaneously. Studies on expressive piano performance report systematic trends in these asynchronies (Palmer, 1996). In particular, when melody tones are emphasized (played louder), their onsets typically occur around 30 ms ahead of the other chord tones (Goebl, 2001). Apart from that, the timbral color of a chord may be shaped deliberately by the performer by altering the intensity balance of the chord tones (Neuhaus, 1973). The proposed chord display detects potential chords in the MIDI data stream, calculates the first-tone lead and the chord spread, and displays the tones of the chords according to their pitch order (vertical axis) and time (horizontal axis). The intensity of the tones are reflected in size and color of the displayed disks (the louder, the larger and the more red on the same scale as in the pattern display). The leading tone is indicated by a black margin. A screenshot of a chord display example is shown in Figure 2. A chord is defined by a simple heuristic: tones belong to a chord when each individual chord tone is no more than 70 ms apart and the total spread does not exceed 300 ms. These figures were approximated from experience with a large corpus of piano performance data. 3

Figure 3 Acoustic Piano Roll. An excerpt of a performance of Schubert s G-flat Major Impromptu is shown (pitch against time in ms). Background shading of the panel corresponds to pedal press; color saturation of the tone bars depicts acoustic decay of the piano tones; thinner decay bars indicate tones that are still sounding after note off due to pedal press. Whenever a tone is played (i.e. when MIDI data is received), the display is updated and displays the last chord played. It works therefore in real time. As simple as this tool is, as explicit is its statement to the performer: It immediately exemplifies the relation of vertical dynamics differentiation to voice anticipation or lag. ACOUSTIC PIANO ROLL DISPLAY Piano roll displays are very common and can be found in every sequencer software package nowadays. They usually represent onset and duration of played tones through the position and extent of bars in a pitch time space. By displaying this, they provide important note information; however, essential data of piano performance is left out, that is e.g., pedal information (especially from the right pedal), dynamics of each tone, and the interaction between pedals and the sounding tones (These data are usually shown in separate channels or on demand; e.g., by clicking on a note bar, its velocity is shown.) Here, we decided to include all performance data derived from a piano into a single comprehensive piano roll representation. A screenshot of this interface is shown in Figure 3, displaying an excerpt of a performance of Schubert s G-flat Major Impromptu. In the pitch time space, we integrated a visual approximation of what is sounding at a given time on the basis of MIDI-like data from a piano performance. This includes beyond displaying the onset and offset of tones with bars the dynamics of each tone, represented by color intensity (on a color map between yellow and red), the decay of the piano tones (represented by color saturation), and the interaction of the right pedal and the tones that sound during pedal press (represented by prolonging the decaying tones with slightly thinner bars). The right pedal itself is shown as grey shading of the background. To model the tone decay of a piano, a set of piano samples recorded from a computer-controlled grand piano (Bösendorfer SE290) was analyzed in terms of their loudness decay over time. 25 different pitches distributed over the whole keyboard were played back on the computercontrolled piano in 5 different dynamic levels (MIDI velocity units from 30 to 110 in steps of 20), each tone sounding for a minimum of 5 seconds (for the low tones 10~s). Loudness representations of these 125 samples were measured in sone according to the Zwicker model (Zwicker & Fastl, 1999) (implemented by Pampalk, Rauber, & Merkl, 2002) and used to interpolate decay functions for all pitches and dynamic levels (MIDI velocities). Essentially, lower tones decay later than higher pitches; softer tones decay faster than loud ones. These data were linearly matched to saturation of an HSV color space; thus, a decaying tone is losing color with time and turning more and more into white. As the other parts of the proposed system, the acoustic piano roll works in real time. To avoid showing always the full range of possible tones of the piano (usually 88 keys from A0 to C8), the interface takes care of dynamically 4

adjusting the displayed pitch range on the vertical axis to the present performance. FINAL REMARKS This paper has presented visualization tools to be used in a pianist s every-day life at practicing piano. They have not yet been extensively tested by pianists in practice. In future work, usability studies will show advantages and drawbacks of the present tools which we will use for further improvements. Foreseen extensions of these interfaces will include other performance parameters (such as tone length), improvement of pattern finding and beat-tracking algorithms, and the identification of other pianistic subtasks. Furthermore, immediate display of touch characteristics could be realized with new computer-monitored instruments as, e.g., found in the latest version of Bösendorfer s computer-controlled piano ( CEUS ) that provides continuous key position data. Acknowledgments This research was supported by the Vienna Science and Technology Fund (WWTF, project CI010 Interfaces to Music ) and by the Austrian Science Fund (FWF, START project Y99-INF and a Schrödinger Fellowship to the first author, J2526). The OFAI acknowledges basic financial support by the Austrian Federal Ministries for Education, Science, and Culture and for Transport, Innovation, and Technology. We want to specially thank Simon Dixon for essential help while programming these tools in JAVA and for valuable comments on earlier versions of this manuscript. References Dixon, S. (2001). Automatic extraction of tempo and beat from expressive performances. Journal of New Music Research, 30(1), 39 58. Goebl, W. (2001). Melody lead in piano performance: Expressive device or artifact? Journal of the Acoustical Society of America, 110(1), 563 572. Hiraga, R., & Matsuda, N. (2004). Visualization of music performance as an aid to listener's comprehension, Proceedings of the Working Conference on Advanced Visual Interfaces (AVI) (pp. 103 106). Gallipoli, Italy: ACM Press. Hiraga, R., Mizaki, R., & Fujishiro, I. (2002). Performance visualization: a new challenge to music through visualization, Proceedings of the tenth ACM International Conference on Multimedia, Juan-les-Pins, France (pp. 239 242). New York: ACM Press. Juslin, P. N., Friberg, A., Schoonderwaldt, E., & Karlsson, J. (2004). Feedback learning of musical expressivity. In A. Williamon (Ed.), Musical Excellence. Strategies and Techniques to Enhance Performance (pp. 247 270). Oxford: Oxford University Press. Neuhaus, H. (1973). The Art of Piano Playing. London: Barrie & Jenkins. Palmer, C. (1996). On the assignment of structure in music performance. Music Perception, 14(1), 23 56. Pampalk, E., Rauber, A., & Merkl, D. (2002). Using smoothed data histograms for cluster visualization in self-organizing maps. In J. R. Dorronsoro (Ed.), Proceedings of the International Conference on Artificial Neural Networks (ICANN'02), Madrid, Spain (pp. 871 876). Berlin: Springer. Pardo, B. (2004). Tempo tracking with a single oscillator, Proceedings of the 5th International Conference on Music Information Retrieval (ISMIR2004) (pp. 154 157). Barcelona: Universitat Pompeu Fabra. Riley-Butler, K. (2001). Comparative performance analysis through feedback technology, Meeting of the Society for Music Perception and Cognition (SMPC2001), August 9 11, 2001 (pp. 27 28). Kingston, Ontario, Canada: Queen's University. Riley-Butler, K. (2002). Teaching expressivity: An aural visual feedback replication model, ESCOM 10th Anniversary Conference on Musical Creativity, April 5 8, 2002. Liège, Belgium: Université de Liège. Smoliar, S. W., Waterworth, J. A., & Kellock, P. R. (1995). pianoforte: A system for piano education beyond notation literacy, Proceedings of the ACM International Conference on Multimedia, San Francisco (pp. 457 465). New York: ACM Press. Webster, P. R. (2002). Computer-based technology and music teaching and learning. In R. J. Colwell & C. Richardson (Eds.), The New Handbook of Research on Music Teaching and Learning (pp. 416 439). Oxford: Oxford University Press. Welch, G. F., Himonides, E., Howard, D. M., & Brereton, J. (2004). VOXed: Technology as a meaningful teaching aid in the singing studio. In R. Parncutt & A. Kessler & F. Zimmer (Eds.), Conference on Interdisciplinary Musicology, April, 15 18. Graz, Austria: University of Graz. Zwicker, E., & Fastl, H. (1999). Psychoacoustics. Facts and Models (Second updated ed.). Berlin, Heidelberg: Springer. 5