Timing In Expressive Performance

Similar documents
Tempo and Beat Analysis

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)

Automatic music transcription

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

Introductions to Music Information Retrieval

CS229 Project Report Polyphonic Piano Transcription

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

Beat Tracking by Dynamic Programming

Analysis of local and global timing and pitch change in ordinary

HUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH

Music Information Retrieval for Jazz

Music Information Retrieval (MIR)

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

Multidimensional analysis of interdependence in a string quartet

A prototype system for rule-based expressive modifications of audio recordings

6.UAP Project. FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System. Daryl Neubieser. May 12, 2016

Music Representations. Beethoven, Bach, and Billions of Bytes. Music. Research Goals. Piano Roll Representation. Player Piano (1900)

MODELING RHYTHM SIMILARITY FOR ELECTRONIC DANCE MUSIC

Voice & Music Pattern Extraction: A Review

EVALUATION OF A SCORE-INFORMED SOURCE SEPARATION SYSTEM

GCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam

Design of a Speaker Recognition Code using MATLAB

Tempo and Beat Tracking

Effects of acoustic degradations on cover song recognition

CSC475 Music Information Retrieval

Laboratory Assignment 3. Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB

Music Database Retrieval Based on Spectral Similarity

y POWER USER MUSIC PRODUCTION and PERFORMANCE With the MOTIF ES Mastering the Sample SLICE function

Music Radar: A Web-based Query by Humming System

Searching for Similar Phrases in Music Audio

La Salle University. I. Listening Answer the following questions about the various works we have listened to in the course so far.

Violin Timbre Space Features

The BAT WAVE ANALYZER project

Supervised Learning in Genre Classification

Audio. Meinard Müller. Beethoven, Bach, and Billions of Bytes. International Audio Laboratories Erlangen. International Audio Laboratories Erlangen

THE importance of music content analysis for musical

Music Structure Analysis

Onset Detection and Music Transcription for the Irish Tin Whistle

Week 14 Music Understanding and Classification

6.5 Percussion scalograms and musical rhythm

CS 591 S1 Computational Audio

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

Audio Structure Analysis

TECHNIQUES FOR AUTOMATIC MUSIC TRANSCRIPTION. Juan Pablo Bello, Giuliano Monti and Mark Sandler

Soundprism: An Online System for Score-Informed Source Separation of Music Audio Zhiyao Duan, Student Member, IEEE, and Bryan Pardo, Member, IEEE

Getting Started. Connect green audio output of SpikerBox/SpikerShield using green cable to your headphones input on iphone/ipad.

Robert Alexandru Dobre, Cristian Negrescu

Lets go through the chart together step by step looking at each bit and understanding what the Chart is asking us to do.

Beethoven, Bach, and Billions of Bytes

Controlling Musical Tempo from Dance Movement in Real-Time: A Possible Approach

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Topic 11. Score-Informed Source Separation. (chroma slides adapted from Meinard Mueller)

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

Chapter Five: The Elements of Music

Composer Style Attribution

Video-based Vibrato Detection and Analysis for Polyphonic String Music

Temporal coordination in string quartet performance

Drum Source Separation using Percussive Feature Detection and Spectral Modulation

A Parametric Autoregressive Model for the Extraction of Electric Network Frequency Fluctuations in Audio Forensic Authentication

Data Driven Music Understanding

Music Similarity and Cover Song Identification: The Case of Jazz

Lecture 10 Harmonic/Percussive Separation

ESTIMATING THE ERROR DISTRIBUTION OF A TAP SEQUENCE WITHOUT GROUND TRUTH 1

Topic 10. Multi-pitch Analysis

CZT vs FFT: Flexibility vs Speed. Abstract

hit), and assume that longer incidental sounds (forest noise, water, wind noise) resemble a Gaussian noise distribution.

Music Information Retrieval (MIR)

ECE438 - Laboratory 4: Sampling and Reconstruction of Continuous-Time Signals

Automatic Music Clustering using Audio Attributes

NAA ENHANCING THE QUALITY OF MARKING PROJECT: THE EFFECT OF SAMPLE SIZE ON INCREASED PRECISION IN DETECTING ERRANT MARKING

Query By Humming: Finding Songs in a Polyphonic Database

Music Alignment and Applications. Introduction

Automatic Rhythmic Notation from Single Voice Audio Sources

A PSYCHOACOUSTICAL INVESTIGATION INTO THE EFFECT OF WALL MATERIAL ON THE SOUND PRODUCED BY LIP-REED INSTRUMENTS

Lab P-6: Synthesis of Sinusoidal Signals A Music Illusion. A k cos.! k t C k / (1)

Interacting with a Virtual Conductor

An Examination of Foote s Self-Similarity Method

A Parametric Autoregressive Model for the Extraction of Electric Network Frequency Fluctuations in Audio Forensic Authentication

Rhythm related MIR tasks

Master's Theses and Graduate Research

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

Spectrum Analyser Basics

AutoChorale An Automatic Music Generator. Jack Mi, Zhengtao Jin

MUSI-6201 Computational Music Analysis

Computer Coordination With Popular Music: A New Research Agenda 1

MMEA Jazz Guitar, Bass, Piano, Vibe Solo/Comp All-

Noise. CHEM 411L Instrumental Analysis Laboratory Revision 2.0

MUSIC is a ubiquitous and vital part of the lives of billions

Pitfalls and Windfalls in Corpus Studies of Pop/Rock Music

PHYSICS OF MUSIC. 1.) Charles Taylor, Exploring Music (Music Library ML3805 T )

SAMPLE ASSESSMENT TASKS MUSIC GENERAL YEAR 12

Music Information Retrieval

Fast Ethernet Consortium Clause 25 PMD-EEE Conformance Test Suite v1.1 Report

Acoustic and musical foundations of the speech/song illusion

Honours Project Dissertation. Digital Music Information Retrieval for Computer Games. Craig Jeffrey

TOWARDS IMPROVING ONSET DETECTION ACCURACY IN NON- PERCUSSIVE SOUNDS USING MULTIMODAL FUSION

PULSE-DEPENDENT ANALYSES OF PERCUSSIVE MUSIC

ARECENT emerging area of activity within the music information

NENS 230 Assignment #2 Data Import, Manipulation, and Basic Plotting

Time Signature Detection by Using a Multi Resolution Audio Similarity Matrix

Transcription:

Timing In Expressive Performance 1 Timing In Expressive Performance Craig A. Hanson Stanford University / CCRMA MUS 151 Final Project

Timing In Expressive Performance Timing In Expressive Performance 2 Introduction Rhythmic timing is arguably the most essential aspect of musical performance. This research project presents an automated method for rhythmic analysis of an audio file (.wav). The analysis process is accomplished through the Matlab programming language in a series of M-Files that may be downloaded here: http://ccrma.stanford.edu/~chanson9/151/151.html. The end goal of this analysis is the extraction of rhythmic inflections in recorded performance. The underlying basis of the Matlab process is the work of Dan Ellis at Columbia University and his paper on Beat Tracking by Dynamic Programming (Ellis 2007 pp 5-10). My additions and alterations to his original code allow the user to run one function and receive analytical charts and information relating to the timing of a particular performance in relation to an extracted tempo. This may be used in conjunction with listening exercises to determine the rhythmic signature of a particular performance and/or performer. My hope is that this process will allow for a deeper understanding of individual rhythmic musical expression. Matlab Methodology The Matlab programs submitted with this paper should provide a generic framework for rhythmic analysis of a.wav file. There are six main M-Files that comprise this application. Upon the successful execution of these six m-files (using the run function), the user receives 3 main outputs: a tempo estimate, an onset strength envelope chart and a rhythmic inflection chart. Below are details of each M-File. 1. RUN.M This m-file is all that a user needs in order to work with this system. Assuming the rest of the m-files are present in the user s working directory, this file will take care of launching the required processes to build the analysis. The user must provide two simple arguments to the run function: a.wav filename and a performer name. The subsequent m- files are then called as needed to perform their various functions. 2. PROCESSWAV.M This m-file is at the heart of everything that occurs in the analysis.

Timing In Expressive Performance 3 It takes in a file and a performer s name and outputs note onsets, beat detection and the steady state comparison at the detected tempo. In order to output this data, the file must perform the following functions: a. Re-sample the input signal to 8kHz to remove unwanted high frequency onset detection interference. b. Take the spectrogram (STFT) of the re-sampled signal. c. Convert the spectrogram to a perceptual representation by mapping 40 Mel bands via a weighted sum of spectrogram values. This is accomplished by calling the fft2melmx function. d. Set any negative values to zero in order to obtain the correct onset strength values. e. Sum the positive differences across all frequency bands. f. Convolve with a Gaussian envelope to obtain onset strength envelope. g. Identify song tempo through the use of tempo2.m. 3. FFT2MELMX.M Fft2melmx.m is called from the processwav function in order to convert the normal FFT frequency bins into Mel bins. Mel frequency band representation is perceptually more accurate than a regular spectrogram. The only arguments required for this function to run is the fft size and sample rate. The function returns a matrix of weights that are applied to the original FFT bins for conversion. 4. TEMPO2.M Tempo2.m is called from the processwav function in order to identify the tempo of the song. As a function, tempo2 takes in an input signal, sample rate, BPM mode estimate and a BPM spread. It will then output a BPM, autocorrelation of note onsets, mel-frequency spectrogram and an onset envelope for the given input file. Information regarding this process of onset detection was found in chapter 3 of Tristan Jehan s PhD thesis Creating Music By Listening (Tristan sec. 3.4). The main function of this m-file is to perform the autocorrelation analysis as follows: a. Perform an autocorrelation of onset strength envelope (i.e. take the inner product of envelope with delayed versions of itself. For delays that line up many peaks, a

Timing In Expressive Performance 4 large correlation occurs). b. Window the autocorrelation at default BPM of 110 (this can be changed, but it works for most songs). The highest peak should be representative of the tempo. 5. LOCALMAX.M Localmax.m is called from tempo2.m in order to identify the maximum peak during the autocorrelation analysis in tempo2.m. 6. CREATEFIGURE.M This m-file is called from within run.m and will generate the charts seen as the analysis finishes processing. These six files comprise all that is necessary to perform the rhythmic analysis of.wav files outlined in this paper. The user should be able to place these M-Files in their Matlab working directory and execute the run function to obtain the information outlined in this paper. Initial Test Analysis For my initial test analysis, I ran seven.wav files through this system and analyzed their respective outputs. Six of these.wav files were actual musical performances. One.wav file was a rigid midi interpretation of the same musical piece. The musical source material for my test analysis was Bach s Invention #1 in C Major. I chose this piece because it is typically performed in a rather straight rhythmic manner and therefore, tempo/onset detection would be more accurate for a first test. In addition, this piece has been interpreted by many musicians, giving me a variety of recorded performance styles to work with. The subsequent page contains a chart of the musical performances chosen for this analysis.

Timing In Expressive Performance 5 Figure 1: Performance Chart There are three main outputs from the Matlab scripts that I used extensively in my analyses of the various performances: the tempo estimate, the onset envelope chart, and the rhythmic inflection chart. The base performance with which all other performances were compared was the rigid midi interpretation, which had a tempo estimate of 86 BPM. The onset envelope chart and rhythmic inflection charts are shown below: Figure 2: Raw Onset Envelope Chart - MIDI

Timing In Expressive Performance 6 Figure 3: Rhythmic Inflection Chart - MIDI The above charts show quite a bit of information pertaining to the midi performance of Bach s Invention #1. The raw onset envelope chart shows the individual note onsets during the course of the entire song. This is measured by the change in Mel magnitude spectrogram bands as the song progresses. The basis for this is the idea that note onsets are typically accompanied by sharp changes in the spectral content of a given sound. By identifying the maximum spectral change on a perceptual basis, you can get a decent idea of the note onsets throughout the song. The rhythmic inflection chart shows how straight the performance is throughout the analyzed recording. There is a diagonal line in black that represents a strict evenness of tempo. The blue line plotted on top of it represents the actual performance inflections contained in the recording. The vertical separation lines represent each musical beat within the piece of music 1. By quickly analyzing this chart, one can see that the midi representation of the music is almost rhythmically exact. This exact sense of timing does not typically lend itself to an expressive musical performance, hence the reason why midi representations of music are usually dull and lifeless. 1 It is of interest to note that the tempo detection is not always 100% accurate in this methodology. The major problem is the identification of tempo being twice as fast or twice as slow as the actual tempo. Because of this, the beat locations may actually show half-beats or double-beats. It is good to keep this in mind as other performances are analyzed.

Timing In Expressive Performance 7 The findings from the midi analysis above become interesting when we compare them with the results of analysis of the real performances. The subsequent pages show both onset envelope and rhythmic inflection charts for each of the performances I analyzed and a brief description of the conclusions drawn from each analysis. 1. Glenn Gould Piano Glenn Gould s performance was clean, with good note onsets and a steady rhythm. The onset envelope chart shows strong onsets and clear delineation between notes. The rhythmic inflection chart really captures the essence of Gould s performance as his inflection line is very close to that of the diagonal reference. Note that the tempo detected during analysis was twice that of the actual tempo. Figure 4: Glenn Gould Figure 5: Janine Jansen

Timing In Expressive Performance 8 2. Janine Jansen Violin and Viola Jansen s rendition for violin and viola is rhythmically very expressive. I believe this is evidenced in the rhythmic inflection chart, as there is a downturn in rhythmic speed halfway through the invention. This downturn is evidenced upon listening. However, I also hear a lot of smaller rhythmic inflections throughout the performance, which are not readily apparent in the chart above. It is interesting to note the dense nature of the onset envelope. I suspect both of these issues are due to the instrumentation used in the performance. Figure 6: Evgeni Koroliov Figure 7: Wanda Landowska

Timing In Expressive Performance 9 3. Evgeni Koroliov - Piano Koroliov s performance is a bit more straight than the above chart eludes. However, this performance showcases some variety from the other performances I analyzed. One can even make out a bit of swing applied to the Bach s bassline in the recording. Koroliov definitely takes his rhythmic liberties, but stays fairly regular. 4. Wanda Landowska Harpsichord Landowska s performance is much slower and more even than most of the others. The onset detection and rhythmic inflections both seemed to work quite well, which I believe is due to the timbre of the harpsichord and the quality of the recording. Figure 8: Max Reger Figure 9: Andras Schiff

Timing In Expressive Performance 10 5. Max Reger Organ Certainly the most adventurous of all renditions of Bach s Invention, Max Reger takes his liberties with rhythm throughout his performance. It is evident in both of the charts above that Reger was uneven throughout the piece, however, he uses this artistically to shape Bach into a work of his own. 6. Andras Schiff Piano Andras Schiff provides the most heavily articulated performance of the lot. I believe this is the reason why the tempo was detected at 129.31 BPM when it is actually played around 60 BPM. The rhythmic inflection chart shows the liberties taken in this performance. Following the tempo while listening will confirm that he does speed up and slow down his performance. However, this chart may be slightly inflated due to articulations throughout the piece. Conclusion The findings from the initial test analysis show that the rhythmic content of a performance may be captured by automated means. Tangible results have been presented from a test scenario using actual musical performances of varied instrumentation. Further analysis of other genres of music are necessary to gain a full understanding of the extent to which this process will work for identifying the underlying rhythmic character of a performance. In the future, I would like to include functionality for the midi file itself to be imported into Matlab and analyzed without having to bounce it to audio. A direct analysis of the midi file and performance deviations can then be made by including Dan Ellis research on aligning midi scores to music audio (Ellis 2009). That said, as a first case test scenario, this proves to be an easy to use system for identifying rhythmic fluctuation throughout a various interpretations of a piece of classical music.

Timing In Expressive Performance 11 References Ellis, D. (2007, 16 July). Beat Tracking by Dynamic Programming. Retrieved March 1, 2009 from the World Wide Web: http://www.ee.columbia.edu/~dpwe/pubs/ellis07- beattrack.pdf Ellis, D. (2009, 26 February). Aligning Midi Scores to Music Audio. Retrieved March 1, 2009 from the World Wide Web: http://labrosa.ee.columbia.edu/matlab/alignmidiwav/#3 Jehan, T. (2005, 17 June). Creating Music By Listening. Retrieved March 3, 2009 from the World Wide Web: http://web.media.mit.edu/~tristan/phd/dissertation/chapter3.html