Semi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis

Similar documents
Music Representations

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

PHYSICS OF MUSIC. 1.) Charles Taylor, Exploring Music (Music Library ML3805 T )

Music Processing Introduction Meinard Müller

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

Transcription An Historical Overview

Music Radar: A Web-based Query by Humming System

Further Topics in MIR

Beethoven, Bach, and Billions of Bytes

Audio. Meinard Müller. Beethoven, Bach, and Billions of Bytes. International Audio Laboratories Erlangen. International Audio Laboratories Erlangen

Laboratory Assignment 3. Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB

Introductions to Music Information Retrieval

Toward a Computationally-Enhanced Acoustic Grand Piano

Music Information Retrieval

Automatic music transcription

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis

A Beat Tracking System for Audio Signals

Reference Manual. Using this Reference Manual...2. Edit Mode...2. Changing detailed operator settings...3

Singing accuracy, listeners tolerance, and pitch analysis

Using the new psychoacoustic tonality analyses Tonality (Hearing Model) 1

6.UAP Project. FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System. Daryl Neubieser. May 12, 2016

Music Representations

We realize that this is really small, if we consider that the atmospheric pressure 2 is

Measurement of overtone frequencies of a toy piano and perception of its pitch

jsymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada

COMPUTATIONAL INVESTIGATIONS INTO BETWEEN-HAND SYNCHRONIZATION IN PIANO PLAYING: MAGALOFF S COMPLETE CHOPIN

Automatic Rhythmic Notation from Single Voice Audio Sources

From quantitative empirï to musical performology: Experience in performance measurements and analyses

An Effective Filtering Algorithm to Mitigate Transient Decaying DC Offset

Robert Alexandru Dobre, Cristian Negrescu

A Matlab toolbox for. Characterisation Of Recorded Underwater Sound (CHORUS) USER S GUIDE

Tempo and Beat Tracking

Tempo and Beat Analysis

Week 14 Music Understanding and Classification

A REAL-TIME SIGNAL PROCESSING FRAMEWORK OF MUSICAL EXPRESSIVE FEATURE EXTRACTION USING MATLAB

Simple Harmonic Motion: What is a Sound Spectrum?

Author Index. Absolu, Brandt 165. Montecchio, Nicola 187 Mukherjee, Bhaswati 285 Müllensiefen, Daniel 365. Bay, Mert 93

Pitch. The perceptual correlate of frequency: the perceptual dimension along which sounds can be ordered from low to high.

Hugo Technology. An introduction into Rob Watts' technology

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

Topic 10. Multi-pitch Analysis

Machine Learning Term Project Write-up Creating Models of Performers of Chopin Mazurkas

Query By Humming: Finding Songs in a Polyphonic Database

AUTOMATIC ACCOMPANIMENT OF VOCAL MELODIES IN THE CONTEXT OF POPULAR MUSIC

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

Automatic Singing Performance Evaluation Using Accompanied Vocals as Reference Bases *

ECE 4220 Real Time Embedded Systems Final Project Spectrum Analyzer

MAutoPitch. Presets button. Left arrow button. Right arrow button. Randomize button. Save button. Panic button. Settings button

A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES

2. AN INTROSPECTION OF THE MORPHING PROCESS

Computer Coordination With Popular Music: A New Research Agenda 1

HST 725 Music Perception & Cognition Assignment #1 =================================================================

6.5 Percussion scalograms and musical rhythm

Note on Posted Slides. Noise and Music. Noise and Music. Pitch. PHY205H1S Physics of Everyday Life Class 15: Musical Sounds

Transcription of the Singing Melody in Polyphonic Music

Musical Acoustics Lecture 16 Interval, Scales, Tuning and Temperament - I

Automatic meter extraction from MIDI files (Extraction automatique de mètres à partir de fichiers MIDI)

CSC475 Music Information Retrieval

Guide to Analysing Full Spectrum/Frequency Division Bat Calls with Audacity (v.2.0.5) by Thomas Foxley

Music Representations. Beethoven, Bach, and Billions of Bytes. Music. Research Goals. Piano Roll Representation. Player Piano (1900)

UNIVERSITY OF DUBLIN TRINITY COLLEGE

Interacting with a Virtual Conductor

ESTIMATING THE ERROR DISTRIBUTION OF A TAP SEQUENCE WITHOUT GROUND TRUTH 1

Experiments on musical instrument separation using multiplecause

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

y POWER USER MUSIC PRODUCTION and PERFORMANCE With the MOTIF ES Mastering the Sample SLICE function

4. ANALOG TV SIGNALS MEASUREMENT

A DISCRETE FILTER BANK APPROACH TO AUDIO TO SCORE MATCHING FOR POLYPHONIC MUSIC

Music Information Retrieval (MIR)

Director Musices: The KTH Performance Rules System

AN ALGORITHM FOR LOCATING FUNDAMENTAL FREQUENCY (F0) MARKERS IN SPEECH

Figure 1: Feature Vector Sequence Generator block diagram.

Hidden Markov Model based dance recognition

Topic 4. Single Pitch Detection

Improving Polyphonic and Poly-Instrumental Music to Score Alignment

The Relationship Between Auditory Imagery and Musical Synchronization Abilities in Musicians

Musical Acoustics Lecture 15 Pitch & Frequency (Psycho-Acoustics)

Lab experience 1: Introduction to LabView

Music Complexity Descriptors. Matt Stabile June 6 th, 2008

Getting Started with the LabVIEW Sound and Vibration Toolkit

Tutorial 3 Normalize step-cycles, average waveform amplitude and the Layout program

MUSIC CURRICULM MAP: KEY STAGE THREE:

EMS DATA ACQUISITION AND MANAGEMENT (LVDAM-EMS) MODEL 9062-C

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng

THE importance of music content analysis for musical

Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals

Acoustic Measurements Using Common Computer Accessories: Do Try This at Home. Dale H. Litwhiler, Terrance D. Lovell

A prototype system for rule-based expressive modifications of audio recordings

2 MHz Lock-In Amplifier

TECHNIQUES FOR AUTOMATIC MUSIC TRANSCRIPTION. Juan Pablo Bello, Giuliano Monti and Mark Sandler

Getting Started. Connect green audio output of SpikerBox/SpikerShield using green cable to your headphones input on iphone/ipad.

Development of an Optical Music Recognizer (O.M.R.).

Analysis of local and global timing and pitch change in ordinary

10 Visualization of Tonal Content in the Symbolic and Audio Domains

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

Speaking in Minor and Major Keys

LOUDNESS EFFECT OF THE DIFFERENT TONES ON THE TIMBRE SUBJECTIVE PERCEPTION EXPERIMENT OF ERHU

DEVELOPMENT OF MIDI ENCODER "Auto-F" FOR CREATING MIDI CONTROLLABLE GENERAL AUDIO CONTENTS

MELODY EXTRACTION FROM POLYPHONIC AUDIO OF WESTERN OPERA: A METHOD BASED ON DETECTION OF THE SINGER S FORMANT

Goebl, Pampalk, Widmer: Exploring Expressive Performance Trajectories. Werner Goebl, Elias Pampalk and Gerhard Widmer (2004) Introduction

Transcription:

Semi-automated extraction of expressive performance information from acoustic recordings of piano music Andrew Earis

Outline Parameters of expressive piano performance Scientific techniques: Fourier transform & wavelet transform CHARM: The semi-automated system

Measurable features of expressive performance Timing (note onset and offset) Dynamics (variations in intensity of notes and chords) Pitch ampli tude time

Timing parameterisation Note onset time. The time at which the note is played. Note offset time. The time at which the note is released. Attack duration. The time from the initial sound until its amplitude is at a maximum. Sustain duration. The time from maximum sound amplitude until the note is released. Decay time. The time from when the note is released until the sound ceases.

Dynamics parameterisation Power of note Maximum amplitude (amplitude of attack) Sustain amplitude (amplitude, or change of amplitude, of the sustain)

Pitch parameterisation Fundamental frequency of note Frequency of individual partials (and how they deviate from the precise harmonic frequencies) Vibrato Frequency of vibrato Amplitude of vibrato Glissando

Aims to extract the expressive performance information from acoustic recordings of piano music with reference to a digitised version of the musical score of the work being performed.

Simple flowchart Digitised musical score Event (note or chord) onset times dynamics Acoustic recording Note-on-note synchronisation algorithm Individual note onset times offset times dynamics Expressive performance data of individual performance

Methods The location in time of individual notes and chords within an acoustic recording must be identified and related to a digitised score of the work being performed. This is a classical time-frequency analysis problem

Fundamental research issues A number of constraints on the time-frequency resolution required. Required frequency resolution determined by the interval between adjacent semitones proportional to the frequency of the lower semitone Spacing varies from about 1.6Hz at the bottom of the piano keyboard to over 120Hz at the top the fundamental frequencies ranging from 27.5Hz to just over 4kHz. Required timing resolution The shortest note duration in rapid note passages is around 50ms. The human ear is capable of perceiving differences of as little as 10ms.

Possible approaches Filterbank techniques Time-frequency analysis (using Fourier Transforms, wavelets and spectrographs) Modelling

Short Time Fourier analysis

Wavelet analysis

Problems with simple approach Synchronization between score and acoustic recording relies on stepping through each event in the score in turn If a particular note or chord in the score is not played, the whole system can break down Other problems too much noise in the recording a high degree of rubato chords are played with too much asynchrony The longer the passage of music to be analysed, the more likely it is that such a system will break down. Time-consuming manual checking of the data would also required to ensure accuracy.

CHARM: Semi-automated system A semi-automated expression extractor, requiring user input for checking and verifying results as the algorithm proceeds i.e. note-on-note. Input to system: MIDI score, acoustic waveform Technology incorporated: filterbank/wavelet-based note onset detector Note/event parameter measurements: onset time, dynamics

Algorithm plan STEP 1 Digitisation of musical score STEP 2 Manual beat tapping to acoustic recording STEP 3 Approximate performance score Estimate of main beat onset times STEP 4 Measurement of actual main beat onset times using algorithm STEP 5 Manual correction of main beat onset times STEP 6 Calculation of other event onset times and all event dynamics, individual note onset times / dynamicsnote onset timings and event/individual note dynamics

Advantages Quick to use, results can be verified as they go along know they are correct! Very easy to use Wide range of scores can be used Quick and easy to develop and automate further Results are verified by the user as measurements are made, ensuring that all results are correct and can be used reliably for musicological analysis Further enhancements can be made (e.g. increase level of automation, sophistication of parameters)

STEP 1 Digitisation of musical score Score is scanned using a standard flatbed scanner Converted to symbolic data with Sharpeye Errors are corrected manually Symbolic data is then converted to the Humdrum data format This is outputted as ASCII tabular data

STEP 2 Manual beat tapping to acoustic recording Standard beat tapping procedures backbeat.exe average of 20 complete taps initially!

STEP 3 Averaged manual beat tapped data combined with the digitized musical score -> Approximate performance score Column Description 1 2 3 4 5 abstime duration note metric level bar number Average absolute time in ms of beat tapped timings Expected duration of note in ms based on score duration MIDI number of note pitch 1 = downbeat 0 = beat -1 = offbeat 6 absolute beat measured from beginning of work

Approximate performance score: Mazurka in A minor op 7 no 2 1912 646 76 1 0 0 2558 463 77 0 1 1 3021 154 76-1 1 1.75 3175 603 57 0 1 2 3175 603 62 0 1 2 3175 603 65 0 1 2 3175 603 74 0 1 2 3778 652 57 1 1 3 3778 652 62 1 1 3 3778 652 65 1 1 3 3778 652 77 1 1 3 4430 1111 77 0 2 4 4914 627 60 0 2 5 4914 627 65 0 2 5

STEP 4 Correction of main beat onset times (from beat tapped times) using wavelet-based algorithm, programmed in MATLAB

Parameters Search window parameters Wavelet parameter - k Wavelet parameter - LPF Wavelet tuning Feedback analysis

Search window parameters

Wavelet parameters - k

Wavelet tuning

Feedback analysis

STEP 5 Manual correction of main beat onset times Sonic Visualiser Output table of corrected main beat onset times

STEP 6 Estimation of subbeat onset times from main beat onset times Correction of estimated subbeat onset times using algorithm Estimation of main beat and subbeat dynamics Estimation of individual note onset times

End product MATLAB code converted to standard.exe files Input data required Excel control spreadsheet - for batch processing of data Input/output file names Analysis parameters Main beat beat tapped input data spreadsheet.wav sound recording

End product Three different algorithms mainbeatcorrection.exe correction of main beat tapped onset times measurement of main beat dynamics (in db) subbeatmeasurement.exe measurement of sub beat onset times measurement of sub beat dynamics individualnotemeasurement.exe measurement of individual note onset timings and dynamics Will be available (or a variation of this) on CHARM website as downloadable resource

End product For example... 3 recordings of Chopin s Mazurka in A minor op 7 no 2 Friedman (1930) Rubinstein (1939) Indjic (1988) Graphs of inter-onset interval for each event in score dynamics of each event (scaled in dbs)

Inter-onset interval 3 2.5 2 1.5 Friedman (1930) Rubinstein (1939) Indjic (1988) 1 0.5 0 0 20 40 60 80 100 120 140 160 180 200 220 240 260 280 300 Beat

Event dynamics 50.00 40.00 30.00 20.00 Friedman (1930) Rubinstein (1939) Indjic (1988) 10.00 0.00 0 9 17 26 34 43 51 60 69 77 86 94 103-10.00 Beat

Conclusions and future work Further scientific development (wavelet parameters, shape, dynamics measurement) Extension to note offset times Testing and refinement of user interface