The Trumpet Shall Sound: De-anonymizing jazz recordings

Similar documents
OBSERVED DIFFERENCES IN RHYTHM BETWEEN PERFORMANCES OF CLASSICAL AND JAZZ VIOLIN STUDENTS

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Introductions to Music Information Retrieval

Effects of acoustic degradations on cover song recognition

A. began in New Orleans during 1890s. B. Jazz a mix of African and European traditions. 1. Storyville District w/ Creoles of Color

Towards Music Performer Recognition Using Timbre Features

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis

TOWARD UNDERSTANDING EXPRESSIVE PERCUSSION THROUGH CONTENT BASED ANALYSIS

Music Similarity and Cover Song Identification: The Case of Jazz

Computational Modelling of Harmony

Data-Driven Solo Voice Enhancement for Jazz Music Retrieval

10 Visualization of Tonal Content in the Symbolic and Audio Domains

Copyright 2009 Pearson Education, Inc. or its affiliate(s). All rights reserved. NES, the NES logo, Pearson, the Pearson logo, and National

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

Track 2 provides different music examples for each style announced.

Articulation * Catherine Schmidt-Jones. 1 What is Articulation? 2 Performing Articulations

Tempo and Beat Analysis

MUSI-6201 Computational Music Analysis

PRESCOTT UNIFIED SCHOOL DISTRICT District Instructional Guide January 2016

NATIONAL SENIOR CERTIFICATE GRADE 12

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

THE SOUND OF SADNESS: THE EFFECT OF PERFORMERS EMOTIONS ON AUDIENCE RATINGS

Jazz Artist Project Directions:

Music Understanding and the Future of Music

Preview Only. Legal Use Requires Purchase. Emily JAZZ. Music by JOHNNY MANDEL Words by JOHNNY MERCER Arranged by LISA DeSPAIN INSTRUMENTATION

MODELING RHYTHM SIMILARITY FOR ELECTRONIC DANCE MUSIC

FINE ARTS Institutional (ILO), Program (PLO), and Course (SLO) Alignment

Query By Humming: Finding Songs in a Polyphonic Database

NCEA Level 2 Music (91275) 2012 page 1 of 6. Assessment Schedule 2012 Music: Demonstrate aural understanding through written representation (91275)

The song remains the same: identifying versions of the same piece using tonal descriptors

Improving Beat Tracking in the presence of highly predominant vocals using source separation techniques: Preliminary study

Senior High School Band District-Developed End-of-Course (DDEOC) Exam Study Guide

Preparatory Orchestra Performance Groups INSTRUMENTAL MUSIC SKILLS

Cambridge International Examinations Cambridge International General Certifi cate of Secondary Education

THE MUSIC ACADEMY AT CCTS.

Chord Classification of an Audio Signal using Artificial Neural Network

Level of Difficulty: Beginning Prerequisites: None

SAMPLE ASSESSMENT TASKS MUSIC JAZZ ATAR YEAR 11

UNIVERSITY OF DUBLIN TRINITY COLLEGE

THE BASIS OF JAZZ ASSESSMENT

Week 14 Music Understanding and Classification

WEST END BLUES / MARK SCHEME

SAMPLE ASSESSMENT TASKS MUSIC CONTEMPORARY ATAR YEAR 11

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors

Robert Alexandru Dobre, Cristian Negrescu

Advanced Orchestra Performance Groups

Middle School Band District-Developed End-of-Course (DDEOC) Exam Study Guide

Advanced Lesson Plan for Young Performers Initiative: Rockin In Rhythm BEFORE THE VIDEO

Middle School General Music Unit Plan Overview

LEVELS IN NATIONAL CURRICULUM MUSIC

LEVELS IN NATIONAL CURRICULUM MUSIC

Music Information Retrieval. Juan Pablo Bello MPATE-GE 2623 Music Information Retrieval New York University

Music Complexity Descriptors. Matt Stabile June 6 th, 2008

Simple Harmonic Motion: What is a Sound Spectrum?

Music Education. Test at a Glance. About this test

Music Information Retrieval with Temporal Features and Timbre

NATIONAL SENIOR CERTIFICATE GRADE 12

New Orleans. Storyville, French Opera House, 1900

Music for Alto Saxophone & Computer

Semi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis

Symphonic Pops Orchestra Performance Groups

Measuring & Modeling Musical Expression

Connecticut State Department of Education Music Standards Middle School Grades 6-8

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

La Salle University. I. Listening Answer the following questions about the various works we have listened to in the course so far.

REHEARSAL STRATEGIES I AIN T GOT NOTHIN BUT THE BLUES BY LOREN SCHOENBERG

TOWARDS IMPROVING ONSET DETECTION ACCURACY IN NON- PERCUSSIVE SOUNDS USING MULTIMODAL FUSION

THE POTENTIAL FOR AUTOMATIC ASSESSMENT OF TRUMPET TONE QUALITY

ESTIMATING THE ERROR DISTRIBUTION OF A TAP SEQUENCE WITHOUT GROUND TRUTH 1

Musicians Adjustment of Performance to Room Acoustics, Part III: Understanding the Variations in Musical Expressions

Middle School Chorus District-Developed End-of-Course (DDEOC) Exam Study Guide

Concise Guide to Jazz

REHEARSAL STRATEGIES HARLEM CONGO BY LOREN SCHOENBERG,

Music Standards for Band. Proficient Apply instrumental technique (e.g., fingerings, bowings, stickings, playing position, tone quality, articulation)

TERM 3 GRADE 5 Music Literacy

Timing In Expressive Performance

International School of Kenya

WHO IS WHO IN THE END? RECOGNIZING PIANISTS BY THEIR FINAL RITARDANDI

La Salle University MUS 150 Art of Listening Final Exam Name

Computer Coordination With Popular Music: A New Research Agenda 1

Jazz is a music genre that started in the early 1900's or earlier, within the African-American communities of the Southern United States.

K-12 Performing Arts - Music Standards Lincoln Community School Sources: ArtsEdge - National Standards for Arts Education

Psychophysiological measures of emotional response to Romantic orchestral music and their musical and acoustic correlates

Version 5: August Requires performance/aural assessment. S1C1-102 Adjusting and matching pitches. Requires performance/aural assessment

Greenwich Music Objectives Grade 3 General Music

CALCULATING SIMILARITY OF FOLK SONG VARIANTS WITH MELODY-BASED FEATURES

MUSIC WESTERN ART. Western Australian Certificate of Education Examination, Question/Answer Booklet. Stage 3

Experiments on musical instrument separation using multiplecause

OF THE ARTS ADMISSIONS GUIDE 2016 ACADEMY

Dimensional Music Emotion Recognition: Combining Standard and Melodic Audio Features

Before I proceed with the specifics of each etude, I would like to give you some general suggestions to help prepare you for your audition.

Automatic Construction of Synthetic Musical Instruments and Performers

HST 725 Music Perception & Cognition Assignment #1 =================================================================

DEVELOPMENTS IN INSTRUMENTAL JAZZ; 1910 TO THE PRESENT DAY: AOS3

Analytic Comparison of Audio Feature Sets using Self-Organising Maps

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

Creating a Successful Audition CD

first year charts Preview Only Legal Use Requires Purchase Pacific Attitude for jazz ensemble JAZZ VINCE GASSI INSTRUMENTATION

Specifying Features for Classical and Non-Classical Melody Evaluation

Music Source Separation

Transcription:

http://dx.doi.org/10.14236/ewic/eva2016.55 The Trumpet Shall Sound: De-anonymizing jazz recordings Janet Lazar Rutgers University New Brunswick, NJ, USA janetlazar@icloud.com Michael Lesk Rutgers University New Brunswick, NJ, USA lesk@acm.org We are experimenting with automated techniques to identify performers on jazz recordings by using stylistic measures of acoustic signals. Many early jazz recordings do not identify individual musicians, leaving them under-appreciated. We look at individual notes and phrasing for recognition of jazz trumpeters as an example. Jazz, performer identification, music analysis. 1. INTRODUCTION For much of the 20th century jazz recordings did not contain full listings of the performers; attributions would only name a group such as "Count Basie and his All American Rhythm Section" or "Duke Ellington and his Orchestra". Who were the actual performers? Our goal is to recognize them automatically, using jazz trumpeters as an example. The pictures below are all from Wikipedia. Louis Armstrong Harry James Wynton Marsalis Identification of flamenco singers and classical pianists has been studied before [Kroher 2014, Saunders 2008]; the jazz problem is more complex because there is no written score to be aligned with the notes played. However, experienced human listeners can recognize the performers, so the problem is feasible. Some researchers have invested in manual creation of the score [Abesser 2015] followed by a complex separation of the playing of each performer. We ve been looking at solo passages, identified by ear, but hoping to recognize them mechanically in the future. Why not try this as a very general machine learning problem? One could feed all the data into WEKA and sit back and watch. However, there isn t enough data: we have at most hundreds, not millions, of samples. Worse yet, there are many accidental properties of the acoustic signals. For example, different recording studios used microphones with different frequency limits. Until the 1950s many microphones recorded only up to 10kHz [Ford]. We would not wish to train a system on whether a recording was made at RCA in Camden, NJ or at Columbia in New York. What features would be characteristic of musical style? The diagram below is from [Ramirez 2007] and shows the intensity contour of a single note: 279

Multi-note features, derived from phrasing, include: Staccato/legato notes separated or continuous? Beat timing regularly spaced notes or ragged time. 2. ST. LOUIS BLUES What features might be exploited for machine classification? Single-note features include: Vibrato are notes steady or wavering? Tone complexity are the notes simple tones or many additional frequencies? Onset speed: do the notes rise quickly or slowly in intensity? Decay speed: do the notes stop quickly or does the performer tail off each note? For demonstration purposes, and to test software, we are using recordings of W. C. Handy s St. Louis Blues, written in 1914 and recorded more than 100 times. Here are sound spectrograms for snippets of sound by Louis Armstrong, Harry James and Wynton Marsalis. The software used in this paper includes BeatRoot [Dixon] and the MIR Toolbox [Lartillot]; we thank the creators and maintainers of these programs. Figure 1: Sound spectrograms of three trumpeters playing St. Louis Blues. Armstrong has the most complex sound (least dominated by the main note frequency) while Marsalis played fewer tones in each note. Marsalis playing is the most staccato; Armstrong and James played more continuously. Looking at frequency stability, Marsalis plays with the most stable notes, i.e., the least vibrato, while James is a bit more variable and Armstrong still more. For another comparison, Figure 2 shows sound spectrograms for about 0.2 seconds (a single note, roughly) taken from three different places for each performer. All are again St. Louis Blues. Look here at the extent to which the pure note and its overtones dominate the signal. Marsalis is playing with the least sound beyond the specific note; James has a more complex note, with extra overtones; and Armstrong has much more in the way of low frequency components in the notes. 280

Figure 2: Single-note sound spectrograms. Figure 3: Single-note, Benny Goodman (top), Harry James (bottom). What would we see if we compared two different clarinetists? The next pair of spectra, in Figure 4, show Benny Goodman above and Artie Shaw. Figure 4: Benny Goodman (top), Artie Shaw (bottom). Compared to the trumpet both are weighted to lower frequency and simpler in structure. Comparing these two, Benny Goodman s notes are purer and contain fewer frequencies. 281

3. CLARINET AND HARP What happens if we look at other instruments? Figure 3 shows a comparison of Benny Goodman (above, clarinet) with Harry James (below, trumpet). Note the generally lower frequency spectrum of the clarinet and the complexity of the trumpet notes in terms of frequencies. As another example, we took sound spectra of four different harpists. In Figure 5, the top left spectrum is Lucia Bova, top right is Csilla Gulyas, bottom left is Maria Graf and bottom right is Judy Loman. They are all playing C. P. E. Bach s Harp Sonata in G major, Wq 139. Figure 5: Four harpists. Left column: Lucia Bova, Maria Graf. Right column: Csilla Gulyas, Judy Loman. We then calculated the basic tempo for each and the attack time, measuring off the sound spectra, using two samples for each player. Below is a plot showing that the performers differ but each tends to repeat her characteristic choices. 4. CONCLUSION The longer-run purpose of this work is to help with cataloging old recordings. Since music had no requirement for compulsory deposit in the United States until the 1970s, the Library of Congress has an unusually incomplete collection. Rutgers University, at its Institute of Jazz Studies in Newark, NJ, holds more than 100,000 sound recordings, and this is the largest jazz repository. Unfortunately, practical difficulties, such as fragility of records, and legal difficulties, such as copyright ownership of recordings made by companies that may be long out of business, impede the study of these recordings. We hope that by automating the creation of metadata we can help the scholars and bring recognition to artists whose contributions are fading from memory and insufficiently documented. 5, REFERENCES Figure 6: Distribution of tempi and attack time. Abesser, J., Cano, E., Frieier, K., Pfleidere, M., Zaddach, W.-G. (2015) Score-Informed Analysis of Intonation and Pitch Modulations in Jazz Solos. 16 th conference, International Society for Music Information Retrieval. 282

Dixon, S. (2001) An Interactive Beat Tracking and Visualisation System. In Proceedings of the 2001 International Computer Music Conference (ICMC'2001). Ford, T. (2005) A recent history of ribbon microphones. Ty Ford Audio and Video, Blogspot. http://tyfordaudiovideo.blogspot.com/2012/02/recen t-history-of-ribbon-microphones.html (retrieved 14 June 2016). Saunders, C., Hardoon, D., Shawe-Taylor, J., Widmer, G. (2008) Using string kernels to identify famous performers from their playing style. Intelligent Data Analysis, 12(4), pp. 425 440. Kroher, Nadine; Gómez, Emilia (2014). Automatic Singer Identification For Improvisational Styles Based On Vibrato, Timbre And Statistical Performance Descriptors. Proceedings ICMCISMCI2014 (Joint International Computer Music and Sound and Music conference), 14 20 September, Athens, Greece, pp. 1160 1165. Lartillot, O., Toiviainen, P., and Eerola, T, (2008). A matlab toolbox for music information retrieval. In C. Preisach, P. D. H. Burkhardt, P. D. L. Schmidt-Thieme, and P. D. R. Decker (eds.), Data Analysis, Machine Learning and Applications, Studies in Classification, Data Analysis, and Knowledge Organization, pp. 261 268. Springer, Berlin/Heidelberg. Ramirez, R., Maestre, E., Pertusa, A., Gómez, E., and Serra, X. (2007) Performance-based interpreter identification in saxophone audio recordings. IEEE Transactions on Circuits and Systems for Video Technology, 17(3), pp. 356 364. 283