Data Driven Music Understanding

Similar documents
Data Driven Music Understanding

Music Information Retrieval for Jazz

Extracting Information from Music Audio

Extracting and Using Music Audio Information

Lecture 11: Chroma and Chords

Searching for Similar Phrases in Music Audio

Lecture 15: Research at LabROSA

Effects of acoustic degradations on cover song recognition

Lecture 12: Alignment and Matching

A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION

Content-based music retrieval

Lecture 10 Harmonic/Percussive Separation

MUSI-6201 Computational Music Analysis

Music Information Retrieval

Tempo and Beat Analysis

Beethoven, Bach, and Billions of Bytes

A Survey of Audio-Based Music Classification and Annotation

Music Representations. Beethoven, Bach, and Billions of Bytes. Music. Research Goals. Piano Roll Representation. Player Piano (1900)

Topic 10. Multi-pitch Analysis

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)

THE importance of music content analysis for musical

Music Genre Classification and Variance Comparison on Number of Genres

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

Music Alignment and Applications. Introduction

MELODY EXTRACTION BASED ON HARMONIC CODED STRUCTURE

CTP431- Music and Audio Computing Music Information Retrieval. Graduate School of Culture Technology KAIST Juhan Nam

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University

Topic 11. Score-Informed Source Separation. (chroma slides adapted from Meinard Mueller)

GCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam

Automatic Rhythmic Notation from Single Voice Audio Sources

Music Mood Classification - an SVM based approach. Sebastian Napiorkowski

Music Information Retrieval

Classification-Based Melody Transcription

Sparse Representation Classification-Based Automatic Chord Recognition For Noisy Music

Recognition and Summarization of Chord Progressions and Their Application to Music Information Retrieval

Lecture 9 Source Separation

Audio. Meinard Müller. Beethoven, Bach, and Billions of Bytes. International Audio Laboratories Erlangen. International Audio Laboratories Erlangen

DOWNBEAT TRACKING WITH MULTIPLE FEATURES AND DEEP NEURAL NETWORKS

MUSIC is a ubiquitous and vital part of the lives of billions

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES

Music Similarity and Cover Song Identification: The Case of Jazz

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Further Topics in MIR

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

Trevor de Clercq. Music Informatics Interest Group Meeting Society for Music Theory November 3, 2018 San Antonio, TX

Supervised Musical Source Separation from Mono and Stereo Mixtures based on Sinusoidal Modeling

Beat-Synchronous Chroma Representations for Music Analysis

Automatic Piano Music Transcription

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

Music Synchronization. Music Synchronization. Music Data. Music Data. General Goals. Music Information Retrieval (MIR)

Classification-based melody transcription

Automatic music transcription

Content-based Music Structure Analysis with Applications to Music Semantics Understanding

EE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function

Singer Recognition and Modeling Singer Error

Music Genre Classification

Tempo and Beat Tracking

Computational Modelling of Harmony

CS 591 S1 Computational Audio

A CHROMA-BASED SALIENCE FUNCTION FOR MELODY AND BASS LINE ESTIMATION FROM MUSIC AUDIO SIGNALS

Beethoven, Bach und Billionen Bytes

Singing Pitch Extraction and Singing Voice Separation

A New Method for Calculating Music Similarity

Transcription of the Singing Melody in Polyphonic Music

Outline. Why do we classify? Audio Classification

Features for Audio and Music Classification

The MPC X & MPC Live Bible 1

Piano Transcription MUMT611 Presentation III 1 March, Hankinson, 1/15

Book: Fundamentals of Music Processing. Audio Features. Book: Fundamentals of Music Processing. Book: Fundamentals of Music Processing

Chord Classification of an Audio Signal using Artificial Neural Network

arxiv: v1 [cs.sd] 8 Jun 2016

Music Information Retrieval (MIR)

Week 14 Music Understanding and Classification

A CLASSIFICATION-BASED POLYPHONIC PIANO TRANSCRIPTION APPROACH USING LEARNED FEATURE REPRESENTATIONS

Introductions to Music Information Retrieval

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis

Video-based Vibrato Detection and Analysis for Polyphonic String Music

Universal Parallel Computing Research Center The Center for New Music and Audio Technologies University of California, Berkeley

Data-Driven Solo Voice Enhancement for Jazz Music Retrieval

Performance Improvement of Music Mood Classification Using Hyper Music Features

On Human Capability and Acoustic Cues for Discriminating Singing and Speaking Voices

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

Automatic Music Similarity Assessment and Recommendation. A Thesis. Submitted to the Faculty. Drexel University. Donald Shaul Williamson

Subjective Similarity of Music: Data Collection for Individuality Analysis

gresearch Focus Cognitive Sciences

Music Information Retrieval. Juan P Bello

Beat Tracking by Dynamic Programming

Chord Recognition. Aspects of Music. Musical Chords. Harmony: The Basis of Music. Musical Chords. Musical Chords. Music Processing.

Modeling Music Similarity: Signal-based Models of Subjective Preference Daniel P.W. Ellis, Electrical Engineering, Columbia University

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

Automatic Music Genre Classification

MUSIC CONTENT ANALYSIS : KEY, CHORD AND RHYTHM TRACKING IN ACOUSTIC SIGNALS

MUSICAL INSTRUMENT RECOGNITION USING BIOLOGICALLY INSPIRED FILTERING OF TEMPORAL DICTIONARY ATOMS

EVALUATION OF A SCORE-INFORMED SOURCE SEPARATION SYSTEM

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

Semi-supervised Musical Instrument Recognition

Aalborg Universitet. Feature Extraction for Music Information Retrieval Jensen, Jesper Højvang. Publication date: 2009

Efficient Vocal Melody Extraction from Polyphonic Music Signals

Transcription:

Data Driven Music Understanding Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Engineering, Columbia University, NY USA http://labrosa.ee.columbia.edu/ 1. Motivation: What is Music? 2. Eigenrhythms 3. Melodic-Harmonic Fragments 4. Example Applications Data-driven music understanding - Ellis 2008-10-03 p. 1 /24

LabROSA Overview Information Extraction Music Machine Learning Recognition Separation Retrieval Speech Environment Signal Processing Data-driven music understanding - Ellis 2008-10-03 p. 2 /24

1. Motivation: What is music? What does music evoke in a listener s mind?? Which are the things that we call music? Data-driven music understanding - Ellis 2008-10-03 p. 3 /24

Oodles of Music What can you do with a million tracks? Data-driven music understanding - Ellis 2008-10-03 p. 4 /24

Re-use in Music 60 Scatter of PCA(3:6) of 12x16 beatchroma What are the most popular chord progressions in pop music? 50 40 30 20 10 60 50 40 30 20 10 10 20 30 40 50 60 10 20 30 40 50 60 Data-driven music understanding - Ellis 2008-10-03 p. 5 /24

Potential Applications Compression Classification Manipulation Data-driven music understanding - Ellis 2008-10-03 p. 6 /24

2. Eigenrhythms: Drum Track Structure To first order, all pop music has the same beat: Ellis & Arroyo ISMIR 04 Can we learn this from examples? Data-driven music understanding - Ellis 2008-10-03 p. 7 /24

Basis Sets Combine a few basic patterns to make a larger dataset data weights X = W H patterns 1 = 0-1 Data-driven music understanding - Ellis 2008-10-03 p. 8/24

Drum Pattern Data Tempo normalization + downbeat alignment Data-driven music understanding - Ellis 2008-10-03 p. 9 /24

NMF Eigenrhythms Posirhythm 1 Posirhythm 2 HH HH SN SN BD BD Posirhythm 3 Posirhythm 4 HH HH SN SN BD BD Posirhythm 5 Posirhythm 6 HH HH SN SN BD BD 0 50 100 150 200 250 300 350 400 0 50 100 150 200 250 300 350 samples (@ 2 1 2 3 4 1 2 3 4 beats (@ 120 Nonnegative: only add beat-weight 0.1 0-0.1 Data-driven music understanding - Ellis 2008-10-03 p. 10/24

Eigenrhythm BeatBox Resynthesize rhythms from eigen-space Data-driven music understanding - Ellis 2008-10-03 p. 11 /24

3. Melodic-Harmonic Fragments How similar are two pieces? Can we find all the pop-music clichés? Data-driven music understanding - Ellis 2008-10-03 p. 12/24

MFCC Features Used in speech recognition Let It Be (LIB-1) - log-freq specgram freq / Hz 5915 1396 329 MFCCs Coefficient 12 10 8 6 4 2 Noise excited MFCC resynthesis (LIB-2) freq / Hz 5915 1396 329 0 5 10 15 20 25 time / sec Data-driven music understanding - Ellis 2008-10-03 p. 13/24

Chroma Features To capture musical content Let It Be - log-freq specgram (LIB-1) freq / Hz 6000 1400 300 chroma bin B A G E D C Chroma features Shepard tone resynthesis of chroma (LIB-3) freq / Hz 6000 1400 300 MFCC-filtered shepard tones (LIB-4) freq / Hz 6000 1400 300 0 5 10 15 20 25 time / sec Data-driven music understanding - Ellis 2008-10-03 p. 14/24

Beat-Synchronous Chroma Compact representation of harmonies Let It Be - log-freq specgram (LIB-1) freq / Hz 6000 1400 300 Onset envelope + beat times chroma bin B A G E D C Beat-synchronous chroma Beat-synchronous chroma + Shepard resynthesis (LIB-6) freq / Hz 6000 1400 300 0 5 10 15 20 25 time / sec Data-driven music understanding - Ellis 2008-10-03 p. 15/24

Finding Cover Songs Ellis & Poliner 07 freq / khz Little similarity in surface audio... Let It Be - The Beatles Let It Be / Beatles / verse 1 4 3 2 freq / khz Let It Be - Nick Cave Let It Be / Nick Cave / verse 1 4 3 2 1 1 0 2 4 6 8 10 time / sec 0 2 4 6 8 10 time / se chroma.. but appears in beat-chroma G F D C Beat-sync chroma features chroma G F D C Beat-sync chroma features A 5 10 15 20 25 beats A 5 10 15 20 25 beat Data-driven music understanding - Ellis 2008-10-03 p. 16/24

Finding Common Fragments Cluster beat-synchronous chroma patches chroma bins G F D C A G F D C A G F D C A G F D C A #32273-13 instances #51917-13 instances #65512-10 instances #9667-9 instances #10929-7 instances #61202-6 instances #55881-5 instances 5 10 15 20 a20-top10x5-cp4-4p0 #68445-5 instances 5 10 15 time / beats Data-driven music understanding - Ellis 2008-10-03 p. 17/24

Clustered Fragments chroma bins G F D C A depeche mode 13-Ice Machine 199.5-204.6s roxette 03-Fireworks 107.9-114.8s G F D C roxette 04-Waiting For The Rain 80.1-93.5s tori amos 11-Playboy Mommy 157.8-171.3s A 5 10 15 20 5 10 15 time / beats... for a dictionary of common themes? Data-driven music understanding - Ellis 2008-10-03 p. 18/24

4. Example Applications: Music Discovery Berenzweig & Ellis 03 Connecting listeners to musicians Data-driven music understanding - Ellis 2008-10-03 p. 19/24

Playlist Generation Mandel, Poliner, Ellis 06 Incremental learning of listeners preferences Data-driven music understanding - Ellis 2008-10-03 p. 20/24

MajorMiner: Music Tagging Describe music using words Mandel & Ellis 07, 08 Data-driven music understanding - Ellis 2008-10-03 p. 21/24

Music Transcription Training data and features: MIDI, multi-track recordings, playback piano, & resampled audio (less than 28 mins of train audio). Normalized magnitude STFT. Classification: N-binary SVMs (one for ea. note). Independent frame-level classification on 10 ms grid. Dist. to class bndy as posterior. Temporal Smoothing: Two state (on/off) independent HMM for ea. note. Parameters learned from training data. Find Viterbi sequence for ea. note. feature representation classification posteriors hmm smoothing Poliner & Ellis 05, 06, 07 feature vector Data-driven music understanding - Ellis 2008-10-03 p. 22/24

MEAPsoft Music Engineering Art Projects collaboration between EE and Computer Music Center with Douglas Repetto, Ron Weiss, and the rest of the MEAP team Data-driven music understanding - Ellis 2008-10-03 p. 23/24

Conclusions Low-level features Classification and Similarity browsing discovery production Music audio Melody and notes Key and chords Tempo and beat Music Structure Discovery modeling generation curiosity Lots of data + noisy transcription + weak clustering musical insights? Data-driven music understanding - Ellis 2008-10-03 p. 24/24