Topics in Computer Music Instrument Identification. Ioanna Karydi

Similar documents
Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

Topic 10. Multi-pitch Analysis

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

POLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES

Music Information Retrieval with Temporal Features and Timbre

LEARNING SPECTRAL FILTERS FOR SINGLE- AND MULTI-LABEL CLASSIFICATION OF MUSICAL INSTRUMENTS. Patrick Joseph Donnelly

MUSI-6201 Computational Music Analysis

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

Efficient Computer-Aided Pitch Track and Note Estimation for Scientific Applications. Matthias Mauch Chris Cannam György Fazekas

Automatic Identification of Instrument Type in Music Signal using Wavelet and MFCC

Recognising Cello Performers Using Timbre Models

MUSICAL NOTE AND INSTRUMENT CLASSIFICATION WITH LIKELIHOOD-FREQUENCY-TIME ANALYSIS AND SUPPORT VECTOR MACHINES

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

Features for Audio and Music Classification

Semi-supervised Musical Instrument Recognition

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES

HUMANS have a remarkable ability to recognize objects

GCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam

Recognising Cello Performers using Timbre Models

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

Acoustic Scene Classification

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

A Survey on: Sound Source Separation Methods

Interactive Classification of Sound Objects for Polyphonic Electro-Acoustic Music Annotation

Melody Retrieval On The Web

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

Supervised Learning in Genre Classification

Musical Acoustics Lecture 15 Pitch & Frequency (Psycho-Acoustics)

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

Classification of Timbre Similarity

Automatic Labelling of tabla signals

Creating a Feature Vector to Identify Similarity between MIDI Files

Tempo and Beat Analysis

THE importance of music content analysis for musical

Musical Instrument Identification based on F0-dependent Multivariate Normal Distribution

Music Database Retrieval Based on Spectral Similarity

Music Complexity Descriptors. Matt Stabile June 6 th, 2008

TYING SEMANTIC LABELS TO COMPUTATIONAL DESCRIPTORS OF SIMILAR TIMBRES

Statistical Modeling and Retrieval of Polyphonic Music

Singer Traits Identification using Deep Neural Network

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

Automatic Construction of Synthetic Musical Instruments and Performers

Cross-Dataset Validation of Feature Sets in Musical Instrument Classification

Transcription of the Singing Melody in Polyphonic Music

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

Analysis, Synthesis, and Perception of Musical Sounds

Comparison Parameters and Speaker Similarity Coincidence Criteria:

Contextual music information retrieval and recommendation: State of the art and challenges

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

CS229 Project Report Polyphonic Piano Transcription

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

Video-based Vibrato Detection and Analysis for Polyphonic String Music

Towards Music Performer Recognition Using Timbre Features

Musical instrument identification in continuous recordings

pitch estimation and instrument identification by joint modeling of sustained and attack sounds.

Application Of Missing Feature Theory To The Recognition Of Musical Instruments In Polyphonic Audio

A repetition-based framework for lyric alignment in popular songs

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

CS 591 S1 Computational Audio

A Need for Universal Audio Terminologies and Improved Knowledge Transfer to the Consumer

MIRAI: Multi-hierarchical, FS-tree based Music Information Retrieval System

HUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH

A prototype system for rule-based expressive modifications of audio recordings

Music BCI ( )

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

A System for Automatic Chord Transcription from Audio Using Genre-Specific Hidden Markov Models

Transcription An Historical Overview

TOWARD UNDERSTANDING EXPRESSIVE PERCUSSION THROUGH CONTENT BASED ANALYSIS

Computational Modelling of Harmony

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis

Improving Frame Based Automatic Laughter Detection

Automatic Music Clustering using Audio Attributes

A Music Retrieval System Using Melody and Lyric

Music Genre Classification and Variance Comparison on Number of Genres

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng

WE ADDRESS the development of a novel computational

Pitch Perception and Grouping. HST.723 Neural Coding and Perception of Sound

Automatic Piano Music Transcription

The song remains the same: identifying versions of the same piece using tonal descriptors

Perceiving Differences and Similarities in Music: Melodic Categorization During the First Years of Life

MOTIVATION AGENDA MUSIC, EMOTION, AND TIMBRE CHARACTERIZING THE EMOTION OF INDIVIDUAL PIANO AND OTHER MUSICAL INSTRUMENT SOUNDS

Chord Classification of an Audio Signal using Artificial Neural Network

EE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function

Effects of acoustic degradations on cover song recognition

TIMBRE-CONSTRAINED RECURSIVE TIME-VARYING ANALYSIS FOR MUSICAL NOTE SEPARATION

IMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM

MELODY EXTRACTION BASED ON HARMONIC CODED STRUCTURE

Pitch. The perceptual correlate of frequency: the perceptual dimension along which sounds can be ordered from low to high.

Musical Instrument Recognizer Instrogram and Its Application to Music Retrieval based on Instrumentation Similarity

Multiple classifiers for different features in timbre estimation

Outline. Why do we classify? Audio Classification

We realize that this is really small, if we consider that the atmospheric pressure 2 is

Transcription:

Topics in Computer Music Instrument Identification Ioanna Karydi

Presentation overview What is instrument identification? Sound attributes & Timbre Human performance The ideal algorithm Selected approaches Overview Steps Results Conclusions 2

What is instrument identification? Instrument identification The procedure of training a learning machine to identify an instrument by hearing a sound clip. 3

Sound attributes Loudness Pitch Timbre Duration Pressure +++ + + + Frequency + +++ ++ + Spectrum + + +++ + Duration + + + +++ Envelop + + ++ + Dependence of subjective qualities of sound on physical parameters.[1] 4

Timbre Timbre: That attribute of auditory sensation in terms of which a listener can judge that two sounds similarly presented and having the same loudness and pitch are dissimilar (American National Standards Institute) 5

Human performance Musical instrument identification systems attempt to reproduce how humans can recognize and identify the musical sounds populating their environment. [1] Number of instruments Identification rates 2 97.6% 3 95.4% 9 90.2% 27 55.7% 6

An ideal Algorithm Characteristics: Generalization Robustness Meaningful behavior Reasonable computational requirements Modularity An ideal musical instrument identification system [1] 7

Agostini (1/4) Overview Segmentation and estimation of pitch Features describe spectral characteristics of monophonic sounds (intended to be heard as a single channel of sound perceived as coming from one position) Dataset composed by 1.007 tones from 27 musical instruments (from orchestral sounds to pop/electronic instruments) Classification using pattern recognition techniques 8

Agostini (2/4) Feature Extraction A set of features related to the harmonic properties of sounds is extracted from monophonic musical signals Steps: Audio Segmentation Pitch Tracking Calculation of Features 9

Agostini (3/4) Classification Techniques Discriminant Analysis Canonical discriminant analysis Separates two instrument labels in a plane by means of a line Quadratic Discriminant Analysis the hyper-surfaces delimiting the regions of classification are quadratic forms Support Vector Machines A representation of the examples as points in space, mapped so that the examples of the separate categories are divided by a clear gap that is as wide as possible k-nearest Neighbors An object is classified by a majority vote of its neighbors, with the object being assigned to the instrument label most common among its k nearest neighbors 10

Agostini (4/4) Results Classifiers performances for different number of instruments [2] 11

Barbedo Tzanetakis (1/6) Overview 4 instruments: alto saxophone (S), violin (V), piano (P) and acoustic guitar (G) Pairwise comparison scheme Requires single note from each instrument, the more isolated partials available the best accuracy Correct identification even in presence of interference Accurate estimates are possible using only a small number of partials, without requiring the entire note Using only one carefully designed feature more effective than traditional machine learning methods 12

Barbedo Tzanetakis (2/6) Setting it up 1.000 mixtures used None of the instrument samples used in the test set Separation accuracy for each pair of instruments [3] 13

Barbedo Tzanetakis (3/6) Setting it up Segment the signal Determine the number of instruments present in each segment Estimate the fundamental frequency (F0) of each instrument 14

Barbedo Tzanetakis (4/6) Feature selection & extraction The features to be extracted depend directly on which instruments are being considered The feature selection for each pair aimed for the best linear separation The features are calculated individually for each partial 15

Barbedo Tzanetakis (5/6) Instrument identification procedure Number of instruments and respective fundamental frequencies are known Pairwise comparison for each isolated partial - an instrument is chosen as the winner for each pair Repeat the same procedure for all partials related to that fundamental frequency Then, the predominant instrument is taken as the correct one Repeat the same procedure for all fundamental frequencies 16

Barbedo Tzanetakis (6/6) Experimental results Isolated partials available Accuracy One Close to 91% More than six Up to 96% 3 factors for such good performance: 1. Only four instruments are considered 2. All instruments are taken from the same database 3. A very effective feature was found to the difficult pair piano - acoustic guitar, which significantly improved the overall results 17

Conclusions Agostini: Common used classifiers could not provide results similar to QDA, since extracted features follow such distribution. Feature set still lacking of temporal descriptors of the signal. Next steps: introduction of new features and inclusion of percussive sounds and live-instruments. Barbedo-Tzanetakis Robust and accurate results. Proved that pairwise comparison approach is effective. Next steps: more instruments and signals from other databases. 18

Thank you for your attention! 19

References [1] Nicolas D. Chétry, (2006), Computer Models for Musical Instrument Identification http://c4dm.eecs.qmul.ac.uk/papers/2006/chetry06- phdthesis.pdf [2] G. Agostini, M. Longari, and E. Pollastri. Content-based classification of musical instrument timbres. International Workshop on Content-Based Multimedia Indexing, 2001. http://www.music.mcgill.ca/~ich/classes/mumt614/mir/agostinitimbre.pdf [3] Jayme Garcia Arnal Barbedo and George Tzanetakis (2010). Instrument identification in polyphonic music signals based on individual partials. http://webhome.csc.uvic.ca/~gtzan/output/icassp2010gtzan.pdf 20