Automatic Labelling of tabla signals

Similar documents
Hidden Markov Model based dance recognition

Automatic Notes Generation for Musical Instrument Tabla

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

Automatic Rhythmic Notation from Single Voice Audio Sources

Concatenated Tabla Sound Synthesis to Help Musicians

Automatic Music Genre Classification

Jazz Melody Generation and Recognition

ACOUSTIC FEATURES FOR DETERMINING GOODNESS OF TABLA STROKES

Topics in Computer Music Instrument Identification. Ioanna Karydi

MUSI-6201 Computational Music Analysis

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

Composer Style Attribution

A System for Automatic Chord Transcription from Audio Using Genre-Specific Hidden Markov Models

Outline. Why do we classify? Audio Classification

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

Automatic Detection of Hindustani Talas

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

MODAL ANALYSIS AND TRANSCRIPTION OF STROKES OF THE MRIDANGAM USING NON-NEGATIVE MATRIX FACTORIZATION

Week 14 Music Understanding and Classification

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors

THE importance of music content analysis for musical

Music Alignment and Applications. Introduction

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t

Automatic Music Clustering using Audio Attributes

Automatic Identification of Instrument Type in Music Signal using Wavelet and MFCC

Transcription of the Singing Melody in Polyphonic Music

Supervised Learning in Genre Classification

Phone-based Plosive Detection

TRACKING THE ODD : METER INFERENCE IN A CULTURALLY DIVERSE MUSIC CORPUS

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES

TOWARD UNDERSTANDING EXPRESSIVE PERCUSSION THROUGH CONTENT BASED ANALYSIS

Musicological perspective. Martin Clayton

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University

Application Of Missing Feature Theory To The Recognition Of Musical Instruments In Polyphonic Audio

Improving Frame Based Automatic Laughter Detection

Computational Modelling of Harmony

Music 209 Advanced Topics in Computer Music Lecture 4 Time Warping

... A Pseudo-Statistical Approach to Commercial Boundary Detection. Prasanna V Rangarajan Dept of Electrical Engineering Columbia University

Music Understanding and the Future of Music

Automatic Piano Music Transcription

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

A Bayesian Network for Real-Time Musical Accompaniment

Detecting Musical Key with Supervised Learning

Automatic Construction of Synthetic Musical Instruments and Performers

International Journal of Advance Engineering and Research Development. Digi-Taal Instrument

arxiv: v1 [cs.ir] 16 Jan 2019

Music Information Retrieval for Jazz

Semi-supervised Musical Instrument Recognition

TOWARDS IMPROVING ONSET DETECTION ACCURACY IN NON- PERCUSSIVE SOUNDS USING MULTIMODAL FUSION

Categorization of ICMR Using Feature Extraction Strategy And MIR With Ensemble Learning

WE ADDRESS the development of a novel computational

GENERATIVE RHYTHMIC MODELS

MODELS of music begin with a representation of the

Automatic music transcription

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

Audio. Meinard Müller. Beethoven, Bach, and Billions of Bytes. International Audio Laboratories Erlangen. International Audio Laboratories Erlangen

EE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

Music Information Retrieval

Topic 10. Multi-pitch Analysis

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

Analysis and Clustering of Musical Compositions using Melody-based Features

IMPROVING RHYTHMIC SIMILARITY COMPUTATION BY BEAT HISTOGRAM TRANSFORMATIONS

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University

Feature-Based Analysis of Haydn String Quartets

Searching for Similar Phrases in Music Audio

HUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH

/$ IEEE

Beethoven, Bach, and Billions of Bytes

DETECTION OF KEY CHANGE IN CLASSICAL PIANO MUSIC

A DISCRETE MIXTURE MODEL FOR CHORD LABELLING

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception

ON FINDING MELODIC LINES IN AUDIO RECORDINGS. Matija Marolt

Lecture 11: Chroma and Chords

TOWARDS CHARACTERISATION OF MUSIC VIA RHYTHMIC PATTERNS

Tempo and Beat Analysis

Toward Automatic Music Audio Summary Generation from Signal Analysis

Musical Hit Detection

Multiple instrument tracking based on reconstruction error, pitch continuity and instrument activity

Rhythm related MIR tasks

DETECTION OF SLOW-MOTION REPLAY SEGMENTS IN SPORTS VIDEO FOR HIGHLIGHTS GENERATION

1 Introduction Steganography and Steganalysis as Empirical Sciences Objective and Approach Outline... 4

Topic 11. Score-Informed Source Separation. (chroma slides adapted from Meinard Mueller)

Transcription and Separation of Drum Signals From Polyphonic Music

STOCHASTIC MODELING OF A MUSICAL PERFORMANCE WITH EXPRESSIVE REPRESENTATIONS FROM THE MUSICAL SCORE

Drum Stroke Computing: Multimodal Signal Processing for Drum Stroke Identification and Performance Metrics

Video-based Vibrato Detection and Analysis for Polyphonic String Music

Tempo and Beat Tracking

Available online at ScienceDirect. Procedia Computer Science 46 (2015 )

Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES

Algorithms for melody search and transcription. Antti Laaksonen

Music Radar: A Web-based Query by Humming System

Musical Instrument Identification based on F0-dependent Multivariate Normal Distribution

Analysis of local and global timing and pitch change in ordinary

Transcription:

ISMIR 2003 Oct. 27th 30th 2003 Baltimore (USA) Automatic Labelling of tabla signals Olivier K. GILLET, Gaël RICHARD

Introduction Exponential growth of available digital information need for Indexing and Retrieval technique For musical signals, a transcription would include: Descriptors such as genre, style, instruments of a piece Descriptors such as beat, note, chords, nuances, etc Many efforts in instrument recognition (Kaminskyj2001, Martin 1999, Marques & al. 1999 Brown 1999, Brown & al.2001, Herrera & al.2000, Eronen2001) Less efforts in percussive instrument recognition (Herrera & al. 2003, Paulus&al.2003, McDonald&al.1997) Most effort on isolated sounds Almost no effort on non-western instrument recognition OBJECTIVE :Automatic transcription of real performances of an Indian instrument: the tabla Page 2

Outline Introduction Presentation of the tabla Transcription of tabla phrases Architecture of the system Features extraction Learning and classification Experimental results Database and evaluation protocols Results Tablascope: a fully integrated environment Description & applications Demonstration Conclusion Page 3

Presentation of the tabla The tabla: an percussive instrument played in Indian classical and semi-classical music The Dayan: wooden treble drum played by the right hand The Bayan: metallic bass drum played by the left hand Page 4

Presentation of the tabla (2) Musical tradition in India is mostly oral Use of mnemonic syllables (or bol ) for each stroke Common bols: Ge, Ke (bayan bols), Na, Tin, Tun, Ti, Te (dayan bols) Dha (Na+Ge), Dhin (Tin + Ge), Dhun (Tun + Ge) Some specificities of this notation system Different bols may sound very similar (ex. Ti and Te) Existence of «words» : «TiReKiTe or «GeReNaGe» A mnemonic may change depending on the context Complex rythmic structure based on Matra (i.e main beat), Vibhag (i.e measure) and avartan (i.e phrase) Page 5

Presentation of tabla (3) In summary: A tabla phrase is then composed of successive bols of different duration (note, half note, quarter note) embeded in a rythmic structure Grouping characteristics (words) : similarity with spoken and written languages: Interest of «Language models» or sequence models In this study, the transcription is limited to the recognition of successives bols The relative duration (note, half note, quarter note) of each bol. Page 6

Transcription of tabla phrases Architecture of the system Page 7

Parametric representation Segmentation in strokes Extraction of a low frequency envelope (sampled at 220.5 Hz) Simple Onset detection based on the difference between two successives samples of the envelope. Tempo extraction Estimated as the maximum of the autocorrelation function of the envelope signal in the range {60 240 bpm} Page 8

Features extraction Ge Na Dha = Ge + Na Ti Ke Page 9

Features extraction 4 frequency bands B1 = [0 150] Hz B2 = [150 220] Hz B3 = [220 380] Hz B4 = [700 900] Hz In the case of single mixture, each band is modelled by a Gaussian Feature vector F = f 1..f 12 (mean, variance and relative weight of each of the 4 Gaussians) Page 10

Learning and Classification of bols 4 classification techniques were used. K-nearest Neighbors (k-nn) Naive Bayes Kernel density estimator HMM sequence modelling Page 11

Learning and Classification of bols Context-dependant models (HMM) Page 12

Learning and Classification of bols Hidden Markov Models States: a couple of Bols B 1 B 2 is associated to each state Transitions: if state i is labelled by B 1 B 2 and j by B 2 B 3 then the transition from state to state is given by: Emissions probabilities: Each state i labelled by B 1 B 2 emits a feature vector according to a distribution characteristics of the bol B 2 preceded by B 1 Page 13

Learning and Classification of bols Training Transition probabilities are estimated by counting occurrences in the training database Emission probabilities are estimated with mean and variance estimators on the set of feature vectors in the case of simple Gaussian model 8 iterations of the Expectation-Maximisation (EM) algorithm in the case of a mixture model Recognition Performed using the traditionnal Viterbi algorithm Page 14

Experimental results Database 64 phrases with a total of 5715 bols A mix of long compositions with themes / variations (kaïda), shorter pieces (kudra) and basic taals. 3 specific sets corresponding to three different tablas: Tabla quality Dayan tuning Recording quality Tabla #1 Low (cheap) in C#3 Studio equipment Tabla #2 High In D3 Studio equiment Tabla #3 High In D3 Noisier environment Page 15

Evaluation protocols Protocol #1: Cross-validation procedure Database split in10 subsets (randomly selected) 9 subsets for training, 1 subset for testing Iteration by rotating the 10 subsets Results are average of the 10 runs Protocol #2: Training database consists in 100% of 2 sets Test is 100% of the remining sets Different instruments and/or conditions are used for training and testing Page 16

Experimental results (protocol #1) Page 17

Experimental results (protocol #2) HMM approaches are more robust to variability Simpler classifiers fail to generalise and to adapt to different recording conditions or instruments Page 18

Experimental results Confusion matrix by bol category (HMM 4-grams, 2 mixture classifier) Page 19

Tablascope: a fully integrated environment Applications: Tabla transcription Tabla sequence synthesis Tabla-controlled synthesizer Page 20

Conclusion A system for automatic labelling of tabla signals was presented Low error rate for transcription (6.5%) Several applications were integrated in a friendly environment called Tablascope. This work can be generalised to other types of percussive instruments still need a larger database to confirm the results.. Page 21