AUTOMATIC TIMBRE CLASSIFICATION OF ETHNOMUSICOLOGICAL AUDIO RECORDINGS

Size: px
Start display at page:

Download "AUTOMATIC TIMBRE CLASSIFICATION OF ETHNOMUSICOLOGICAL AUDIO RECORDINGS"

Transcription

1 AUTOMATIC TIMBRE CLASSIFICATION OF ETHNOMUSICOLOGICAL AUDIO RECORDINGS Dominique Fourer, Jean-Luc Rouas, Pierre Hanna, Matthias Robine LaBRI - CNRS UMR University of Boreaux {fourer, rouas, hanna, robine}@labri.fr ABSTRACT Automatic timbre characterization of auio signals can help to measure similarities between souns an is of interest for automatic or semi-automatic atabases inexing. The most effective methos use machine learning approaches which require qualitative an iversifie training atabases to obtain accurate results. In this paper, we introuce a iversifie atabase compose of worlwie nonwestern instruments auio recorings on which is evaluate an effective timbre classification metho. A comparative evaluation base on the well stuie Iowa musical instruments atabase shows results comparable with those of state-of-the-art methos. Thus, the propose metho offers a practical solution for automatic ethnomusicological inexing of a atabase compose of iversifie souns with various quality. The relevance of auio features for the timbre characterization is also iscusse in the context of non-western instruments analysis. 1. INTRODUCTION Characterizing musical timbre perception remains a challenging task relate to the human auitory mechanism an to the physics of musical instruments [4]. This task is full of interest for many applications like automatic atabase inexing, measuring similarities between souns or for automatic soun recognition. Existing psychoacoustical stuies moel the timbre as a multiimensional phenomenon inepenent from musical parameters (e.g. pitch, uration or louness) [7, 8]. A quantitative interpretation of instrument s timbre base on acoustic features compute from auio signals was first propose in [9] an pursue in more recent stuies [12] which aim at organizing auio timbre escriptors efficiently. Nowaays, effective automatic timbre classification methos [13] use supervise statistical learning approaches base on auio signals features compute from analyze ata. Thus, the performance obtaine with such systems epens on the taxonomy, the size an the iversity of training atabases. However, most c Dominique Fourer, Jean-Luc Rouas, Pierre Hanna, Matthias Robine. License uner a Creative Commons Attribution 4.0 International License (CC BY 4.0). Attribution: Dominique Fourer, Jean-Luc Rouas, Pierre Hanna, Matthias Robine. Automatic timbre classification of ethnomusicological auio recorings, 15th International Society for Music Information Retrieval Conference, of existing research atabases (e.g. RWC [6], Iowa [5]) are only compose of common western instruments annotate with specific taxonomies. In this work, we revisit the automatic instrument classification problem from an ethnomusicological point of view by introucing a iversifie an manually annotate research atabase provie by the Centre e Recherche en Ethno-Musicologie (CREM). This atabase is aily supplie by researchers an has the particularity of being compose of uncommon non-western musical instrument recorings from aroun the worl. This work is motivate by practical applications to automatic inexing of online auio recorings atabase which have to be computationally efficient while proviing accurate results. Thus, we aim at valiating the efficiency an the robustness of the statistical learning approach using a constraine stanar taxonomy, applie to recorings of various quality. In this stuy, we expect to show the atabase influence, the relevance of timbre auio features an the choice of taxonomy for the automatic instrument classification process. A result comparison an a crossatabase evaluation is performe using the well-stuie university of Iowa musical instrument atabase. This paper is organize as follows. The CREM atabase is introuce in Section 2. The timbre quantization principle base on mathematical functions escribing auio features is presente in Section 3. An efficient timbre classification metho is escribe in Section 4. Experiments an results base on the propose metho are etaile in Section 5. Conclusion an future works are finally iscusse in Section THE CREM ETHNOMUSICOLOGICAL DATABASE The CREM research atabase 1 is compose of iversifie soun samples irectly recore by ethnomusicologists in various conitions (i.e. no recoring stuio) an from iversifie places all aroun the worl. It contains more than 7000 hours of auio ata recore since 1932 to nowaays using ifferent supports like magnetic tapes or vinyl iscs. The vintage auio recorings of the atabase were carefully igitize to preserve the authenticity of the originals an contain various environment noise. The more recent auio recorings can be irectly igital recore with a high-quality. Most of the musical instruments which com- 1 CREM auio archives freely available online at: 295

2 pose this atabase are non-western an can be uncommon while covering a large range of musical instrument families (see Figure 1(a)). Among uncommon instruments, one can fin the lute or the Ngbaka harp as corophones. More uncommon instruments like Oscillating bamboo, struck machete an struck girer were classifie by ethnomusicologists as iiophones. In this paper, we restricte our stuy to the solo excerpts (where only one monophonic or polyphonic instrument is active) to reuce the interference problems which may occur uring auio analysis. A escription of the selecte CREM sub-atabase is presente in Table 1. Accoring to this table, one can observe that this atabase is actually inhomogeneous. The aerophones are overrepresente while membranophones are unerrepresente. Due to its iversity an the various quality of the composing souns, the automatic ethnomusicological classification of this atabase may appear as challenging. Class name Duration (s) # aerophones-blowe 1, corophones-struck corophones-plucke 715 1, corophones-bowe iiophones-struck iiophones-plucke iiophones-clinke membranophones-struck Total 3, Table 1. Content of the CREM sub-atabase with uration an number of 10-secons segmente excerpts. 3. TIMBRE QUANTIZATION AND CLASSIFICATION 3.1 Timbre quantization Since preliminaries works on the timbre escription of perceive souns, Peeters et al. propose in [12] a large set of auio features escriptors which can be compute from auio signals. The auio escriptors efine numerical functions which aim at proviing cues about specific acoustic features (e.g. brightness is often associate with the spectral centroi accoring to [14]). Thus, the auio escriptors can be organize as follows: Temporal escriptors convey information about the time evolution of a signal (e.g. log attack time, temporal increase, zero-crossing rate, etc.). Harmonic escriptors are compute from the etecte pitch events associate with a funamental frequency (F 0 ). Thus, one can use a prior waveform moel of quasi-harmonic souns which have an equally space Dirac comb shape in the magnitue spectrum. The tonal part of souns can be isolate from signal mixture an be escribe (e.g. noisiness, inharmonicity, etc.). Spectral escriptors are compute from signal timefrequency representation (e.g. Short-Term Fourier Transform) without prior waveform moel (e.g. spectral centroi, spectral ecrease, etc.) Perceptual escriptors are compute from auitoryfiltere banwith versions of signals which aim at approximating the human perception of souns. This can be efficiently compute using Equivalent Rectangular Banwith (ERB) scale [10] which can be combine with gammatone filter-bank [3] (e.g. louness, ERB spectral centroi, etc.) In this stuy, we focus on the soun escriptors liste in table 2 which can be estimate using the timbre toolbox 2 an etaile in [12]. All escriptors are compute for each analyze soun excerpt an may return null values. The harmonic escriptors of polyphonic souns are compute using the prominent etecte F 0 caniate (single F 0 estimation). To normalize the uration of analyze soun, we separate each excerpt in 10-secons length segments without istinction of silence or pitch events. Thus, each segment is represente by a real vector where the corresponing time series of each escriptor is summarize by a statistic. The meian an the Inter Quartile Range (IQR) statistics were chosen for their robustness to outliers. Acronym Descriptor name # Att Attack uration (see ADSR moel [15]) 1 AttSlp Attack slope (ADSR) 1 Dec Decay uration (ADSR) 1 DecSlp Decay slope (ADSR) 1 Rel Release uration (ADSR) 1 LAT Log Attack Time 1 Tcent Temporal centroi 1 Eur Effective uration 1 FreqMo, AmpMo Total energy moulation (frequency,amplitue) 2 RMSenv RMS envelope 2 ACor Signal Auto-Correlation function (12 first coef.) 24 ZCR Zero-Crossing Rate 2 HCent Harmonic spectral centroi 2 HSpr Harmonic spectral sprea 2 HSkew Harmonic skewness 2 HKurt Harmonic kurtosis 2 HSlp Harmonic slope 2 HDec Harmonic ecrease 2 HRoff Harmonic rolloff 2 HVar Harmonic variation 2 HErg, HNErg, HFErg, Harmonic energy, noise energy an frame energy 6 HNois Noisiness 2 HF0 Funamental frequency F 0 2 HinH Inharmonicity 2 HTris Harmonic tristimulus 6 HoevR Harmonic o to even partials ratio 2 Hev Harmonic eviation 2 SCent, ECent Spectral centroi of the magnitue an energy spectrum 4 SSpr, ESpr Spectral sprea of the magnitue an energy spectrum 4 SSkew, ESkew Spectral skewness of the magnitue an energy spectrum 4 SKurt, EKurt Spectral kurtosis of the magnitue an energy spectrum 4 SSlp, ESlp Spectral slope of the magnitue an energy spectrum 4 SDec, EDec Spectral ecrease of the magnitue an energy spectrum 4 SRoff, ERoff Spectral rolloff of the magnitue an energy spectrum 4 SVar, EVar Spectral variation of the magnitue an energy spectrum 4 SFErg, EFErg Spectral frame energy of the magnitue an energy spectrum 4 Sflat, ESflat Spectral flatness of the magnitue an energy spectrum 4 Scre, EScre Spectral crest of the magnitue an energy spectrum 4 ErbCent, ErbGCent ERB scale magnitue spectrogram / gammatone centroi 4 ErbSpr, ErbGSpr ERB scale magnitue spectrogram / gammatone sprea 4 ErbSkew, ErbGSkew ERB scale magnitue spectrogram / gammatone skewness 4 ErbKurt, ErbGKurt ERB scale magnitue spectrogram / gammatone kurtosis 4 ErbSlp, ErbGSlp ERB scale magnitue spectrogram / gammatone slope 4 ErbDec, ErbGDec ERB scale magnitue spectrogram / gammatone ecrease 4 ErbRoff, ErbGRoff ERB scale magnitue spectrogram / gammatone rolloff 4 ErbVar, ErbGVar ERB scale magnitue spectrogram / gammatone variation 4 ErbFErg, ErbGFErg ERB scale magnitue spectrogram / gammatone frame energy 4 ErbSflat, ErbGSflat ERB scale magnitue spectrogram / gammatone flatness 4 ErbScre, ErbGScre ERB scale magnitue spectrogram / gammatone crest 4 Total 164 Table 2. Acronym, name an number of the use timbre escriptors. 2 MATLAB coe available at 296

3 aerophones blowe corophones instrument iiophones bowe plucke struck plucke struck pizzicato (a) Hornbostel an Sachs taxonomy (T1) instrument sustaine clinke membranophones struck strings plucke strings bowe strings flute/rees brass piano violin viola cello oublebass violin viola cello oublebass flute clarinet oboe saxophone bassoon (b) Musician s instrument taxonomy (T2) struck trumpet trombone tuba 4.1 Metho overview Here, each soun segment (cf. Section 3.1) is represente by vector of length p = 164 where each value correspons to a escriptor (see Table 2). The training step of this metho (illustrate in Figure 2) aims at moeling each timbre class using the best projection space for classification. A features selection algorithm is first applie to efficiently reuce the number of escriptors to avoi statistical overlearning. The classification space is compute using iscriminant analysis which consists in estimating optimal weights over the escriptors allowing the best iscrimination between timbre classes. Thus, the classification task consists in projecting an input soun into the best classification space an to select the most probable timbre class using the learne moel. input soun Figure 1. Taxonomies use for the automatic classification of musical instruments as propose by Hornbostel an Sachs taxonomy in [16] (a) an Peeters in [13] (b). 3.2 Classification taxonomy In this stuy, we use two atabases which can be annotate using ifferent taxonomies. Due to its iversity, the CREM atabase was only annotate using the Hornbostel an Sachs taxonomy [16] (T1) illustrate in Figure 1(a) which is wiely use in ethnomusicology. This hierarchical taxonomy is general enough to classify uncommon instruments (e.g. struck bamboo) an conveys information about soun prouction materials an playing styles. From an another han, the Iowa musical instruments atabase [5] use in our experiments was initially annotate using a musician s instrument taxonomy (T2) as propose in [13] an illustrate in Figure 1(b). This atabase is compose of common western pitche instruments which can easily be annotate using T1 as escribe in Table 3. One can notice that the Iowa atabase is only compose of aerophones an corophones instruments. If we consier the playing style, only 4 classes are represente if we apply T1 taxonomy to the Iowa atabase. T1 class name T2 equivalence Duration (s) # aero-blowe ree/flute an brass 5, coro-struck struck strings 5, coro-plucke plucke strings 5, coro-bowe bowe strings 7, Total 24,597 2,735 Table 3. Content of the Iowa atabase using musician s instrument taxonomy (T2) an equivalence with the Hornbostel an Sachs taxonomy (T1). 4. AUTOMATIC INSTRUMENT TIMBRE CLASSIFICATION METHOD The escribe metho aims at estimating the corresponing taxonomy class name of a given input soun. features computation features selection (LDA, MI, IRMFSP) classification space computation (LDA) class moeling class affectation (annotate) Figure 2. Training step of the propose metho. 4.2 Linear iscriminant analysis The goal of Linear Discriminant Analysis (LDA) [1] is to fin the best projection or linear combination of all escriptors which maximizes the average istance between classes (inter-class istance) while minimizing istance between iniviuals from the same class (intra-class istance). This metho assumes that the class affectation of each iniviual is a priori known. Its principle can be escribe as follows. First consier the n p real matrix M where each row is a vector of escriptors associate to a soun (iniviual). We assume that each iniviual is a member of a unique class k [1, K]. Now we efine W as the intraclass variance-covariance matrix which can be estimate by: W = 1 n K n k W k, (1) k=1 where W k is the variance-covariance matrix compute from the n k p sub-matrix of M compose of the n k iniviuals inclue into the class k. We also efine B the inter-class variance-covariance matrix expresse as follows: B = 1 n K n k (µ k µ)(µ k µ) T, (2) k=1 297

4 where µ k correspons to the mean vector of class k an µ is the mean vector of the entire ataset. Accoring to [1], it can be shown that the eigenvectors of matrix D = (B + W ) 1 B solve this optimization problem. When the matrix A = (B + W ) is not invertible, a computational solution consists in using pseuoinverse of matrix A which can be calculate using A T (AA T ) Features selection algorithms Features selection aims at computing the optimal relevance of each escriptor which can be measure with a weight or a rank. The resulting escriptors subset has to be the most iscriminant as possible with the minimal reunancy. In this stuy, we investigate the three approaches escribe below LDA features selection The LDA metho etaile in Section 4.2 can also be use for selecting the most relevant features. In fact, the compute eigenvectors which correspon to linear combination of escriptors convey a relative weight applie to each escriptor. Thus, the significance (or weight) S of a escriptor can be compute using a summation over a efine range [1, R] of the eigenvectors of matrix D as follows: S = R v r,, (3) r=1 where v r, is the -th coefficient of the r-th eigenvector associate to the eigenvalues sorte by escening orer (i.e. r = 1 correspons to the maximal eigenvalue of matrix D). In our implementation, we fixe R = Mutual information Features selection algorithms aim at computing a subset of escriptors that conveys the maximal amount of information to moel classes. From a statistical point of view, if we consier classes an feature escriptors as realizations of ranom variables C an F. The relevance can be measure with the mutual information efine by: I(C, F ) = c P (c, f) P (c, f) P (c)p (f), (4) f where P (c) enotes the probability of C = c which can be estimate from the approximate probability ensity functions (pf) using a compute histogram. Accoring to Bayes theorem one can compute P (c, f) = P (f c)p (c) where P (f c) is the pf of the feature escriptor value f into class c. This metho can be improve using [2] by reucing simultaneously the reunancy by consiering the mutual information between previously selecte escriptors Inertia Ratio Maximisation using features space projection (IRMFSP) This algorithm was first propose in [11] to reuce the number of escriptors use by timbre classification methos. It consists in maximizing the relevance of the escriptors subset for the classification task while minimizing the reunancy between the selecte ones. This iterative metho (ι p) is compose of two steps. The first one selects at iteration ι the non-previously selecte escriptor which maximizes the ratio between inter-class inertia an the total inertia expresse as follow: ˆ (ι) = arg max K n k (µ,k µ )(µ,k µ ) T k=1, (5) n (f (ι),i µ )(f (ι),i µ ) T i=1 where f (ι),i enotes the value of escriptor [1, p] affecte to the iniviual i. µ,k an µ respectively enote the average value of escriptor into the class k an for the total ataset. The secon step of this algorithm aims at orthogonalizing the remaining ata for the next iteration as follows: ( ) f (ι+1) = f (ι) f (ι) g ˆ g ˆ ˆ (ι), (6) where f (ι) is the vector of the previously selecte escriptor ˆ (ι) for all the iniviuals of the entire ataset an ˆ g ˆ = f (ι) (ι) / f is its normalize form. ˆ ˆ 4.4 Class moeling an automatic classification Each instrument class is moele into the projecte classification space resulting from the application of LDA. Thus, each class can be represente by its gravity center ˆµ k which correspons to the vector of the average values of the projecte iniviuals which compose the class k. The classification ecision which affect a class ˆk to an input soun represente by a projecte vector ˆx is simply performe by minimizing the Eucliean istance with the gravity center of each class as follows: ˆk = arg min ˆµ k ˆx 2 k [1, K], (7) k where v 2 enotes the l 2 norm of vector v. Despite its simplicity, this metho seems to obtain goo results comparable with those of the literature [12]. 5. EXPERIMENTS AND RESULTS In this section we present the classification results obtaine using the propose metho escribe in Section Metho evaluation base on self atabase classification In this experiment, we evaluate the classification of each istinct atabase using ifferent taxonomies. We applie the 3-fol cross valiation methoology which consists in partitioning the atabase in 3 istinct ranom subsets compose with 33% of each class (no collision between sets). Thus, the automatic classification applie on each subset is base on training applie on the remaining 66% of the 298

5 atabase. Figure 5.1 compares the classification accuracy obtaine as a function of the number of use escriptors. The resulting confusion matrix of the CREM atabase using 20 auio escriptors is presente in Table 4 an shows an average classification accuracy of 80% where each instrument is well classifie with a minimal accuracy of 70% for the aerophones. These results are goo an seems comparable with those escribe in the literature [11] using the same number of escriptor. The most relevant feature escriptors (selecte among the top ten) estimate by the IRMSFP an use for the classification task are etaile in Table 7. This result reveals significant ifferences between the two atabases. As an example, harmonic escriptors are only iscriminative for the CREM atabase but not for the Iowa atabase. This may be explaine by the presence of membranophone in the CREM atabase which are not present in the Iowa atabase. Contrarily, spectral an perceptual escriptors seems more relevant for the Iowa atabase than for the CREM atabase. Some escriptors appear to be relevant for both atabase like the Spectral flatness (Sflat) an the ERB scale frame energy (ErbFErg) which escribe the spectral envelope of signal. aero c-struc c-pluc c-bowe i-pluc i-struc i-clink membr aero c-struc c-pluc c-bowe i-pluc i-struc i-clink 100 membr Accuracy ratio Accuracy ratio Accuracy as a function of the number of escriptor [17 classes] LDA number of escriptors (a) Iowa atabase using T2 MI IRMFSP Accuracy as a function of the number of escriptor [4 classes] LDA number of escriptors MI IRMFSP Table 4. Confusion matrix (expresse in percent of the souns of the original class liste on the left) of the CREM atabase using the 20 most relevant escriptors selecte by IRMSFP (b) Iowa atabase using T1 Accuracy as a function of the number of escriptor [8 classes] Cross-atabase evaluation In this experiments (see Table 5), we merge the two atabases an we applie the 3-fol cross valiation metho base on the T1 taxonomy to evaluate the classification accuracy on both atabase. The resulting average accuracy is about 68% which is lower than the accuracy obtaine on the istinct classification of each atabase. The results of cross-atabase evaluation applie between atabases using the T1 taxonomy are presente in Table 6 an obtain a poor average accuracy of 30%. This seems to confirm our intuition that the Iowa atabase conveys insufficient information to istinguish the ifferent playing styles between the non-western corophones instruments of the CREM atabase. 6. CONCLUSION AND FUTURE WORKS We applie a computationally efficient automatic timbre classification metho which was successfully evaluate on an introuce iversifie atabase using an ethnomusicological taxonomy. This metho obtains goo classification results (> 80% of accuracy) for both evaluate atabases which are comparable to those of the literature. However, Accuracy ratio LDA number of escriptors (c) CREM atabase using T1 MI IRMFSP Figure 3. Comparison of the 3-fol cross valiation classification accuracy as a function of the number of optimally selecte escriptors. the cross-atabase evaluation shows that each atabase cannot be use to infer a classification to the other. This can be explaine by significant ifferences between these atabases. Interestingly, results on the merge atabase obtain an acceptable accuracy of about 70%. As shown in previous work [11], our experiments confirm the efficiency of IRMFSP algorithm for automatic features selection applie to timbre classification. The interpretation of the 299

6 aero c-struc c-pluc c-bowe i-pluc i-struc i-clink membr aero c-struc c-pluc c-bowe i-pluc i-struc i-clink membr Table 5. Confusion matrix (expresse in percent of the souns of the original class liste on the left) of the evaluate fusion between the CREM an the Iowa atabase using the 20 most relevant escriptors selecte by IRMSFP. aero c-struc c-pluc c-bowe aero c-struc c-pluc c-bowe Table 6. Confusion matrix (expresse in percent of the souns of the original class liste on the left) of the CREM atabase classification base on Iowa atabase training. CREM T1 Iowa T1 Iowa T2 CREM+Iowa T1 Eur AttSlp AttSlp AmpMo Acor Dec Acor Acor ZCR RMSenv Hev Hnois HTris3 Sflat SFErg Sflat Sflat ERoff SRoff SVar SSkew SKurt Scre ErbGKurt ErbKurt ErbSpr ErbFErg ErbFErg ErbFErg ErbRoff ErbRoff ErbSlp ErbGSpr ErbGCent Table 7. Comparison of the most relevant escriptors estimate by IRMFSP. most relevant selecte features shows a significant effect of the content of atabase rather than on the taxonomy. However the timbre moeling interpretation applie to timbre classification remains ifficult. Future works will consist in further investigating the role of escriptors by manually constraining selection before the classification process. 7. ACKNOWLEDGMENTS This research was partly supporte by the French ANR (Agence Nationale e la Recherche) DIADEMS (Description,Inexation, Acces aux Documents Ethnomusicologiques et Sonores) project (ANR-12-CORD-0022). 8. REFERENCES [1] T. W. Anerson. An Introuction to Multivariate Statistical Analysis. Wiley-Blackwell, New York, USA, [2] R. Battiti. Using mutual information for selecting features in supervise neural net learning. IEEE Trans. on Neural Networks, 5(4): , Jul [3] E.Ambikairajah, J. Epps, an L. Lin. Wieban speech an auio coing using gammatone filter banks. In Proc. IEEE ICASSP 01, volume 2, pages , [4] N. F. Fletcher an T. D. Rossing. The Physics of Musical Instruments. Springer-Verlag, [5] L. Fritts. Musical instrument samples. Univ. Iowa Electronic Music Stuios, [Online]. Available: [6] M. Goto, H. Hashiguchi, T. Nishimura, an R. Oka. Rwc music atabase: Music genre atabase an musical instrument soun atabase. In Proc. ISMIR, pages , Oct [7] J. M. Grey an J. W. Goron. Perceptual effects of spctral moifications on musical timbre. Journal of Acoustic Society of America (JASA), 5(63): , [8] S. McAams, S. Winsberg, S. Donnaieu, G. Soete, an J. Krimphoff. Perceptual scaling of synthesize musical timbres: Common imensions, specificities, an latent subject classes. Psychological Research, 58(3): , [9] N. Misariis, K. Bennett, D. Pressnitzer, P. Susini, an S. McAams. Valiation of a multiimensional istance moel for perceptual issimilarities among musical timbres. In Proc. ICA & ASA, volume 103, Seattle, USA, Jun [10] B.C.J. Moore an B.R. Glasberg. Suggeste formulae for calculating auitory-filter banwiths an excitation patterns. Journal of the Acoustical Society of America, 74: , [11] G. Peeters. Automatic classification of large musical instrument atabases using hierarchical classifiers with intertia ratio maximization. In 115th convention of AES, New York, USA, Oct [12] G. Peeters, B. Giorano, P. Susini, N. Misariis, an S. McAams. The timbre toolbox: Auio escriptors of musical signals. Journal of Acoustic Society of America (JASA), 5(130): , Nov [13] G. Peeters an X. Roet. Automatically selecting signal escriptors for soun classification. In Proc. ICMC, Göteborg, Sween, [14] E. Schubert, J. Wolfe, an A. Tarnopolsky. Spectral centroi an timbre in complex, multiple instrumental textures. In Proc. 8th Int. Conf. on Music Perception & Cognition (ICMPC), Evanston, Aug [15] G. Torelli an G. Caironi. New polyphonic soun generator chip with integrate microprocessorprogrammable asr envelope shaper. IEEE Trans. on Consumer Electronics, CE-29(3): , [16] E. v. Hornbostel an C. Sachs. The classification of musical instruments. Galpin Society Journal, 3(25):3 29,

MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES

MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES Jun Wu, Yu Kitano, Stanislaw Andrzej Raczynski, Shigeki Miyabe, Takuya Nishimoto, Nobutaka Ono and Shigeki Sagayama The Graduate

More information

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu

More information

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Róisín Loughran roisin.loughran@ul.ie Jacqueline Walker jacqueline.walker@ul.ie Michael O Neill University

More information

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods Kazuyoshi Yoshii, Masataka Goto and Hiroshi G. Okuno Department of Intelligence Science and Technology National

More information

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 AN HMM BASED INVESTIGATION OF DIFFERENCES BETWEEN MUSICAL INSTRUMENTS OF THE SAME TYPE PACS: 43.75.-z Eichner, Matthias; Wolff, Matthias;

More information

Musical Instrument Identification based on F0-dependent Multivariate Normal Distribution

Musical Instrument Identification based on F0-dependent Multivariate Normal Distribution Musical Instrument Identification based on F0-dependent Multivariate Normal Distribution Tetsuro Kitahara* Masataka Goto** Hiroshi G. Okuno* *Grad. Sch l of Informatics, Kyoto Univ. **PRESTO JST / Nat

More information

Topics in Computer Music Instrument Identification. Ioanna Karydi

Topics in Computer Music Instrument Identification. Ioanna Karydi Topics in Computer Music Instrument Identification Ioanna Karydi Presentation overview What is instrument identification? Sound attributes & Timbre Human performance The ideal algorithm Selected approaches

More information

MUSI-6201 Computational Music Analysis

MUSI-6201 Computational Music Analysis MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)

More information

POLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING

POLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING POLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING Luis Gustavo Martins Telecommunications and Multimedia Unit INESC Porto Porto, Portugal lmartins@inescporto.pt Juan José Burred Communication

More information

GCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam

GCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam GCT535- Sound Technology for Multimedia Timbre Analysis Graduate School of Culture Technology KAIST Juhan Nam 1 Outlines Timbre Analysis Definition of Timbre Timbre Features Zero-crossing rate Spectral

More information

Topic 10. Multi-pitch Analysis

Topic 10. Multi-pitch Analysis Topic 10 Multi-pitch Analysis What is pitch? Common elements of music are pitch, rhythm, dynamics, and the sonic qualities of timbre and texture. An auditory perceptual attribute in terms of which sounds

More information

MUSICAL NOTE AND INSTRUMENT CLASSIFICATION WITH LIKELIHOOD-FREQUENCY-TIME ANALYSIS AND SUPPORT VECTOR MACHINES

MUSICAL NOTE AND INSTRUMENT CLASSIFICATION WITH LIKELIHOOD-FREQUENCY-TIME ANALYSIS AND SUPPORT VECTOR MACHINES MUSICAL NOTE AND INSTRUMENT CLASSIFICATION WITH LIKELIHOOD-FREQUENCY-TIME ANALYSIS AND SUPPORT VECTOR MACHINES Mehmet Erdal Özbek 1, Claude Delpha 2, and Pierre Duhamel 2 1 Dept. of Electrical and Electronics

More information

Musical instrument identification in continuous recordings

Musical instrument identification in continuous recordings Musical instrument identification in continuous recordings Arie Livshin, Xavier Rodet To cite this version: Arie Livshin, Xavier Rodet. Musical instrument identification in continuous recordings. Digital

More information

WE ADDRESS the development of a novel computational

WE ADDRESS the development of a novel computational IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 3, MARCH 2010 663 Dynamic Spectral Envelope Modeling for Timbre Analysis of Musical Instrument Sounds Juan José Burred, Member,

More information

A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES

A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES Panayiotis Kokoras School of Music Studies Aristotle University of Thessaloniki email@panayiotiskokoras.com Abstract. This article proposes a theoretical

More information

AMusical Instrument Sample Database of Isolated Notes

AMusical Instrument Sample Database of Isolated Notes 1046 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 5, JULY 2009 Purging Musical Instrument Sample Databases Using Automatic Musical Instrument Recognition Methods Arie Livshin

More information

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS Andrew N. Robertson, Mark D. Plumbley Centre for Digital Music

More information

Application Of Missing Feature Theory To The Recognition Of Musical Instruments In Polyphonic Audio

Application Of Missing Feature Theory To The Recognition Of Musical Instruments In Polyphonic Audio Application Of Missing Feature Theory To The Recognition Of Musical Instruments In Polyphonic Audio Jana Eggink and Guy J. Brown Department of Computer Science, University of Sheffield Regent Court, 11

More information

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors Priyanka S. Jadhav M.E. (Computer Engineering) G. H. Raisoni College of Engg. & Mgmt. Wagholi, Pune, India E-mail:

More information

THE importance of music content analysis for musical

THE importance of music content analysis for musical IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 15, NO. 1, JANUARY 2007 333 Drum Sound Recognition for Polyphonic Audio Signals by Adaptation and Matching of Spectrogram Templates With

More information

Semi-supervised Musical Instrument Recognition

Semi-supervised Musical Instrument Recognition Semi-supervised Musical Instrument Recognition Master s Thesis Presentation Aleksandr Diment 1 1 Tampere niversity of Technology, Finland Supervisors: Adj.Prof. Tuomas Virtanen, MSc Toni Heittola 17 May

More information

MOTIVATION AGENDA MUSIC, EMOTION, AND TIMBRE CHARACTERIZING THE EMOTION OF INDIVIDUAL PIANO AND OTHER MUSICAL INSTRUMENT SOUNDS

MOTIVATION AGENDA MUSIC, EMOTION, AND TIMBRE CHARACTERIZING THE EMOTION OF INDIVIDUAL PIANO AND OTHER MUSICAL INSTRUMENT SOUNDS MOTIVATION Thank you YouTube! Why do composers spend tremendous effort for the right combination of musical instruments? CHARACTERIZING THE EMOTION OF INDIVIDUAL PIANO AND OTHER MUSICAL INSTRUMENT SOUNDS

More information

Instrument identification in solo and ensemble music using independent subspace analysis

Instrument identification in solo and ensemble music using independent subspace analysis Instrument identification in solo and ensemble music using independent subspace analysis Emmanuel Vincent, Xavier Rodet To cite this version: Emmanuel Vincent, Xavier Rodet. Instrument identification in

More information

Supervised Musical Source Separation from Mono and Stereo Mixtures based on Sinusoidal Modeling

Supervised Musical Source Separation from Mono and Stereo Mixtures based on Sinusoidal Modeling Supervised Musical Source Separation from Mono and Stereo Mixtures based on Sinusoidal Modeling Juan José Burred Équipe Analyse/Synthèse, IRCAM burred@ircam.fr Communication Systems Group Technische Universität

More information

Experiments on musical instrument separation using multiplecause

Experiments on musical instrument separation using multiplecause Experiments on musical instrument separation using multiplecause models J Klingseisen and M D Plumbley* Department of Electronic Engineering King's College London * - Corresponding Author - mark.plumbley@kcl.ac.uk

More information

Singing Voice Conversion Using Posted Waveform Data on Music Social Media

Singing Voice Conversion Using Posted Waveform Data on Music Social Media Singing Voice Conversion Using Poste Waveform Data on Music Social Meia Koki Sena, Yukiya Hono, Kei Sawaa, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku an Keiichi Tokua Department of Computer Science

More information

Week 14 Music Understanding and Classification

Week 14 Music Understanding and Classification Week 14 Music Understanding and Classification Roger B. Dannenberg Professor of Computer Science, Music & Art Overview n Music Style Classification n What s a classifier? n Naïve Bayesian Classifiers n

More information

Recognising Cello Performers using Timbre Models

Recognising Cello Performers using Timbre Models Recognising Cello Performers using Timbre Models Chudy, Magdalena; Dixon, Simon For additional information about this publication click this link. http://qmro.qmul.ac.uk/jspui/handle/123456789/5013 Information

More information

Automatic music transcription

Automatic music transcription Music transcription 1 Music transcription 2 Automatic music transcription Sources: * Klapuri, Introduction to music transcription, 2006. www.cs.tut.fi/sgn/arg/klap/amt-intro.pdf * Klapuri, Eronen, Astola:

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM A QUER B EAMPLE MUSIC RETRIEVAL ALGORITHM H. HARB AND L. CHEN Maths-Info department, Ecole Centrale de Lyon. 36, av. Guy de Collongue, 69134, Ecully, France, EUROPE E-mail: {hadi.harb, liming.chen}@ec-lyon.fr

More information

CS229 Project Report Polyphonic Piano Transcription

CS229 Project Report Polyphonic Piano Transcription CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project

More information

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION Halfdan Rump, Shigeki Miyabe, Emiru Tsunoo, Nobukata Ono, Shigeki Sagama The University of Tokyo, Graduate

More information

Lecture 9 Source Separation

Lecture 9 Source Separation 10420CS 573100 音樂資訊檢索 Music Information Retrieval Lecture 9 Source Separation Yi-Hsuan Yang Ph.D. http://www.citi.sinica.edu.tw/pages/yang/ yang@citi.sinica.edu.tw Music & Audio Computing Lab, Research

More information

Automatic Construction of Synthetic Musical Instruments and Performers

Automatic Construction of Synthetic Musical Instruments and Performers Ph.D. Thesis Proposal Automatic Construction of Synthetic Musical Instruments and Performers Ning Hu Carnegie Mellon University Thesis Committee Roger B. Dannenberg, Chair Michael S. Lewicki Richard M.

More information

Multipitch estimation by joint modeling of harmonic and transient sounds

Multipitch estimation by joint modeling of harmonic and transient sounds Multipitch estimation by joint modeling of harmonic and transient sounds Jun Wu, Emmanuel Vincent, Stanislaw Raczynski, Takuya Nishimoto, Nobutaka Ono, Shigeki Sagayama To cite this version: Jun Wu, Emmanuel

More information

Transcription of the Singing Melody in Polyphonic Music

Transcription of the Singing Melody in Polyphonic Music Transcription of the Singing Melody in Polyphonic Music Matti Ryynänen and Anssi Klapuri Institute of Signal Processing, Tampere University Of Technology P.O.Box 553, FI-33101 Tampere, Finland {matti.ryynanen,

More information

Chord Classification of an Audio Signal using Artificial Neural Network

Chord Classification of an Audio Signal using Artificial Neural Network Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------

More information

Recognising Cello Performers Using Timbre Models

Recognising Cello Performers Using Timbre Models Recognising Cello Performers Using Timbre Models Magdalena Chudy and Simon Dixon Abstract In this paper, we compare timbre features of various cello performers playing the same instrument in solo cello

More information

Subjective Similarity of Music: Data Collection for Individuality Analysis

Subjective Similarity of Music: Data Collection for Individuality Analysis Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp

More information

Research Article. ISSN (Print) *Corresponding author Shireen Fathima

Research Article. ISSN (Print) *Corresponding author Shireen Fathima Scholars Journal of Engineering and Technology (SJET) Sch. J. Eng. Tech., 2014; 2(4C):613-620 Scholars Academic and Scientific Publisher (An International Publisher for Academic and Scientific Resources)

More information

Automatic Identification of Instrument Type in Music Signal using Wavelet and MFCC

Automatic Identification of Instrument Type in Music Signal using Wavelet and MFCC Automatic Identification of Instrument Type in Music Signal using Wavelet and MFCC Arijit Ghosal, Rudrasis Chakraborty, Bibhas Chandra Dhara +, and Sanjoy Kumar Saha! * CSE Dept., Institute of Technology

More information

Keywords Separation of sound, percussive instruments, non-percussive instruments, flexible audio source separation toolbox

Keywords Separation of sound, percussive instruments, non-percussive instruments, flexible audio source separation toolbox Volume 4, Issue 4, April 2014 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Investigation

More information

Music Information Retrieval with Temporal Features and Timbre

Music Information Retrieval with Temporal Features and Timbre Music Information Retrieval with Temporal Features and Timbre Angelina A. Tzacheva and Keith J. Bell University of South Carolina Upstate, Department of Informatics 800 University Way, Spartanburg, SC

More information

Environmental sound description : comparison and generalization of 4 timbre studies

Environmental sound description : comparison and generalization of 4 timbre studies Environmental sound description : comparison and generaliation of 4 timbre studies A. Minard, P. Susini, N. Misdariis, G. Lemaitre STMS-IRCAM-CNRS 1 place Igor Stravinsky, 75004 Paris, France. antoine.minard@ircam.fr

More information

MUSICAL INSTRUMENT RECOGNITION USING BIOLOGICALLY INSPIRED FILTERING OF TEMPORAL DICTIONARY ATOMS

MUSICAL INSTRUMENT RECOGNITION USING BIOLOGICALLY INSPIRED FILTERING OF TEMPORAL DICTIONARY ATOMS MUSICAL INSTRUMENT RECOGNITION USING BIOLOGICALLY INSPIRED FILTERING OF TEMPORAL DICTIONARY ATOMS Steven K. Tjoa and K. J. Ray Liu Signals and Information Group, Department of Electrical and Computer Engineering

More information

LOUDNESS EFFECT OF THE DIFFERENT TONES ON THE TIMBRE SUBJECTIVE PERCEPTION EXPERIMENT OF ERHU

LOUDNESS EFFECT OF THE DIFFERENT TONES ON THE TIMBRE SUBJECTIVE PERCEPTION EXPERIMENT OF ERHU The 21 st International Congress on Sound and Vibration 13-17 July, 2014, Beijing/China LOUDNESS EFFECT OF THE DIFFERENT TONES ON THE TIMBRE SUBJECTIVE PERCEPTION EXPERIMENT OF ERHU Siyu Zhu, Peifeng Ji,

More information

Music Complexity Descriptors. Matt Stabile June 6 th, 2008

Music Complexity Descriptors. Matt Stabile June 6 th, 2008 Music Complexity Descriptors Matt Stabile June 6 th, 2008 Musical Complexity as a Semantic Descriptor Modern digital audio collections need new criteria for categorization and searching. Applicable to:

More information

2018 Fall CTP431: Music and Audio Computing Fundamentals of Musical Acoustics

2018 Fall CTP431: Music and Audio Computing Fundamentals of Musical Acoustics 2018 Fall CTP431: Music and Audio Computing Fundamentals of Musical Acoustics Graduate School of Culture Technology, KAIST Juhan Nam Outlines Introduction to musical tones Musical tone generation - String

More information

A prototype system for rule-based expressive modifications of audio recordings

A prototype system for rule-based expressive modifications of audio recordings International Symposium on Performance Science ISBN 0-00-000000-0 / 000-0-00-000000-0 The Author 2007, Published by the AEC All rights reserved A prototype system for rule-based expressive modifications

More information

EE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function

EE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function EE391 Special Report (Spring 25) Automatic Chord Recognition Using A Summary Autocorrelation Function Advisor: Professor Julius Smith Kyogu Lee Center for Computer Research in Music and Acoustics (CCRMA)

More information

Feature-based Characterization of Violin Timbre

Feature-based Characterization of Violin Timbre 7 th European Signal Processing Conference (EUSIPCO) Feature-based Characterization of Violin Timbre Francesco Setragno, Massimiliano Zanoni, Augusto Sarti and Fabio Antonacci Dipartimento di Elettronica,

More information

Interactive Classification of Sound Objects for Polyphonic Electro-Acoustic Music Annotation

Interactive Classification of Sound Objects for Polyphonic Electro-Acoustic Music Annotation for Polyphonic Electro-Acoustic Music Annotation Sebastien Gulluni 2, Slim Essid 2, Olivier Buisson, and Gaël Richard 2 Institut National de l Audiovisuel, 4 avenue de l Europe 94366 Bry-sur-marne Cedex,

More information

Violin Timbre Space Features

Violin Timbre Space Features Violin Timbre Space Features J. A. Charles φ, D. Fitzgerald*, E. Coyle φ φ School of Control Systems and Electrical Engineering, Dublin Institute of Technology, IRELAND E-mail: φ jane.charles@dit.ie Eugene.Coyle@dit.ie

More information

Classification of Timbre Similarity

Classification of Timbre Similarity Classification of Timbre Similarity Corey Kereliuk McGill University March 15, 2007 1 / 16 1 Definition of Timbre What Timbre is Not What Timbre is A 2-dimensional Timbre Space 2 3 Considerations Common

More information

Automatic Rhythmic Notation from Single Voice Audio Sources

Automatic Rhythmic Notation from Single Voice Audio Sources Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung

More information

GOOD-SOUNDS.ORG: A FRAMEWORK TO EXPLORE GOODNESS IN INSTRUMENTAL SOUNDS

GOOD-SOUNDS.ORG: A FRAMEWORK TO EXPLORE GOODNESS IN INSTRUMENTAL SOUNDS GOOD-SOUNDS.ORG: A FRAMEWORK TO EXPLORE GOODNESS IN INSTRUMENTAL SOUNDS Giuseppe Bandiera 1 Oriol Romani Picas 1 Hiroshi Tokuda 2 Wataru Hariya 2 Koji Oishi 2 Xavier Serra 1 1 Music Technology Group, Universitat

More information

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng S. Zhu, P. Ji, W. Kuang and J. Yang Institute of Acoustics, CAS, O.21, Bei-Si-huan-Xi Road, 100190 Beijing,

More information

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION ULAŞ BAĞCI AND ENGIN ERZIN arxiv:0907.3220v1 [cs.sd] 18 Jul 2009 ABSTRACT. Music genre classification is an essential tool for

More information

Topic 4. Single Pitch Detection

Topic 4. Single Pitch Detection Topic 4 Single Pitch Detection What is pitch? A perceptual attribute, so subjective Only defined for (quasi) harmonic sounds Harmonic sounds are periodic, and the period is 1/F0. Can be reliably matched

More information

TYING SEMANTIC LABELS TO COMPUTATIONAL DESCRIPTORS OF SIMILAR TIMBRES

TYING SEMANTIC LABELS TO COMPUTATIONAL DESCRIPTORS OF SIMILAR TIMBRES TYING SEMANTIC LABELS TO COMPUTATIONAL DESCRIPTORS OF SIMILAR TIMBRES Rosemary A. Fitzgerald Department of Music Lancaster University, Lancaster, LA1 4YW, UK r.a.fitzgerald@lancaster.ac.uk ABSTRACT This

More information

Convention Paper Presented at the 115th Convention 2003 October New York, NY, USA

Convention Paper Presented at the 115th Convention 2003 October New York, NY, USA Audio Engineering Society Convention Paper Presented at the 5th Convention 23 October 3 New York, NY, USA This convention paper has been reproduced from the author s advance manuscript, without editing,

More information

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES Vishweshwara Rao and Preeti Rao Digital Audio Processing Lab, Electrical Engineering Department, IIT-Bombay, Powai,

More information

HIT SONG SCIENCE IS NOT YET A SCIENCE

HIT SONG SCIENCE IS NOT YET A SCIENCE HIT SONG SCIENCE IS NOT YET A SCIENCE François Pachet Sony CSL pachet@csl.sony.fr Pierre Roy Sony CSL roy@csl.sony.fr ABSTRACT We describe a large-scale experiment aiming at validating the hypothesis that

More information

Subjective evaluation of common singing skills using the rank ordering method

Subjective evaluation of common singing skills using the rank ordering method lma Mater Studiorum University of ologna, ugust 22-26 2006 Subjective evaluation of common singing skills using the rank ordering method Tomoyasu Nakano Graduate School of Library, Information and Media

More information

Neural Network for Music Instrument Identi cation

Neural Network for Music Instrument Identi cation Neural Network for Music Instrument Identi cation Zhiwen Zhang(MSE), Hanze Tu(CCRMA), Yuan Li(CCRMA) SUN ID: zhiwen, hanze, yuanli92 Abstract - In the context of music, instrument identi cation would contribute

More information

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST) Computational Models of Music Similarity 1 Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST) Abstract The perceived similarity of two pieces of music is multi-dimensional,

More information

An Accurate Timbre Model for Musical Instruments and its Application to Classification

An Accurate Timbre Model for Musical Instruments and its Application to Classification An Accurate Timbre Model for Musical Instruments and its Application to Classification Juan José Burred 1,AxelRöbel 2, and Xavier Rodet 2 1 Communication Systems Group, Technical University of Berlin,

More information

Parameter Estimation of Virtual Musical Instrument Synthesizers

Parameter Estimation of Virtual Musical Instrument Synthesizers Parameter Estimation of Virtual Musical Instrument Synthesizers Katsutoshi Itoyama Kyoto University itoyama@kuis.kyoto-u.ac.jp Hiroshi G. Okuno Kyoto University okuno@kuis.kyoto-u.ac.jp ABSTRACT A method

More information

A SEGMENTAL SPECTRO-TEMPORAL MODEL OF MUSICAL TIMBRE

A SEGMENTAL SPECTRO-TEMPORAL MODEL OF MUSICAL TIMBRE A SEGMENTAL SPECTRO-TEMPORAL MODEL OF MUSICAL TIMBRE Juan José Burred, Axel Röbel Analysis/Synthesis Team, IRCAM Paris, France {burred,roebel}@ircam.fr ABSTRACT We propose a new statistical model of musical

More information

Robert Alexandru Dobre, Cristian Negrescu

Robert Alexandru Dobre, Cristian Negrescu ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Automatic Music Transcription Software Based on Constant Q

More information

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES PACS: 43.60.Lq Hacihabiboglu, Huseyin 1,2 ; Canagarajah C. Nishan 2 1 Sonic Arts Research Centre (SARC) School of Computer Science Queen s University

More information

CTP431- Music and Audio Computing Musical Acoustics. Graduate School of Culture Technology KAIST Juhan Nam

CTP431- Music and Audio Computing Musical Acoustics. Graduate School of Culture Technology KAIST Juhan Nam CTP431- Music and Audio Computing Musical Acoustics Graduate School of Culture Technology KAIST Juhan Nam 1 Outlines What is sound? Physical view Psychoacoustic view Sound generation Wave equation Wave

More information

A NOVEL CEPSTRAL REPRESENTATION FOR TIMBRE MODELING OF SOUND SOURCES IN POLYPHONIC MIXTURES

A NOVEL CEPSTRAL REPRESENTATION FOR TIMBRE MODELING OF SOUND SOURCES IN POLYPHONIC MIXTURES A NOVEL CEPSTRAL REPRESENTATION FOR TIMBRE MODELING OF SOUND SOURCES IN POLYPHONIC MIXTURES Zhiyao Duan 1, Bryan Pardo 2, Laurent Daudet 3 1 Department of Electrical and Computer Engineering, University

More information

EXPLORING THE USE OF ENF FOR MULTIMEDIA SYNCHRONIZATION

EXPLORING THE USE OF ENF FOR MULTIMEDIA SYNCHRONIZATION EXPLORING THE USE OF ENF FOR MULTIMEDIA SYNCHRONIZATION Hui Su, Adi Hajj-Ahmad, Min Wu, and Douglas W. Oard {hsu, adiha, minwu, oard}@umd.edu University of Maryland, College Park ABSTRACT The electric

More information

Topic 11. Score-Informed Source Separation. (chroma slides adapted from Meinard Mueller)

Topic 11. Score-Informed Source Separation. (chroma slides adapted from Meinard Mueller) Topic 11 Score-Informed Source Separation (chroma slides adapted from Meinard Mueller) Why Score-informed Source Separation? Audio source separation is useful Music transcription, remixing, search Non-satisfying

More information

Proceedings of Meetings on Acoustics

Proceedings of Meetings on Acoustics Proceedings of Meetings on Acoustics Volume 19, 2013 http://acousticalsociety.org/ ICA 2013 Montreal Montreal, Canada 2-7 June 2013 Musical Acoustics Session 3pMU: Perception and Orchestration Practice

More information

Supervised Learning in Genre Classification

Supervised Learning in Genre Classification Supervised Learning in Genre Classification Introduction & Motivation Mohit Rajani and Luke Ekkizogloy {i.mohit,luke.ekkizogloy}@gmail.com Stanford University, CS229: Machine Learning, 2009 Now that music

More information

CSC475 Music Information Retrieval

CSC475 Music Information Retrieval CSC475 Music Information Retrieval Monophonic pitch extraction George Tzanetakis University of Victoria 2014 G. Tzanetakis 1 / 32 Table of Contents I 1 Motivation and Terminology 2 Psychacoustics 3 F0

More information

MPEG-7 AUDIO SPECTRUM BASIS AS A SIGNATURE OF VIOLIN SOUND

MPEG-7 AUDIO SPECTRUM BASIS AS A SIGNATURE OF VIOLIN SOUND MPEG-7 AUDIO SPECTRUM BASIS AS A SIGNATURE OF VIOLIN SOUND Aleksander Kaminiarz, Ewa Łukasik Institute of Computing Science, Poznań University of Technology. Piotrowo 2, 60-965 Poznań, Poland e-mail: Ewa.Lukasik@cs.put.poznan.pl

More information

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

Music Emotion Recognition. Jaesung Lee. Chung-Ang University Music Emotion Recognition Jaesung Lee Chung-Ang University Introduction Searching Music in Music Information Retrieval Some information about target music is available Query by Text: Title, Artist, or

More information

TIMBRE-CONSTRAINED RECURSIVE TIME-VARYING ANALYSIS FOR MUSICAL NOTE SEPARATION

TIMBRE-CONSTRAINED RECURSIVE TIME-VARYING ANALYSIS FOR MUSICAL NOTE SEPARATION IMBRE-CONSRAINED RECURSIVE IME-VARYING ANALYSIS FOR MUSICAL NOE SEPARAION Yu Lin, Wei-Chen Chang, ien-ming Wang, Alvin W.Y. Su, SCREAM Lab., Department of CSIE, National Cheng-Kung University, ainan, aiwan

More information

Audio-Based Video Editing with Two-Channel Microphone

Audio-Based Video Editing with Two-Channel Microphone Audio-Based Video Editing with Two-Channel Microphone Tetsuya Takiguchi Organization of Advanced Science and Technology Kobe University, Japan takigu@kobe-u.ac.jp Yasuo Ariki Organization of Advanced Science

More information

HUMANS have a remarkable ability to recognize objects

HUMANS have a remarkable ability to recognize objects IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 21, NO. 9, SEPTEMBER 2013 1805 Musical Instrument Recognition in Polyphonic Audio Using Missing Feature Approach Dimitrios Giannoulis,

More information

Analysis, Synthesis, and Perception of Musical Sounds

Analysis, Synthesis, and Perception of Musical Sounds Analysis, Synthesis, and Perception of Musical Sounds The Sound of Music James W. Beauchamp Editor University of Illinois at Urbana, USA 4y Springer Contents Preface Acknowledgments vii xv 1. Analysis

More information

SYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS

SYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS Published by Institute of Electrical Engineers (IEE). 1998 IEE, Paul Masri, Nishan Canagarajah Colloquium on "Audio and Music Technology"; November 1998, London. Digest No. 98/470 SYNTHESIS FROM MUSICAL

More information

A Survey on: Sound Source Separation Methods

A Survey on: Sound Source Separation Methods Volume 3, Issue 11, November-2016, pp. 580-584 ISSN (O): 2349-7084 International Journal of Computer Engineering In Research Trends Available online at: www.ijcert.org A Survey on: Sound Source Separation

More information

A DISPLAY INDEPENDENT HIGH DYNAMIC RANGE TELEVISION SYSTEM

A DISPLAY INDEPENDENT HIGH DYNAMIC RANGE TELEVISION SYSTEM A DISPLAY INDEPENDENT HIGH DYNAMIC RANGE TELEVISION SYSTEM T. Borer an A. Cotton BBC R&D, 56 Woo Lane, Lonon, W12 7SB, UK ABSTRACT High Dynamic Range (HDR) television has capture the imagination of the

More information

Automatic morphological description of sounds

Automatic morphological description of sounds Automatic morphological description of sounds G. G. F. Peeters and E. Deruty Ircam, 1, pl. Igor Stravinsky, 75004 Paris, France peeters@ircam.fr 5783 Morphological description of sound has been proposed

More information

A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION

A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION Graham E. Poliner and Daniel P.W. Ellis LabROSA, Dept. of Electrical Engineering Columbia University, New York NY 127 USA {graham,dpwe}@ee.columbia.edu

More information

Cross-Dataset Validation of Feature Sets in Musical Instrument Classification

Cross-Dataset Validation of Feature Sets in Musical Instrument Classification Cross-Dataset Validation of Feature Sets in Musical Instrument Classification Patrick J. Donnelly and John W. Sheppard Department of Computer Science Montana State University Bozeman, MT 59715 {patrick.donnelly2,

More information

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC Scientific Journal of Impact Factor (SJIF): 5.71 International Journal of Advance Engineering and Research Development Volume 5, Issue 04, April -2018 e-issn (O): 2348-4470 p-issn (P): 2348-6406 MUSICAL

More information

Music Synchronization. Music Synchronization. Music Data. Music Data. General Goals. Music Information Retrieval (MIR)

Music Synchronization. Music Synchronization. Music Data. Music Data. General Goals. Music Information Retrieval (MIR) Advanced Course Computer Science Music Processing Summer Term 2010 Music ata Meinard Müller Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Music Synchronization Music ata Various interpretations

More information

Psychophysical quantification of individual differences in timbre perception

Psychophysical quantification of individual differences in timbre perception Psychophysical quantification of individual differences in timbre perception Stephen McAdams & Suzanne Winsberg IRCAM-CNRS place Igor Stravinsky F-75004 Paris smc@ircam.fr SUMMARY New multidimensional

More information

Music Source Separation

Music Source Separation Music Source Separation Hao-Wei Tseng Electrical and Engineering System University of Michigan Ann Arbor, Michigan Email: blakesen@umich.edu Abstract In popular music, a cover version or cover song, or

More information

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC Vishweshwara Rao, Sachin Pant, Madhumita Bhaskar and Preeti Rao Department of Electrical Engineering, IIT Bombay {vishu, sachinp,

More information

DXR.1 Digital Audio Codec

DXR.1 Digital Audio Codec DXR.1 Digital Auio Coec SECTION 1...INTRODUCTION... 3...DIGITAL SERVICES... 3...WHAT COMES WITH THE DXR.1?... 3 2...SETUP... 4...DATA CONNECTION... 4...POWER CONNECTION... 4...AUDIO CONNECTIONS... 5...CONTACT

More information

Musical Instrument Recognizer Instrogram and Its Application to Music Retrieval based on Instrumentation Similarity

Musical Instrument Recognizer Instrogram and Its Application to Music Retrieval based on Instrumentation Similarity Musical Instrument Recognizer Instrogram and Its Application to Music Retrieval based on Instrumentation Similarity Tetsuro Kitahara, Masataka Goto, Kazunori Komatani, Tetsuya Ogata and Hiroshi G. Okuno

More information

Audio classification from time-frequency texture

Audio classification from time-frequency texture Audio classification from time-frequency texture The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published Publisher Guoshen,

More information

PREDICTING THE PERCEIVED SPACIOUSNESS OF STEREOPHONIC MUSIC RECORDINGS

PREDICTING THE PERCEIVED SPACIOUSNESS OF STEREOPHONIC MUSIC RECORDINGS PREDICTING THE PERCEIVED SPACIOUSNESS OF STEREOPHONIC MUSIC RECORDINGS Andy M. Sarroff and Juan P. Bello New York University andy.sarroff@nyu.edu ABSTRACT In a stereophonic music production, music producers

More information