Research Projects. Measuring music similarity and recommending music. Douglas Eck Research Statement 2

Size: px
Start display at page:

Download "Research Projects. Measuring music similarity and recommending music. Douglas Eck Research Statement 2"

Transcription

1 Research Statement Douglas Eck Assistant Professor University of Montreal Department of Computer Science Montreal, QC, Canada Overview and Background Since 2003 I have been an assistant professor in the University of Montreal Department of Computer Science. My training is in machine learning and cognitive science (Ph.D. in Computer Science & Cognitive Science, Indiana University, 2000). I design machine learning algorithms in the domain music. I have also done research in computational models of musical rhythm and meter. Primarily I am interested in bringing knowledge about human music perception and production as well as knowledge about music structure to bear on the task of computational modeling and algorithm design in the domain of music. Representative tasks include large-scale music recommendation, low-dimensional representations of music audio for similarity, beat tracking and synchronization with music audio, expressive music performance and music generation. My research is multidisciplinary. In addition to publishing in my primary domain of machine learning (NIPS, Machine Learning, Neural Computation, Neural Networks), I have also published in psychology and cognitive science (Music Perception, Psychology Research), signal processing (IEEE Sig Proc, EURASIP Journal Appl. Sig. Proc.), music information retrieval (ISMIR), and music technology (Journal of New Music Research). I have a long history of collaboration. I have active collaborations with fellow machine learning faculty at the University of Montreal (Yoshua Bengio, Balázs Kégl, Pascal Vincent). I work closely with psychologists and neuroscientists from the Brain Music and Sound (BRAMS) research center. This includes projects between myself and my students and the labs of Caroline Palmer (McGill psychology), Isabelle Peretz (University of Montreal psychology) and Robert Zatorre (Montreal Neurological Institute). Via McGill s Center for Interdisciplinary Research in Music Media and Tech (CIRMMT), I collaborate with the music and information retrieval group of Ichiro Fujinaga. I work closely with students and enjoy my role as lab leader. I currently advise or co-advise seven graduate students. Most of my publications are co-authored by one of my students. I offer my students a wide range of research options and try to connect them with other researchers in the field doing similar work. To illustrate, I currently have students working on projects such as sequence generation algorithms trained on music scores (symbolic sequence learning), expressive piano performance (cognitive psychology, motion capture), audio chord recognition (audio signal processing, matrix factorization), music similarity and recommendation (dimensionality reduction, classification).

2 Douglas Eck Research Statement 2 I also enjoy working with industry partners. Recently I participated in two technology transfer projects related to music recommendation. In the first project Dan Levitin (McGill psychology), Yoshua Bengio (University of Montreal Computer Science) and myself developed the core technologies for a Montreal music recommendation site In the second project I worked on similar but larger-scale music recommendation project with Sun Labs, Boston (described below). The value of industry collaboration in the domain of machine learning is self-evident: one is forced to design algorithms that actually perform well on real-world problems. Additionally such contacts are useful for finding jobs for students who do not want to continue in academia. For example my Master s student Norman Casagrande is now a machine learning researcher at the popular web music recommendation site last.fm in London. In the future I am interested in expanding ties with video game developers (context-aware music generation) and developers of online communities such as Second Life (immersive musical environments). Research Projects Measuring music similarity and recommending music Representative publications: D. Eck, P. Lamere, T. Bertin-Mahieux, and S. Green. Automatic generation of social tags for music recommendation. In Neural Information Processing Systems Conference (NIPS) 20, J. Bergstra, N. Casagrande, D. Erhan, D. Eck, and B. Kégl. Aggregate features and AdaBoost for music classification. Machine Learning, 65(2-3): , Presume that you want to measure how similar are two songs based solely on their acoustic content. If you could do this in a sophisticated manner it would be possible to recommend new music to listeners based on how their favorite songs sound. One place to start is to represent each song as a vector such that similar songs are near-neighbors in this vector space. Similarity can then be measured using a vector distance measure such as cosine distance. Unfortunately it is difficult to construct a vectorial representation that is compact enough to yield meaningful vector distances and yet still descriptive of the music contained in the song. The root of the challenge lies in the high dimensionality of music. Nearly 1 million integer values are required to represent 10 seconds of CD-quality stereo audio (44, bit samples per second second for each stereo channel = 882, 000) yielding almost 16 million values for a three minute song. It is of course possible to downsample and otherwise transform audio in order to reduce this dimensionality. For example, in speech recognition, Fourier-based transforms are commonly used to generate representations containing on the order of 200 values for every second of audio (e.g. 20 Mel-Frequency Cepstral Coefficients or MFCCs calculated at 100ms intervals). But even then our three minute song is encoded using 36, 000 values, too many for effective vector-based similarity in this domain. In addition to this challenge of high-dimensional inputs, there is also the problem of how people judge music similarity. First, we attend to different timescales when measuring similarity. For example we might cluster music based on rhythm and tempo cues (e.g. I want music to jog to ) encoded in timescales

3 Douglas Eck Research Statement 3 between roughly.5hz and 5Hz. Or we might instead pa attention to timbral cues (e.g. I want to hear accordion music ) encoded in much faster timescales. Second, our criteria for judging song similarity change based on context. For example one might add the Beatles song Taxman to a jogging mix because it has an upbeat tempo. In another context one might select the same song for different reasons. Perhaps the album Revolver is put into a dinner party mix because it is good background for afterdinner conversation. For these reasons it seems clear that standard signal processing tricks are not enough to yield a good measure of music similarity. I have worked on several projects which address issues related to music recommendation and music similarity. With Yoshua Bengio (Machine Learning, University of Montreal) and Dan Levitin (Music Psychology, McGill) I worked on a research project funded by a Canada NSERC I2I Innov industrial partnership grant. The goal of the project was to build a hybrid collaborative filtering / content-based filtering music recommender. Though our technology performed relatively well, the project was not pursued by the industrial partner past our one year involvement. With students James Bergstra and Norman Casagrande, I competed in two contests at MIREX, an international programming contest for music information retrieval (MIR). The contests involved predicting, respectively, the genre and the artist of audio waveforms. Our approach used a multi-class version of the ensemble learning algorithm AdaBoost. The model automatically ranked the audio features, allowing us to work with a large feature set and to iteratively discard unpredictive features when the model did not select them. The model finished first among 15 entries for genre recognition [2] and second among 10 entries for artist recognition. See [3] for more details. In January 2007 I took a six-month leave from the University of Montreal and became a visiting professor at Sun Labs, Sun Microsystems Boston where I worked on the Search Inside the Music (SITM), a project to build an industrial-scale music recommender. My role was to lead the machine learning part of the project. We devised a data mining method for collecting large numbers of music-related tags assigned by listeners to songs and artists on social music networks (e.g. Pandora or Last.fm). We then learned to predict these tags from audio. The vector of predicted tags can be used as a measure of inter-song or inter-artist similarity and can thus be used to recommend new or otherwise untagged music. In [10] we present results from experiments on a social tag database of millions of tags applied to 100, 000 artists and an audio database of 90, 000 songs spanning many of the more popular of these artists. Future Work One interesting area of further investigation is that of using non-supervised dimensionality reduction techniques to construct a low-dimensional music space using features extracted from digital audio. For example, in Figure 1 see the results of finding the shortest path through a nearest neighbor graph. The graph was built using similarity distances from the SITM project [10]. A musical version of this example is available at eckdoug/tagspace I am also exploring alternative interfaces for browsing the space of similar songs such as immersive environments where listeners can change an audio stream by moving around a space constructed using graphical hints which indicate how certain movements will change the music.

4 Douglas Eck Research Statement 4 E)-.'I5/1'!"#$%&'()*'+,,-./(,*'-/'7.,'E5/#%&B 3"-/-)& -5%@J./@ 6B*-.@/@ 6/"2 6%*&,5J6/*&$5%-,5 5/:H @%)*/ *,$$)(, 1,-)2 K)LL %*6-5"1,*-)2 %*#%,.%@J./@ &5"*&, &/-.%: I"*H I/2H I,1)2,(/:)2%6-6,4@,5%1,*-)2,1/ #,)-.1,-)2 #)*:, :/"*-5B :2)66%:5/:H :2)66%:)2 :.%22/"- C5%-@/@ C5%-%6. C2",6 )1C%,*- ):/"6-%: FG6!"#$%&'()*'+,,-./(,* 01)*",2'34 +,)"4'35-6'75%/ 3*/")5'+5).,1 7./1)6'8,$1)* 9%:.),2';%)::.%*/ 7.,'<,)#'7,4)* ;)2,5%,'=-5)-%>", 35-%6-!/6:%2 <,)#'?/1@/6,56 A/B'?)1,1C,5 D5C%-)2 0)-'=-)-%: 7.,'E5/#%&B PNM QNR GNF GNG JGNF JQNR JPNM JONP JMNG Figure 1: Shortest path through tag space from Beethoven to hardcore/industrial band The Prodigy calculated using the SITM AdaBoost model (see 2007 NIPS paper [10]). Music sequence generation Representative publications: H. Jaeger and D. Eck. Can t get you out of my head: A connectionist model of cyclic rehearsal. In Modeling Communications with Robots and Virtual Humans, LNCS. Springer-Verlag, To appear. J. Paiement, D. Eck, S. Bengio, and D. Barber. A graphical model for chord progressions embedded in a psychoacoustic space. In ICML 05: Proceedings of the 22nd international conference on Machine learning, pages , New York, NY, USA, ACM Press. D. Eck and J. Schmidhuber. Finding temporal structure in music: Blues improvisation with LSTM recurrent networks. In H. Bourlard, editor, Neural Networks for Signal Processing XII, Proceedings of the 2002 IEEE Workshop, pages , New York, IEEE. Despite significant advances in our understanding of how people perceive and create music, we still suffer from a relatively primitive set of algorithms for working with music. There are indeed good synthesis algorithms for the realistic simulation of acoustic and electronic instruments. There are also effective methods for editing, combining and embellishing pieces of music. What is lacking, I believe, are algorithms for generating interesting new sequences of music, either alone as software composers or in real-time collaboration with musicians as software improvisers. Automatic music composition algorithms like these would serve two audiences. They would improve the tools already available to tech-savvy musicians while at the

5 Douglas Eck Research Statement 5 same time they would involve a new group of users, those who enjoy working with music but who lack the music-theoretic skills to compose. In addition such tools may be useful for generating music based on the changing context of a video game or film. Though all of these areas have received considerable attention, especially in traditional artificial intelligence (for example the work of David Cope [5]), fundamental challenges remain. Music sequences are difficult to learn using standard time-series analysis techniques. This is due in part to the fact that musical structure is elaborated over long time spans and is hierarchical, making it difficult or impossible to learn from local information. As an example compare the phrase John ate the [?] to the musical sequence B C E [?] In the linguistic case, the words John ate the constrain the possible values of next word [?]. In the musical case, the nearby notes B C and E do little to constrain what can come next. This is not to say that music lacks constraints. In music the constraints come from elaboration and repetition within the context of musical structure. Fortunately (for us), listeners are able to find this structure easily. For example, children old enough to jump rope can learn songs containing repeated variations of a musical phrase. Moreover the phrase can be relatively long, usually spanning several measures of music. Unfortunately (for standard sequence learning tools) this illustrates that music is elaborated over long time spans and at different timescales in parallel. Furthermore, wholesale repetition of sequences or modifications of sequences is common. Together, this leads to a difficult sequence learning task for standard timeseries models such as recurrent neural networks, hidden Markov models, ARMA models, etc Probabilistic piano roll Time Notes Figure 2: A piano roll representation showing note probabilities. Densities are taken from the transitions learned by an LSTM recurrent network trained on folk music. (Graphic from manuscript currently under preparation with graduate student Jasmin Lapalme.) One promising method for composing music automatically is to use a recurrent neural network (RNN) to capture the long-timescale regularities that constitute musical style. An RNN is a neural network with

6 Douglas Eck Research Statement 6 self-connected units. These self-connections allow an RNN to represent context and thus to discover correlations that span long time-lags in a sequence. Unfortunately, the gradient descent learning used in traditional RNNs leads to exponential decay of vital error information [1], greatly limiting in practice the kinds of sequence learning problems RNNs can handle. A potential solution to this problem lies in the hybrid RNN Long Short-Term Memory (LSTM) [12]. In previously published research done with one of LSTM s developers, I used LSTM to create an algorithm for learning to improvise music in a bebopjazz style [9]. Unlike competing methods, my LSTM composer could learn to be a stable generator of stylistically-correct music. Another promising method lies in generative graphical models. In recent work with PhD student Jean- François Paiement and co-supervisor Samy Bengio (Google) we developed a hierarchical graphical model that can generate new chord voicings for jazz standards and can also generate melodies from chord voicings [17, 16]. By modeling distance patterns, this approach can also be extended to musical rhythm [18]. Finally I have been collaborating with Herbert Jaeger from Jacobs University Bremen. We have been applying his Echo State Networks to the task of music sequence generation. Early experiments on the learning and rehearsal of cyclic patterns were promising [13] but much more work is needed to know whether this approach will scale from short cyclic patterns to the more deeply-structured music sequences. Future Work All of the approaches I have tried have yielded partial success but no machine learning algorithm has been a clear winner. (This is not surprising given the complexity of music). One direction I continue to explore is to represent metrical hierarchy either as a feature in the input sequence or in the structure of the learner. By having access to such structure a model is able to know where to look in the sequence for repetition boundaries and, more generally, for correlations spanning long timescales. Also, in certain domains it may be possible to solve a simpler problem than the general problem of composing novel musical sequences by example. In the case of video games, for example, a simpler yet still useful sequence generation task might be to adapt existing music compositions to meet specific contexts in a video game. That is, the goal of the model would not be to generate new melodies to fit a situation but rather to morph the timbre, harmonics and performance dynamics of known compositions to fit a particular game state (e.g. very dangerous ). This would require representing a music composition such that variations can be generated by adjusting parameters. One could then learn a mapping between those composition parameters and the changing state of the video game. Expression: Realistic Performance Timing and Dynamics Representative publications: D. Eck. Beat tracking using an autocorrelation phase matrix. In Proceedings of the 2007 International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages IEEE Signal Processing Society, 2007.

7 Douglas Eck Research Statement 7 D. Eck. Finding long-timescale musical structure with an autocorrelation phase matrix. Music Perception, 24(2): , D. Eck. Finding downbeats with a relaxation oscillator. Psychol. Research, 66(1):18 25, What is the difference between a musical score and a musical performance? With the growth of computergenerated music, this question is of practical as well as theoretical importance. Almost anyone who listens to a computer play a musical score notices how flat and wooden it sounds. The challenge here lies not in the ability for the computer to generate realistic musical instrument timbres. Rather it concerns how to generate realistic timing and dynamics. That is, much of what makes a performance a good one lies not in which notes are played, but rather when and how they are played 1. These aspects of performance have been researched by cognitive psychologists for nearly 20 years (see [19] for a slightly-dated but still relevant review). However it has proved difficult to quantify expressive performance such that machines can generate convincing performances. See [20] for a review of recent AI and machine learning attempts. Expressive music performance is a new area of research for me. However I have a long background in working with the related task of synchronizing in real time with music. In work that extended my dissertation I introduced a model that uses the dynamics of continually-spiking neurons [8] to find beats. The benefits of my approach are seen when large numbers of oscillating neurons are used to sample multiple levels of the metrical hierarchy. In this case, the fact that model neurons exhibit so-called relaxation oscillation becomes important. Relaxation oscillators are optimally-fast at synchronizing when coupled together in groups. This leads to rapid and stable phase-locking with a music-like input, even when dozens or hundreds of neurons are employed. For example, see Figure which displays an array of Fitzhugh-Nagumo neurons synchronizing with a metronome (middle) and a performance of the Beatles song Yesterday (right) played on a MIDI-enabled piano. More recently I developed an algorithm that extends autocorrelation-based meter detection [4] to preserve the necessary phase information for doing beat tracking [6, 7]. The core data structure of the algorithm is an phase-by-period matrix called an autocorrelation phase matrix (APM) Figure. The APM encapsulates important long-timescale dependencies in a performance. By limiting search to those lags suggested by the APM, it is possible to implement on-line beat tracking and improvisation. Finally, I have made advances in the related task of extracting note onsets from digital audio. Advances in this domain will make it possible to extract performance data directly from audio rather than from MIDI recordings which require a MIDI-enabled instrument. In work with student Alexandre Lacoste, I developed a note onset detection algorithm that used a supervised learning step to remove noise [14]. Two variants of the algorithm were 1st and 2nd out of was 1st out of nine entries in the Audio Onset Detection contest at MIREX 2005 [15]. 1 For those who are not convinced please compare two performances of a Chopin étude, one ca/ eckdoug/performance/deadpan.mp3 is played from the score by a computer (no performance dynamics or pedaling) while the other one eckdoug/performance/expressive.mp3 was played by a professional pianist. Note that the same score was used to generate both performances and that both were stored in MIDI format. The only differences are changes in timing, note velocities and the use of the pedal. For most listeners the differences between the two versions is vast.

8 Douglas Eck Research Statement 8 Figure 3: Array of 25 Fitzhugh-Nagumo spiking neurons (single waveform at left) synchronizing with a metronome (middle) and a performance of the Beatles Yesterday (right) played on a MIDI-enabled piano. Future Work In my current research I am focusing on two music performance issues. The first is the question of exactly how expressive timing and dynamics correlate with meter and grouping structure in music. Second is the question of how expressive timing and dynamics effects are achieved by performers? Currently I am in the process of collecting performance data from a large number of pianists recorded on my lab Boesendorfer concert grand piano. With student Stanislaus Lauly I am working on a number of algorithms for modeling expressive performance. To date one of the most promising is a sequential version of the Deep Belief Network (DBN) [11], an algorithm composed of multiple connected Restricted Boltzmann Machines (RBMs) trained using contrastive divergence. References [1] Y. Bengio, P. Simard, and P. Frasconi. Learning long-term dependencies with gradient descent is difficult. IEEE Transactions on Neural Networks, 5(2): , [2] J. Bergstra, N. Casagrande, and D. Eck. Genre classification: Timbre- and rhythm-based multiresolution audio classification. URL mirex-results. MIREX genre classification contest, [3] J. Bergstra, N. Casagrande, D. Erhan, D. Eck, and B. Kégl. Aggregate features and AdaBoost for music classification. Machine Learning, 65(2-3): , [4] J. Brown. Determination of meter of musical scores by autocorrelation. Journal of the Acoustical Society of America, 94: , [5] D. Cope. Computers and Musical Style. A-R Editions, Inc., Madison, Wisconsin, [6] D. Eck. Beat tracking using an autocorrelation phase matrix. In Proceedings of the 2007 International

9 Douglas Eck Research Statement 9 Figure 4: Autocorrelation phase matrix (APM) for a Cha-Cha-Cha song. The APM is useful for beat tracking and for finding deep structure (metrical hierarchy) in music audio. Conference on Acoustics, Speech and Signal Processing (ICASSP), pages IEEE Signal Processing Society, [7] D. Eck. Finding long-timescale musical structure with an autocorrelation phase matrix. Music Perception, 24(2): , [8] D. Eck. Finding downbeats with a relaxation oscillator. Psychol. Research, 66(1):18 25, [9] D. Eck and J. Schmidhuber. Finding temporal structure in music: Blues improvisation with LSTM recurrent networks. In H. Bourlard, editor, Neural Networks for Signal Processing XII, Proceedings of the 2002 IEEE Workshop, pages , New York, IEEE. [10] D. Eck, P. Lamere, T. Bertin-Mahieux, and S. Green. Automatic generation of social tags for music recommendation. In Neural Information Processing Systems Conference (NIPS) 20, [11] G. E. Hinton, S. Osindero, and Y.-W. Teh. A fast learning algorithm for deep belief nets. Neural Comput., 18(7): , ISSN doi: [12] S. Hochreiter and J. Schmidhuber. Long Short-Term Memory. Neural Computation, 9(8): , [13] H. Jaeger and D. Eck. Can t get you out of my head: A connectionist model of cyclic rehearsal. In Modeling Communications with Robots and Virtual Humans, LNCS. Springer-Verlag, To appear. [14] A. Lacoste and D. Eck. A supervised classification algorithm for note onset detection. EURASIP Journal on Applied Signal Processing, 2007(ID 43745):1 13, 2007.

10 Douglas Eck Research Statement 10 [15] A. Lacoste and D. Eck. Onset detection with artificial neural networks. URL music-ir.org/evaluation/mirex-results. MIREX note onset detection contest, [16] J. Paiement, D. Eck, and S. Bengio. A probabilistic model for chord progressions. In Proceedings of the 6th International Conference on Music Information Retrieval (ISMIR 2005), pages , London: University of London, [17] J. Paiement, D. Eck, S. Bengio, and D. Barber. A graphical model for chord progressions embedded in a psychoacoustic space. In ICML 05: Proceedings of the 22nd international conference on Machine learning, pages , New York, NY, USA, ACM Press. [18] J. Paiement, Y. Grandvalet, S. Bengio, and D. Eck. A generative model for rhythms. NIPS 2007 Workshop on Music, Brain and Cognition, [19] C. Palmer. Music performance. Annual Review of Psychology, 48: , URL http: //upload.mcgill.ca/spl/annrev97.pdf. [20] G. Widmer and W. Goebl. Computational models of expressive music performance: The state of the art. Journal of New Music Research, 33(3): , 2004.

Measuring & Modeling Musical Expression

Measuring & Modeling Musical Expression Measuring & Modeling Musical Expression Douglas Eck University of Montreal Department of Computer Science BRAMS Brain Music and Sound International Laboratory for Brain, Music and Sound Research Overview

More information

arxiv: v1 [cs.lg] 15 Jun 2016

arxiv: v1 [cs.lg] 15 Jun 2016 Deep Learning for Music arxiv:1606.04930v1 [cs.lg] 15 Jun 2016 Allen Huang Department of Management Science and Engineering Stanford University allenh@cs.stanford.edu Abstract Raymond Wu Department of

More information

Learning Musical Structure Directly from Sequences of Music

Learning Musical Structure Directly from Sequences of Music Learning Musical Structure Directly from Sequences of Music Douglas Eck and Jasmin Lapalme Dept. IRO, Université de Montréal C.P. 6128, Montreal, Qc, H3C 3J7, Canada Technical Report 1300 Abstract This

More information

Supervised Learning in Genre Classification

Supervised Learning in Genre Classification Supervised Learning in Genre Classification Introduction & Motivation Mohit Rajani and Luke Ekkizogloy {i.mohit,luke.ekkizogloy}@gmail.com Stanford University, CS229: Machine Learning, 2009 Now that music

More information

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? NICHOLAS BORG AND GEORGE HOKKANEN Abstract. The possibility of a hit song prediction algorithm is both academically interesting and industry motivated.

More information

Introductions to Music Information Retrieval

Introductions to Music Information Retrieval Introductions to Music Information Retrieval ECE 272/472 Audio Signal Processing Bochen Li University of Rochester Wish List For music learners/performers While I play the piano, turn the page for me Tell

More information

Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies

Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies Judy Franklin Computer Science Department Smith College Northampton, MA 01063 Abstract Recurrent (neural) networks have

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

MUSI-6201 Computational Music Analysis

MUSI-6201 Computational Music Analysis MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)

More information

Computer Coordination With Popular Music: A New Research Agenda 1

Computer Coordination With Popular Music: A New Research Agenda 1 Computer Coordination With Popular Music: A New Research Agenda 1 Roger B. Dannenberg roger.dannenberg@cs.cmu.edu http://www.cs.cmu.edu/~rbd School of Computer Science Carnegie Mellon University Pittsburgh,

More information

Music Composition with RNN

Music Composition with RNN Music Composition with RNN Jason Wang Department of Statistics Stanford University zwang01@stanford.edu Abstract Music composition is an interesting problem that tests the creativity capacities of artificial

More information

Music Information Retrieval with Temporal Features and Timbre

Music Information Retrieval with Temporal Features and Timbre Music Information Retrieval with Temporal Features and Timbre Angelina A. Tzacheva and Keith J. Bell University of South Carolina Upstate, Department of Informatics 800 University Way, Spartanburg, SC

More information

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC Scientific Journal of Impact Factor (SJIF): 5.71 International Journal of Advance Engineering and Research Development Volume 5, Issue 04, April -2018 e-issn (O): 2348-4470 p-issn (P): 2348-6406 MUSICAL

More information

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION ULAŞ BAĞCI AND ENGIN ERZIN arxiv:0907.3220v1 [cs.sd] 18 Jul 2009 ABSTRACT. Music genre classification is an essential tool for

More information

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception LEARNING AUDIO SHEET MUSIC CORRESPONDENCES Matthias Dorfer Department of Computational Perception Short Introduction... I am a PhD Candidate in the Department of Computational Perception at Johannes Kepler

More information

Subjective Similarity of Music: Data Collection for Individuality Analysis

Subjective Similarity of Music: Data Collection for Individuality Analysis Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp

More information

Chord Classification of an Audio Signal using Artificial Neural Network

Chord Classification of an Audio Signal using Artificial Neural Network Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------

More information

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval DAY 1 Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval Jay LeBoeuf Imagine Research jay{at}imagine-research.com Rebecca

More information

Robert Alexandru Dobre, Cristian Negrescu

Robert Alexandru Dobre, Cristian Negrescu ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Automatic Music Transcription Software Based on Constant Q

More information

Automatic Music Genre Classification

Automatic Music Genre Classification Automatic Music Genre Classification Nathan YongHoon Kwon, SUNY Binghamton Ingrid Tchakoua, Jackson State University Matthew Pietrosanu, University of Alberta Freya Fu, Colorado State University Yue Wang,

More information

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Aric Bartle (abartle@stanford.edu) December 14, 2012 1 Background The field of composer recognition has

More information

LSTM Neural Style Transfer in Music Using Computational Musicology

LSTM Neural Style Transfer in Music Using Computational Musicology LSTM Neural Style Transfer in Music Using Computational Musicology Jett Oristaglio Dartmouth College, June 4 2017 1. Introduction In the 2016 paper A Neural Algorithm of Artistic Style, Gatys et al. discovered

More information

CS229 Project Report Polyphonic Piano Transcription

CS229 Project Report Polyphonic Piano Transcription CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project

More information

Effects of acoustic degradations on cover song recognition

Effects of acoustic degradations on cover song recognition Signal Processing in Acoustics: Paper 68 Effects of acoustic degradations on cover song recognition Julien Osmalskyj (a), Jean-Jacques Embrechts (b) (a) University of Liège, Belgium, josmalsky@ulg.ac.be

More information

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon A Study of Synchronization of Audio Data with Symbolic Data Music254 Project Report Spring 2007 SongHui Chon Abstract This paper provides an overview of the problem of audio and symbolic synchronization.

More information

Enhancing Music Maps

Enhancing Music Maps Enhancing Music Maps Jakob Frank Vienna University of Technology, Vienna, Austria http://www.ifs.tuwien.ac.at/mir frank@ifs.tuwien.ac.at Abstract. Private as well as commercial music collections keep growing

More information

Singer Traits Identification using Deep Neural Network

Singer Traits Identification using Deep Neural Network Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic

More information

Hidden Markov Model based dance recognition

Hidden Markov Model based dance recognition Hidden Markov Model based dance recognition Dragutin Hrenek, Nenad Mikša, Robert Perica, Pavle Prentašić and Boris Trubić University of Zagreb, Faculty of Electrical Engineering and Computing Unska 3,

More information

Computational Modelling of Harmony

Computational Modelling of Harmony Computational Modelling of Harmony Simon Dixon Centre for Digital Music, Queen Mary University of London, Mile End Rd, London E1 4NS, UK simon.dixon@elec.qmul.ac.uk http://www.elec.qmul.ac.uk/people/simond

More information

Statistical Modeling and Retrieval of Polyphonic Music

Statistical Modeling and Retrieval of Polyphonic Music Statistical Modeling and Retrieval of Polyphonic Music Erdem Unal Panayiotis G. Georgiou and Shrikanth S. Narayanan Speech Analysis and Interpretation Laboratory University of Southern California Los Angeles,

More information

Automatic Rhythmic Notation from Single Voice Audio Sources

Automatic Rhythmic Notation from Single Voice Audio Sources Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung

More information

Music Similarity and Cover Song Identification: The Case of Jazz

Music Similarity and Cover Song Identification: The Case of Jazz Music Similarity and Cover Song Identification: The Case of Jazz Simon Dixon and Peter Foster s.e.dixon@qmul.ac.uk Centre for Digital Music School of Electronic Engineering and Computer Science Queen Mary

More information

The Million Song Dataset

The Million Song Dataset The Million Song Dataset AUDIO FEATURES The Million Song Dataset There is no data like more data Bob Mercer of IBM (1985). T. Bertin-Mahieux, D.P.W. Ellis, B. Whitman, P. Lamere, The Million Song Dataset,

More information

hit), and assume that longer incidental sounds (forest noise, water, wind noise) resemble a Gaussian noise distribution.

hit), and assume that longer incidental sounds (forest noise, water, wind noise) resemble a Gaussian noise distribution. CS 229 FINAL PROJECT A SOUNDHOUND FOR THE SOUNDS OF HOUNDS WEAKLY SUPERVISED MODELING OF ANIMAL SOUNDS ROBERT COLCORD, ETHAN GELLER, MATTHEW HORTON Abstract: We propose a hybrid approach to generating

More information

Deep learning for music data processing

Deep learning for music data processing Deep learning for music data processing A personal (re)view of the state-of-the-art Jordi Pons www.jordipons.me Music Technology Group, DTIC, Universitat Pompeu Fabra, Barcelona. 31st January 2017 Jordi

More information

Week 14 Music Understanding and Classification

Week 14 Music Understanding and Classification Week 14 Music Understanding and Classification Roger B. Dannenberg Professor of Computer Science, Music & Art Overview n Music Style Classification n What s a classifier? n Naïve Bayesian Classifiers n

More information

An ecological approach to multimodal subjective music similarity perception

An ecological approach to multimodal subjective music similarity perception An ecological approach to multimodal subjective music similarity perception Stephan Baumann German Research Center for AI, Germany www.dfki.uni-kl.de/~baumann John Halloran Interact Lab, Department of

More information

Speech and Speaker Recognition for the Command of an Industrial Robot

Speech and Speaker Recognition for the Command of an Industrial Robot Speech and Speaker Recognition for the Command of an Industrial Robot CLAUDIA MOISA*, HELGA SILAGHI*, ANDREI SILAGHI** *Dept. of Electric Drives and Automation University of Oradea University Street, nr.

More information

SYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS

SYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS Published by Institute of Electrical Engineers (IEE). 1998 IEE, Paul Masri, Nishan Canagarajah Colloquium on "Audio and Music Technology"; November 1998, London. Digest No. 98/470 SYNTHESIS FROM MUSICAL

More information

A Beat Tracking System for Audio Signals

A Beat Tracking System for Audio Signals A Beat Tracking System for Audio Signals Simon Dixon Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria. simon@ai.univie.ac.at April 7, 2000 Abstract We present

More information

Audio. Meinard Müller. Beethoven, Bach, and Billions of Bytes. International Audio Laboratories Erlangen. International Audio Laboratories Erlangen

Audio. Meinard Müller. Beethoven, Bach, and Billions of Bytes. International Audio Laboratories Erlangen. International Audio Laboratories Erlangen Meinard Müller Beethoven, Bach, and Billions of Bytes When Music meets Computer Science Meinard Müller International Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de School of Mathematics University

More information

Data Driven Music Understanding

Data Driven Music Understanding Data Driven Music Understanding Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Engineering, Columbia University, NY USA http://labrosa.ee.columbia.edu/ 1. Motivation:

More information

Extracting Significant Patterns from Musical Strings: Some Interesting Problems.

Extracting Significant Patterns from Musical Strings: Some Interesting Problems. Extracting Significant Patterns from Musical Strings: Some Interesting Problems. Emilios Cambouropoulos Austrian Research Institute for Artificial Intelligence Vienna, Austria emilios@ai.univie.ac.at Abstract

More information

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene Beat Extraction from Expressive Musical Performances Simon Dixon, Werner Goebl and Emilios Cambouropoulos Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria.

More information

Music Radar: A Web-based Query by Humming System

Music Radar: A Web-based Query by Humming System Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,

More information

Acoustic Scene Classification

Acoustic Scene Classification Acoustic Scene Classification Marc-Christoph Gerasch Seminar Topics in Computer Music - Acoustic Scene Classification 6/24/2015 1 Outline Acoustic Scene Classification - definition History and state of

More information

IMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM

IMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM IMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM Thomas Lidy, Andreas Rauber Vienna University of Technology, Austria Department of Software

More information

A Graphical Model for Chord Progressions Embedded in a Psychoacoustic Space

A Graphical Model for Chord Progressions Embedded in a Psychoacoustic Space Embedded in a Psychoacoustic Space Jean-François Paiement paiement@idiap.ch IDIAP Research Institute, Rue du Simplon 4, Case Postale 592, CH-1920 Martigny, Switzerland Douglas Eck eckdoug@iro.umontreal.ca

More information

CTP431- Music and Audio Computing Music Information Retrieval. Graduate School of Culture Technology KAIST Juhan Nam

CTP431- Music and Audio Computing Music Information Retrieval. Graduate School of Culture Technology KAIST Juhan Nam CTP431- Music and Audio Computing Music Information Retrieval Graduate School of Culture Technology KAIST Juhan Nam 1 Introduction ü Instrument: Piano ü Genre: Classical ü Composer: Chopin ü Key: E-minor

More information

Music Genre Classification and Variance Comparison on Number of Genres

Music Genre Classification and Variance Comparison on Number of Genres Music Genre Classification and Variance Comparison on Number of Genres Miguel Francisco, miguelf@stanford.edu Dong Myung Kim, dmk8265@stanford.edu 1 Abstract In this project we apply machine learning techniques

More information

A STUDY ON LSTM NETWORKS FOR POLYPHONIC MUSIC SEQUENCE MODELLING

A STUDY ON LSTM NETWORKS FOR POLYPHONIC MUSIC SEQUENCE MODELLING A STUDY ON LSTM NETWORKS FOR POLYPHONIC MUSIC SEQUENCE MODELLING Adrien Ycart and Emmanouil Benetos Centre for Digital Music, Queen Mary University of London, UK {a.ycart, emmanouil.benetos}@qmul.ac.uk

More information

A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS

A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS Mutian Fu 1 Guangyu Xia 2 Roger Dannenberg 2 Larry Wasserman 2 1 School of Music, Carnegie Mellon University, USA 2 School of Computer

More information

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 AN HMM BASED INVESTIGATION OF DIFFERENCES BETWEEN MUSICAL INSTRUMENTS OF THE SAME TYPE PACS: 43.75.-z Eichner, Matthias; Wolff, Matthias;

More information

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS Andrew N. Robertson, Mark D. Plumbley Centre for Digital Music

More information

Lecture 9 Source Separation

Lecture 9 Source Separation 10420CS 573100 音樂資訊檢索 Music Information Retrieval Lecture 9 Source Separation Yi-Hsuan Yang Ph.D. http://www.citi.sinica.edu.tw/pages/yang/ yang@citi.sinica.edu.tw Music & Audio Computing Lab, Research

More information

Detecting Musical Key with Supervised Learning

Detecting Musical Key with Supervised Learning Detecting Musical Key with Supervised Learning Robert Mahieu Department of Electrical Engineering Stanford University rmahieu@stanford.edu Abstract This paper proposes and tests performance of two different

More information

ESP: Expression Synthesis Project

ESP: Expression Synthesis Project ESP: Expression Synthesis Project 1. Research Team Project Leader: Other Faculty: Graduate Students: Undergraduate Students: Prof. Elaine Chew, Industrial and Systems Engineering Prof. Alexandre R.J. François,

More information

Toward a Computationally-Enhanced Acoustic Grand Piano

Toward a Computationally-Enhanced Acoustic Grand Piano Toward a Computationally-Enhanced Acoustic Grand Piano Andrew McPherson Electrical & Computer Engineering Drexel University 3141 Chestnut St. Philadelphia, PA 19104 USA apm@drexel.edu Youngmoo Kim Electrical

More information

Tempo and Beat Analysis

Tempo and Beat Analysis Advanced Course Computer Science Music Processing Summer Term 2010 Meinard Müller, Peter Grosche Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Tempo and Beat Analysis Musical Properties:

More information

Algorithmic Music Composition

Algorithmic Music Composition Algorithmic Music Composition MUS-15 Jan Dreier July 6, 2015 1 Introduction The goal of algorithmic music composition is to automate the process of creating music. One wants to create pleasant music without

More information

Methods for the automatic structural analysis of music. Jordan B. L. Smith CIRMMT Workshop on Structural Analysis of Music 26 March 2010

Methods for the automatic structural analysis of music. Jordan B. L. Smith CIRMMT Workshop on Structural Analysis of Music 26 March 2010 1 Methods for the automatic structural analysis of music Jordan B. L. Smith CIRMMT Workshop on Structural Analysis of Music 26 March 2010 2 The problem Going from sound to structure 2 The problem Going

More information

EE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function

EE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function EE391 Special Report (Spring 25) Automatic Chord Recognition Using A Summary Autocorrelation Function Advisor: Professor Julius Smith Kyogu Lee Center for Computer Research in Music and Acoustics (CCRMA)

More information

The Sparsity of Simple Recurrent Networks in Musical Structure Learning

The Sparsity of Simple Recurrent Networks in Musical Structure Learning The Sparsity of Simple Recurrent Networks in Musical Structure Learning Kat R. Agres (kra9@cornell.edu) Department of Psychology, Cornell University, 211 Uris Hall Ithaca, NY 14853 USA Jordan E. DeLong

More information

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the

More information

Music Information Retrieval

Music Information Retrieval CTP 431 Music and Audio Computing Music Information Retrieval Graduate School of Culture Technology (GSCT) Juhan Nam 1 Introduction ü Instrument: Piano ü Composer: Chopin ü Key: E-minor ü Melody - ELO

More information

Sudhanshu Gautam *1, Sarita Soni 2. M-Tech Computer Science, BBAU Central University, Lucknow, Uttar Pradesh, India

Sudhanshu Gautam *1, Sarita Soni 2. M-Tech Computer Science, BBAU Central University, Lucknow, Uttar Pradesh, India International Journal of Scientific Research in Computer Science, Engineering and Information Technology 2018 IJSRCSEIT Volume 3 Issue 3 ISSN : 2456-3307 Artificial Intelligence Techniques for Music Composition

More information

A System for Acoustic Chord Transcription and Key Extraction from Audio Using Hidden Markov models Trained on Synthesized Audio

A System for Acoustic Chord Transcription and Key Extraction from Audio Using Hidden Markov models Trained on Synthesized Audio Curriculum Vitae Kyogu Lee Advanced Technology Center, Gracenote Inc. 2000 Powell Street, Suite 1380 Emeryville, CA 94608 USA Tel) 1-510-428-7296 Fax) 1-510-547-9681 klee@gracenote.com kglee@ccrma.stanford.edu

More information

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST) Computational Models of Music Similarity 1 Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST) Abstract The perceived similarity of two pieces of music is multi-dimensional,

More information

Semi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis

Semi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis Semi-automated extraction of expressive performance information from acoustic recordings of piano music Andrew Earis Outline Parameters of expressive piano performance Scientific techniques: Fourier transform

More information

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION Halfdan Rump, Shigeki Miyabe, Emiru Tsunoo, Nobukata Ono, Shigeki Sagama The University of Tokyo, Graduate

More information

A prototype system for rule-based expressive modifications of audio recordings

A prototype system for rule-based expressive modifications of audio recordings International Symposium on Performance Science ISBN 0-00-000000-0 / 000-0-00-000000-0 The Author 2007, Published by the AEC All rights reserved A prototype system for rule-based expressive modifications

More information

Music Representations

Music Representations Lecture Music Processing Music Representations Meinard Müller International Audio Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de Book: Fundamentals of Music Processing Meinard Müller Fundamentals

More information

Music Information Retrieval

Music Information Retrieval Music Information Retrieval When Music Meets Computer Science Meinard Müller International Audio Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de Berlin MIR Meetup 20.03.2017 Meinard Müller

More information

CHORD GENERATION FROM SYMBOLIC MELODY USING BLSTM NETWORKS

CHORD GENERATION FROM SYMBOLIC MELODY USING BLSTM NETWORKS CHORD GENERATION FROM SYMBOLIC MELODY USING BLSTM NETWORKS Hyungui Lim 1,2, Seungyeon Rhyu 1 and Kyogu Lee 1,2 3 Music and Audio Research Group, Graduate School of Convergence Science and Technology 4

More information

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM A QUER B EAMPLE MUSIC RETRIEVAL ALGORITHM H. HARB AND L. CHEN Maths-Info department, Ecole Centrale de Lyon. 36, av. Guy de Collongue, 69134, Ecully, France, EUROPE E-mail: {hadi.harb, liming.chen}@ec-lyon.fr

More information

Automatic Musical Pattern Feature Extraction Using Convolutional Neural Network

Automatic Musical Pattern Feature Extraction Using Convolutional Neural Network Automatic Musical Pattern Feature Extraction Using Convolutional Neural Network Tom LH. Li, Antoni B. Chan and Andy HW. Chun Abstract Music genre classification has been a challenging yet promising task

More information

Audio Feature Extraction for Corpus Analysis

Audio Feature Extraction for Corpus Analysis Audio Feature Extraction for Corpus Analysis Anja Volk Sound and Music Technology 5 Dec 2017 1 Corpus analysis What is corpus analysis study a large corpus of music for gaining insights on general trends

More information

Paulo V. K. Borges. Flat 1, 50A, Cephas Av. London, UK, E1 4AR (+44) PRESENTATION

Paulo V. K. Borges. Flat 1, 50A, Cephas Av. London, UK, E1 4AR (+44) PRESENTATION Paulo V. K. Borges Flat 1, 50A, Cephas Av. London, UK, E1 4AR (+44) 07942084331 vini@ieee.org PRESENTATION Electronic engineer working as researcher at University of London. Doctorate in digital image/video

More information

CLASSIFICATION OF MUSICAL METRE WITH AUTOCORRELATION AND DISCRIMINANT FUNCTIONS

CLASSIFICATION OF MUSICAL METRE WITH AUTOCORRELATION AND DISCRIMINANT FUNCTIONS CLASSIFICATION OF MUSICAL METRE WITH AUTOCORRELATION AND DISCRIMINANT FUNCTIONS Petri Toiviainen Department of Music University of Jyväskylä Finland ptoiviai@campus.jyu.fi Tuomas Eerola Department of Music

More information

Modeling Musical Context Using Word2vec

Modeling Musical Context Using Word2vec Modeling Musical Context Using Word2vec D. Herremans 1 and C.-H. Chuan 2 1 Queen Mary University of London, London, UK 2 University of North Florida, Jacksonville, USA We present a semantic vector space

More information

MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES

MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES Jun Wu, Yu Kitano, Stanislaw Andrzej Raczynski, Shigeki Miyabe, Takuya Nishimoto, Nobutaka Ono and Shigeki Sagayama The Graduate

More information

Figured Bass and Tonality Recognition Jerome Barthélemy Ircam 1 Place Igor Stravinsky Paris France

Figured Bass and Tonality Recognition Jerome Barthélemy Ircam 1 Place Igor Stravinsky Paris France Figured Bass and Tonality Recognition Jerome Barthélemy Ircam 1 Place Igor Stravinsky 75004 Paris France 33 01 44 78 48 43 jerome.barthelemy@ircam.fr Alain Bonardi Ircam 1 Place Igor Stravinsky 75004 Paris

More information

GRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM

GRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM 19th European Signal Processing Conference (EUSIPCO 2011) Barcelona, Spain, August 29 - September 2, 2011 GRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM Tomoko Matsui

More information

Piano Transcription MUMT611 Presentation III 1 March, Hankinson, 1/15

Piano Transcription MUMT611 Presentation III 1 March, Hankinson, 1/15 Piano Transcription MUMT611 Presentation III 1 March, 2007 Hankinson, 1/15 Outline Introduction Techniques Comb Filtering & Autocorrelation HMMs Blackboard Systems & Fuzzy Logic Neural Networks Examples

More information

Estimating the Time to Reach a Target Frequency in Singing

Estimating the Time to Reach a Target Frequency in Singing THE NEUROSCIENCES AND MUSIC III: DISORDERS AND PLASTICITY Estimating the Time to Reach a Target Frequency in Singing Sean Hutchins a and David Campbell b a Department of Psychology, McGill University,

More information

Analysing Musical Pieces Using harmony-analyser.org Tools

Analysing Musical Pieces Using harmony-analyser.org Tools Analysing Musical Pieces Using harmony-analyser.org Tools Ladislav Maršík Dept. of Software Engineering, Faculty of Mathematics and Physics Charles University, Malostranské nám. 25, 118 00 Prague 1, Czech

More information

Music Performance Panel: NICI / MMM Position Statement

Music Performance Panel: NICI / MMM Position Statement Music Performance Panel: NICI / MMM Position Statement Peter Desain, Henkjan Honing and Renee Timmers Music, Mind, Machine Group NICI, University of Nijmegen mmm@nici.kun.nl, www.nici.kun.nl/mmm In this

More information

A System for Automatic Chord Transcription from Audio Using Genre-Specific Hidden Markov Models

A System for Automatic Chord Transcription from Audio Using Genre-Specific Hidden Markov Models A System for Automatic Chord Transcription from Audio Using Genre-Specific Hidden Markov Models Kyogu Lee Center for Computer Research in Music and Acoustics Stanford University, Stanford CA 94305, USA

More information

Blues Improviser. Greg Nelson Nam Nguyen

Blues Improviser. Greg Nelson Nam Nguyen Blues Improviser Greg Nelson (gregoryn@cs.utah.edu) Nam Nguyen (namphuon@cs.utah.edu) Department of Computer Science University of Utah Salt Lake City, UT 84112 Abstract Computer-generated music has long

More information

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring 2009 Week 6 Class Notes Pitch Perception Introduction Pitch may be described as that attribute of auditory sensation in terms

More information

Chord Label Personalization through Deep Learning of Integrated Harmonic Interval-based Representations

Chord Label Personalization through Deep Learning of Integrated Harmonic Interval-based Representations Chord Label Personalization through Deep Learning of Integrated Harmonic Interval-based Representations Hendrik Vincent Koops 1, W. Bas de Haas 2, Jeroen Bransen 2, and Anja Volk 1 arxiv:1706.09552v1 [cs.sd]

More information

Outline. Why do we classify? Audio Classification

Outline. Why do we classify? Audio Classification Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification Implementation Future Work Why do we classify

More information

Music Genre Classification

Music Genre Classification Music Genre Classification chunya25 Fall 2017 1 Introduction A genre is defined as a category of artistic composition, characterized by similarities in form, style, or subject matter. [1] Some researchers

More information

Transcription of the Singing Melody in Polyphonic Music

Transcription of the Singing Melody in Polyphonic Music Transcription of the Singing Melody in Polyphonic Music Matti Ryynänen and Anssi Klapuri Institute of Signal Processing, Tampere University Of Technology P.O.Box 553, FI-33101 Tampere, Finland {matti.ryynanen,

More information

Pitch Perception and Grouping. HST.723 Neural Coding and Perception of Sound

Pitch Perception and Grouping. HST.723 Neural Coding and Perception of Sound Pitch Perception and Grouping HST.723 Neural Coding and Perception of Sound Pitch Perception. I. Pure Tones The pitch of a pure tone is strongly related to the tone s frequency, although there are small

More information

A Categorical Approach for Recognizing Emotional Effects of Music

A Categorical Approach for Recognizing Emotional Effects of Music A Categorical Approach for Recognizing Emotional Effects of Music Mohsen Sahraei Ardakani 1 and Ehsan Arbabi School of Electrical and Computer Engineering, College of Engineering, University of Tehran,

More information

Voice & Music Pattern Extraction: A Review

Voice & Music Pattern Extraction: A Review Voice & Music Pattern Extraction: A Review 1 Pooja Gautam 1 and B S Kaushik 2 Electronics & Telecommunication Department RCET, Bhilai, Bhilai (C.G.) India pooja0309pari@gmail.com 2 Electrical & Instrumentation

More information

Audio-Based Video Editing with Two-Channel Microphone

Audio-Based Video Editing with Two-Channel Microphone Audio-Based Video Editing with Two-Channel Microphone Tetsuya Takiguchi Organization of Advanced Science and Technology Kobe University, Japan takigu@kobe-u.ac.jp Yasuo Ariki Organization of Advanced Science

More information

Automatic Laughter Detection

Automatic Laughter Detection Automatic Laughter Detection Mary Knox Final Project (EECS 94) knoxm@eecs.berkeley.edu December 1, 006 1 Introduction Laughter is a powerful cue in communication. It communicates to listeners the emotional

More information

Automatic music transcription

Automatic music transcription Music transcription 1 Music transcription 2 Automatic music transcription Sources: * Klapuri, Introduction to music transcription, 2006. www.cs.tut.fi/sgn/arg/klap/amt-intro.pdf * Klapuri, Eronen, Astola:

More information