Hidden Markov Model based dance recognition

Size: px
Start display at page:

Download "Hidden Markov Model based dance recognition"

Transcription

1 Hidden Markov Model based dance recognition Dragutin Hrenek, Nenad Mikša, Robert Perica, Pavle Prentašić and Boris Trubić University of Zagreb, Faculty of Electrical Engineering and Computing Unska 3, Zagreb, Croatia Abstract In this paper we describe a dance classification system for compositions written in MIDI format. The system recognizes the following dances: tango, polka, mazurka, waltz, cha-cha-cha and march. The rhytmic structure of a dance is a finite sequence of notes of specified durations that repeats itself through the whole composition, so we can hypothesise that the probability of occurence of specified note duration depends on the duration of the note before it. Hence the implementation of the classifier is made using Hidden Markov Models. The models are used in two basic forms the first assumes discrete note durations, and the other assumes that note durations conform to normal distribution. The system was tested using dance-prototype generated examples with added Gaussian noise, as well as with human-played examples. The results gathered using both kinds of examples are comparable. The system was implemented using the Matlab programming package. I. INTRODUCTION Upon hearing a certain sequence of notes or rhythm, a dance expert or even a dance enthusiast immediately thinks of some type of dance or movement which would best fit the heard music. Thus, he/she easily recognizes the type of dance or music that is being played. Computer is not able to do the same with such ease, as it is unable to focus on a specific musical instrument in the audio recording. MIDI (Musical Instrument Digital Interface) format is commonly used in musical production, besides mp3, wave and similar formats. MIDI is a protocol by which computer communicates with certain external devices, such as keyboards. The protocol is based on exchange of messages between the device and the computer. Those messages can be saved in a file and interpreted later as an audio or as a note inscription. The protocol is a standard which is used by all musical instruments and musical software, but the problem is that most of the devices and software do not honor the protocol specifications exactly. Therefore, it s a common situation when a note inscription written in one program and saved in MIDI format, when opened in an another program, is poorly correlated with the source inscription. The problem resides in the fact that each note can be written in MIDI format in various ways. This increases the possibility of misinterpretaion of the recording. One of the most common notes is a quarter note. In a 4/4 measure it is represented by one tick. For example, if the tempo is 120 and measure 4/4, then there should be 120 quarter notes in one minute of recording. Thus, every quarter note should last exactly half a second. But this holds only on average. Let us assume that each quarter note lasts 100 ticks of the clock. The each quaver should last 50 ticks and dotted quaver should last 75 ticks. The dot in the note increases its duration by 50%. For example, dotted quarter note lasts the same as the quarter note and the quaver together. If the music is played by human then the quarter note lasts 100 ticks only on average, but it can last a bit more or less, e.g. 102 ticks or 85 ticks. This depends on the melody phrasing and other factors. Musical inscription software often adds noise to the duration of notes when saving in the MIDI format in order to achieve the greater fidelity of the recording as if it has been played by a human. This makes the correct interpretation of the note difficult to the computer, as for example, the note that lasts 85 ticks is much closer to a dotted quaver than to a quarter note. Thus, the quarter note is often not a real quarter note. This is the reason for wrong interpretation of notes among different programs. Because of this problem, classification of dances by rhythmic patterns obtained from MIDI files is a very challenging problem in computer science. In this paper we present methods that enable the computer to recognize the dances based on the human-labeled examples. The next section gives an overview of the previous works and solutions of the described problem. The third section describes a method for classification of musical pieces with the Hidden Markov Model. The fourth section describes the results of classification. The fifth section concludes the paper and discusses aspects of future work. II. PREVIOUS WORK The described problem is tightly related with the problem of detecting the rhythmic structure of the musical piece. Takeda et al. define the problem as a search for a sequence of states in a probabilistic model [1]. Since the states are represented with Hidden Markov Models, the most probable sequence of states can be find with the well-known Viterbi algorithm [2]. Therefore, the rhythmic structure is determined by the most probable sequence of states found by the Viterbi algorithm for the given sequence of observations. This method is good for finding the specific rhythmic structure, but it is impractical for classification of rhythmic structures. In [3], the system for extraction of musical features from MIDI recording is described. The described system consists of more subsystems for carrying out the following tasks: identifying basic musical objects (notes, pauses, chords, etc.), searching for accent on each musical object, rhythm recognition, rhythm tracking and note discretization. The rhythmic structure of the piece is recognized by looking into the time

2 interval which consists of certain number of notes. This time interval is determined in advance for each potential rhythmic structure that is being recognized. The actual notes in that interval are then compared to the expected notes and then the classification is performed. This method is not practical for solving our problem as it does not give good results. In [4] methods for note duration discretization and methods for detection and tracking of rhythm are presented. The rhythm detection in this paper is based on Hidden Markov Models in such a way that each state of the model represents the moment in which the note has been played. This enables the modeling of different moments in which a note can appear. This method is very useful for converting MIDI recordings into printable musical inscription. III. METHOD DESCRIPTION It is a general trend to use Hidden Markov Models (HMMs) for solving pattern recognition problems in cases where patterns are time dependant signals, as for example in speech recognition [2]. MIDI signals are time dependant signals and they represent a more abstract way for representing music in computer. It is much easier to extract note characteristics from MIDI recording than from mp3 or wave. Hence, we think that it would be a good idea to use HMMs for classification of musical pieces recorded in MIDI format. The idea behind HMMs assumes the existence of some set of states Q = {q i } N i=1, where N is the number of states. For each state we define probabilities of transition from the current state into all other states and probability for staying in current state. Furthermore, for each state we define its a priori probability (prior), i.e. the probability that the system will start in this state. Besides the set of states, there exists a set of possible outputs of the system V = {v j } M j=1, where M is number of possible outputs. For each state of the system, we define the probability that the system will generate a certain output while being in that state. All that can be formally written in the following way: Hidden Markov Model λ is a tuple λ = (Λ, B, Π) where Λ is a transition probability matrix, B is an output probability matrix and Π is a vector of priors. Elements of the matrix Λ are a ij and represent the probability of transition from state i to state j, i.e. a ij = P (q t+1 = j q t = i) The elements of the matrix B are b ij and represent the probability that the output j will be generated while the system is in state i, i.e. b ij = P (output = v j q t = i) The elements of the vector Π are π i and represent the probability that the system will start its work in the state i, i.e. π i = P (q 1 = i) As a result of such definition of HMM, it is suitable to represent it in a form of a directed graph. Vertices of the graph represent the states of the HMM and the outputs of the system, while the edges represent possible transitions between states and possible outputs of the system for each state. The weights of the edges represent probabilities. An example of a HMM is shown in figure 1. Figure 1. An example of a Hidden Markov Model. X represents the states, Y represents the possible outputs of the system, a represents the transition probabilities and b represents the probabilities of outputs in each state. Possible outputs of the system can be continuous too. In that case for each state we have to model the probability distribution which will generate the outputs of the system in that state, e.g. Gaussian distribution. In general, it is possible to model different probability distribution functions for each state, but it is common to use the same probability distribution function in all states, but with different parameters. This simplifies the usage of the model and the learning algorithm. Possible outputs of the system depend on the problem we try to model using HMMs. On the other hand, the number of states is a parameter of the model and thus influences the complexity of the learning. A. Data preparation and feature selection Our system recognizes dances using their rhythmic structures. Rhythmic structure is a sequence of notes of certain duration, i.e. the alternation of sound and silence in time. Rhythmic structure examples that can be recognized by our classifier are shown in figure 2. The duration of a note is the only feature that is used by our classifier as it is the only required feature to describe the rhythmic structure of a dance. B. Note discretization In cases when we want to test the classification of musical pieces by having notes represented by their class, we first need to perform note discretization, i.e. classify them into some

3 Figure 2. (a) Tango rhythm (b) Polka rhythm (c) Mazurka rhythm (d) Waltz rhythm (e) Cha-cha-cha rhythm (f) March rhythm Rhythmic structures of dances recognizable by our system class of notes is a note a quaver, a quarter note, a half note, etc. For discretization of notes we use a modified k Nearest Neighbours (knn) classifier, which determines the type of note based on its duration and examples read from the learning database. This means that the classifier reads a duration of a note from a MIDI file and then determines whether the given duration is a duration of a quaver, a quarter note, a half note, etc. Every type of note has its own identification number or index which is then used as a feature in a HMM based classifier. Thus semiquaver has an index 1, dotted semiquaver has an index 2, quaver has an index 3, dotted quaver has an index 4, quarter note has an index 5, dotted quarter note has an index 6, half note has an index 7, dotted half note has an index 8 and a whole note has an index 9. Such discrete notes are then used for learning the Hidden Markov Models. The classifier that is used for note discretization is not an usual knn classifier. Actually, it works in the following way: every note duration that has to be made discrete is first compared with the learning examples such that the differences of the duration the note and duration of all notes in the learning set are calculated. Our learning set has 100 examples for each note type. Next, all examples for which the absolute value of the mentioned difference is minimal and mutually equal are chosen. After that, the note is classified in the class that is most frequent among the chosen notes. For example, let us classify a note that has duration of We calculate the differences of that duration and durations of all notes in the learning set. Then we observe the absolute values of calculated differences. Let us assume that notes that correspond to the minimal absolute values of differences are from set of classes {6, 6, 6, 6, 5, 7}. Since the class 6 is the most frequent in the set of closest classes, the note is classified in the class 6, which represents the dotted quarter note. C. Learning the note classifier For classifying the notes we used a Hidden Markov Model based method, as it has been described in the third section. In the next subsection we will explain methods for learning the classifier and then we will describe a method of classification of a new example. The learning processes in cases of discrete and continuous note durations are similar. In both cases we use a Maximum Likelihood criterion. Based on that criterion, we want to determine the parameters of the Hidden Markov Model in a such a way that the generating probability of learning examples for that model will be maximal. Unfortunately, the solution of this maximization problem can not be found in closed form. Therefore, we need to use iterative methods for finding the solution. This can be done in various ways, for example with the Baum-Welch algorithm or with the gradient descent optimization, as is explained in [5]. Instead of Maximum Likelihood criterion, it is possible to use Maximum Mutual Information criterion, for which the gradient descent optimization methods are also required [5]. We learn our classifier with the Maximum Likelihood criterion because the method for iterative maximization of this criterion is already implemented in a Hidden Markov Model toolbox for Matlab software 1 The learning algorithm is stopped if it converges or if it exceeds the maximum allowed number of iterations, which, in our case, was 60 iterations. While learning, we record the log-likelihood in each iteration and show how it grows until it reaches its maximum. The plot of the growth of the log-likelihood is shown in figure 3. We have trained special HMMs for each dance, i.e. each HMM generates the rhythmic structure of the dance it represents with the maximum likelihood. In case of the continuous note durations, the output probabilities of each state of the model are represented with the Gaussian distribution with parameters µ i and σi 2, i.e. with the mean and the variance. The transition probabilities, priors and the Gaussian distribution parameters are determined with the learning algorithm using training examples. The number of states of each HMM is determined with the 3-fold cross-validation using 60 examples. We have determined that HMMs that represent tango, polka, cha-cha-cha and march should have 3 states. Hidden Markov Model that represents mazurka should have four states and HMM that represents waltz should have five states. The interpretation of parameters µ i and σi 2 is obvious. They determine the mean value of note s duration and the 1 murphyk/software/hmm/hmm.html

4 We classify the given example as tango, since the likelihood that the example is tango is maximal. In case of discrete note durations, we first have to discretize the example and then classify it. The classification procedure is the same as in case with continuous note durations, with the exception that the output symbols of each state of HMM are discrete-valued so we do not assume a theoretical distribution that would generate the output examples. IV. RESULTS Figure 3. algorithm The growth of the log-likelihood in iterations of the learning variance around the mean. Interpretation of other parameters, such as number of states and transition probabilities is not so intuitive. The probability that a note is first in the rhythmic structure can be interpreted with the prior. We can interpret the number of states of HMM as a number of different notes in a rhythmic structure. The transition probabilities between states can represent the probabilities that a certain note will appear after another note in a rhythmic structure. For example, a 37 represents in this interpretation a probability that half note will appear after a quaver. In this example, we used indices 3 = quaver i 7 = halfnote. D. Classification of dances After learning the HMMs for each dance, the classification of a new example is simple and intuitive. For each HMM we calculate the likelihood that the model will generate the given example. We then classify the example into a dance category for which the calculated likelihood is maximal. We calculate the likelihood of generating the given example with the forward algorithm described in [2]. If likelihoods for generating example are same for all HMMs, the example will not be classified. Let us show that on an example. Let the rhythmic structure we want to classify be given with X = [ ] Likelihoods of the HMMs for the given example are as following: P (X = tango) = P (X = polka) = P (X = mazurka) = P (X = waltz) = P (X = cha-cha-cha) = P (X = march) = A. Data set Our system discriminates six different dances: tango, polka, mazurka, waltz, cha-cha-cha and march, but it is easy to add more dance types. Based on rhythmic structures that are available on Wikipedia and that are shown in figure 2, we generated the learning examples in two ways: synthetically from the prototypes and by playing the rhythms on a keyboard with the MIDI interface. The synthetic generation of examples was done in the following way: we assumed that the duration of the quarter note is 120 ticks. Based on that assumption we calculated the durations of other notes and generated the prototypes of the rhythmic structures of dances according to rhythms displayed in figure 2. We then added a Gaussian noise to the prototypes in order to get synthetic examples. The mean and variance of the Gaussian noise were randomly changed in order to get the most heterogeneous examples. Beside synthetically generated examples, we played the rhythms displayed in figure 2 on a keyboard with a MIDI interface which can be used to load the played notes into the computer. For each dance we played 70 examples that were used exclusively for learning and validation, whilst for crossvalidation we used additional 30 examples. As examples were really played, they represent a real situation where the note duration may not obey the Gaussian distribution, as it was assumed while synthetically generating the examples. This will also show whether the assumption that the duration of notes obey the Gaussian distribution was correct. B. Classification results The classifier was tested in various ways. First we used the synthetically generated examples with continuous note durations. We generated 50 examples for learning and 100 examples for testing in a way that has been described before. The parameters of the additive Gaussian noise were the following: the variance of each example was randomly selected from interval [0, 2] and the mean from interval [ 5, 5]. This means that for each example we first generated parameters of the additive Gaussian noise, then we generated the noise and finally added the noise to the dance rhythmic structure prototypes. The results of the first experiment are given in table I. The rows of the table represent the dance which was the decision of the classifier, and the columns represent the real dance type.

5 Table I CONFUSION MATRIX FOR CLASSIFIER WHICH USES SYNTHETICALLY GENERATED EXAMPLES AND CONTINUOUS-VALUED DURATION OF NOTES Tango Polka Mazurka Waltz Cha-cha-cha March The accuracy of the classification is 93.33%, and precision and recall are 80%. The F1 micro and F1 macro values are as follows: 1 = 80% 1 = 74.61% We can notice that the classifier is bad in discrimination of polka and march. Great similarity between those dances is the main reason for such behaviour if you see the rhythmic structures of those dances on figure 2, you can see very similar quaver and semiquaver patterns. Semiquavers and quavers are very short notes so it is very difficult discriminating them in this case, especially if examples have big variance and deviations of note duration means. Poorer discrimination of waltz from cha-cha-cha is a consequence of noise in examples. We can see that the classifier is much more accurate and precise than just randomly picking a dance. Namely, the accuracy of a random pick would, on average, be equal to the probability of randomly picking the correct dance. As the system discriminates six dances, the accuracy of the random choice would be %. If we make the examples from previous experiment discrete, as we described earlier, and if we then use our classifier in a discrete domain, the results become even better (see the table II). The classification accuracy has increased to 95.78%, and precision and recall to 87.33%. The F1 micro and F1 macro values are as follows: 1 = 87.33% 1 = 85.57% We can now notice that the classifier better discriminates polka from march, but still the most of the polka examples are misclassified, what can of course be explained with the great similarity of theoretical rhythmic structures of polka and march. In a realistic situation few musical pieces have the real theoretical rhythmic structure. Therefore, we tested our classifier with the examples we played by ourselves on the keyboard with the MIDI interface. In the following experiments, we used 25 examples of each dance for learning the classifier and 45 examples for testing the classification. We first carried out an experiment with continuous-valued note durations. The results of the experiment are shown in table III. The classification accuracy is 96.91%, and the precision and recall are 90, 74%. The F1 micro and F1 macro values are as follows: 1 = 90.74% 1 = 89.97% We can now notice that the classifier better discriminates waltz from cha-cha-cha, which is the consequence of better learning and testing examples. If we use 30 examples for learning, instead of 25, the precision and accuracy of classification raise to 100%. When carrying out the last experiment, we used the same examples as in previous experiment, but before classification we discretized them and used the classifier in a discrete domain. The results of the experiment are shown in table IV. The classification accuracy has raised to 99.32%, and precision to 98.50%. The F1 micro and F1 macro values are as follows: 1 = 97.95% 1 = 97.95% The recall is 97.41% because not all examples have been classified, i.e. the classifier has refused to classify one example of tango and two examples of waltz so those examples were used as false negatives when calculating the F values and recall. This lowered the recall of the classification. We can notice that in last experiment the discrimination rate between polka and march has raised a lot. This tells us that this type of classification is the best for general use. The refusal of classification is often regarded as better than misclassification as this enables people to manually classify the examples that classifier refused to classify. V. CONCLUSION In this paper we described a system which is able to recognize dances based on MIDI recordings of the pieces. This enables the enthusiasts that do not understand the musical notation to recognize their favourite dances. This is an interesting problem because MIDI recordings are usually played by humans, so it is not possible to determine the type of the note or rhythmic structure with the full certainty. We made a classifier which can classify the examples with both discrete- and continuous-valued note durations. The classifier is based on Hidden Markov Models. The discretization of notes was made with the modified knn classifier. The

6 Table II CONFUSION MATRIX FOR CLASSIFIER WHICH USES SYNTHETICALLY GENERATED EXAMPLES AND DISCRETE-VALUED NOTE DURATIONS Tango Polka Mazurka Waltz Cha-cha-cha March Table III CONFUSION MATRIX FOR CLASSIFIER WHICH USES HUMAN-PLAYED EXAMPLES AND CONTINUOUS-VALUED NOTE DURATIONS Tango Polka Mazurka Waltz Cha-cha-cha March Table IV CONFUSION MATRIX FOR CLASSIFIER WHICH USES HUMAN-PLAYED EXAMPLES AND DISCRETE-VALUED NOTE DURATIONS Tango Polka Mazurka Waltz Cha-cha-cha March classifier was trained with synthetically generated and humanplayed examples, with both continuous- and discrete-valued note durations. The best classification rates were achieved with human-played examples with discrete-valued note durations. We have accomplished everything planned, although future expansions of the system are possible, For example, it would be possible to create a classifier which would automatically find the characteristic rhythmic structure of a piece during the training phase. With a greater number of examples, such a classifier would be better at learning specific music pieces, in contrast to our work where we used theoretical rhytmic structure. [4] M. Hamanaka, M. Goto, H. Asoh, and N. Otsu, A learning-based quantization: Estimation of onset times in a musical score, in Proceedings of the 5th World Multi-conference on Systemics, Cybernetics and Informatics (SCI 2001, vol. 10, 2001, pp [5] N. Warakagoda, A Hybrid ANN-HMM ASR System with NN-based Adaptive Preprocessing. Institutt for teleteknikk, NTH, REFERENCES [1] H. Takeda, N. Saito, T. Otsuki, M. Nakai, H. Shimodaira, and S. Sagayama, Hidden Markov model for automatic transcription of MIDI signals, in Multimedia Signal Processing, 2002 IEEE Workshop on. IEEE, 2003, pp [2] L. Rabiner, A tutorial on hidden Markov models and selected applications in speech recognition, Proceedings of the IEEE, vol. 77, no. 2, pp , [3] E. Cambouropoulos, From MIDI to traditional musical notation, in Proceedings of the AAAI Workshop on Artificial Intelligence and Music: Towards Formal Models for Composition, Performance and Analysis, vol. 30, 2000.

Supervised Learning in Genre Classification

Supervised Learning in Genre Classification Supervised Learning in Genre Classification Introduction & Motivation Mohit Rajani and Luke Ekkizogloy {i.mohit,luke.ekkizogloy}@gmail.com Stanford University, CS229: Machine Learning, 2009 Now that music

More information

Automatic Labelling of tabla signals

Automatic Labelling of tabla signals ISMIR 2003 Oct. 27th 30th 2003 Baltimore (USA) Automatic Labelling of tabla signals Olivier K. GILLET, Gaël RICHARD Introduction Exponential growth of available digital information need for Indexing and

More information

Detecting Musical Key with Supervised Learning

Detecting Musical Key with Supervised Learning Detecting Musical Key with Supervised Learning Robert Mahieu Department of Electrical Engineering Stanford University rmahieu@stanford.edu Abstract This paper proposes and tests performance of two different

More information

Outline. Why do we classify? Audio Classification

Outline. Why do we classify? Audio Classification Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification Implementation Future Work Why do we classify

More information

Composer Style Attribution

Composer Style Attribution Composer Style Attribution Jacqueline Speiser, Vishesh Gupta Introduction Josquin des Prez (1450 1521) is one of the most famous composers of the Renaissance. Despite his fame, there exists a significant

More information

Automatic Rhythmic Notation from Single Voice Audio Sources

Automatic Rhythmic Notation from Single Voice Audio Sources Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung

More information

A Bayesian Network for Real-Time Musical Accompaniment

A Bayesian Network for Real-Time Musical Accompaniment A Bayesian Network for Real-Time Musical Accompaniment Christopher Raphael Department of Mathematics and Statistics, University of Massachusetts at Amherst, Amherst, MA 01003-4515, raphael~math.umass.edu

More information

CS229 Project Report Polyphonic Piano Transcription

CS229 Project Report Polyphonic Piano Transcription CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project

More information

Chord Classification of an Audio Signal using Artificial Neural Network

Chord Classification of an Audio Signal using Artificial Neural Network Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------

More information

AUTOMATIC ACCOMPANIMENT OF VOCAL MELODIES IN THE CONTEXT OF POPULAR MUSIC

AUTOMATIC ACCOMPANIMENT OF VOCAL MELODIES IN THE CONTEXT OF POPULAR MUSIC AUTOMATIC ACCOMPANIMENT OF VOCAL MELODIES IN THE CONTEXT OF POPULAR MUSIC A Thesis Presented to The Academic Faculty by Xiang Cao In Partial Fulfillment of the Requirements for the Degree Master of Science

More information

A System for Automatic Chord Transcription from Audio Using Genre-Specific Hidden Markov Models

A System for Automatic Chord Transcription from Audio Using Genre-Specific Hidden Markov Models A System for Automatic Chord Transcription from Audio Using Genre-Specific Hidden Markov Models Kyogu Lee Center for Computer Research in Music and Acoustics Stanford University, Stanford CA 94305, USA

More information

MUSI-6201 Computational Music Analysis

MUSI-6201 Computational Music Analysis MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)

More information

Music Alignment and Applications. Introduction

Music Alignment and Applications. Introduction Music Alignment and Applications Roger B. Dannenberg Schools of Computer Science, Art, and Music Introduction Music information comes in many forms Digital Audio Multi-track Audio Music Notation MIDI Structured

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM A QUER B EAMPLE MUSIC RETRIEVAL ALGORITHM H. HARB AND L. CHEN Maths-Info department, Ecole Centrale de Lyon. 36, av. Guy de Collongue, 69134, Ecully, France, EUROPE E-mail: {hadi.harb, liming.chen}@ec-lyon.fr

More information

Jazz Melody Generation and Recognition

Jazz Melody Generation and Recognition Jazz Melody Generation and Recognition Joseph Victor December 14, 2012 Introduction In this project, we attempt to use machine learning methods to study jazz solos. The reason we study jazz in particular

More information

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Aric Bartle (abartle@stanford.edu) December 14, 2012 1 Background The field of composer recognition has

More information

Transcription of the Singing Melody in Polyphonic Music

Transcription of the Singing Melody in Polyphonic Music Transcription of the Singing Melody in Polyphonic Music Matti Ryynänen and Anssi Klapuri Institute of Signal Processing, Tampere University Of Technology P.O.Box 553, FI-33101 Tampere, Finland {matti.ryynanen,

More information

Music Composition with RNN

Music Composition with RNN Music Composition with RNN Jason Wang Department of Statistics Stanford University zwang01@stanford.edu Abstract Music composition is an interesting problem that tests the creativity capacities of artificial

More information

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? NICHOLAS BORG AND GEORGE HOKKANEN Abstract. The possibility of a hit song prediction algorithm is both academically interesting and industry motivated.

More information

Robert Alexandru Dobre, Cristian Negrescu

Robert Alexandru Dobre, Cristian Negrescu ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Automatic Music Transcription Software Based on Constant Q

More information

Available online at ScienceDirect. Procedia Computer Science 46 (2015 )

Available online at  ScienceDirect. Procedia Computer Science 46 (2015 ) Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 46 (2015 ) 381 387 International Conference on Information and Communication Technologies (ICICT 2014) Music Information

More information

Music Segmentation Using Markov Chain Methods

Music Segmentation Using Markov Chain Methods Music Segmentation Using Markov Chain Methods Paul Finkelstein March 8, 2011 Abstract This paper will present just how far the use of Markov Chains has spread in the 21 st century. We will explain some

More information

Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment

Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment Gus G. Xia Dartmouth College Neukom Institute Hanover, NH, USA gxia@dartmouth.edu Roger B. Dannenberg Carnegie

More information

Music Representations

Music Representations Advanced Course Computer Science Music Processing Summer Term 00 Music Representations Meinard Müller Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Music Representations Music Representations

More information

Algorithmic Music Composition

Algorithmic Music Composition Algorithmic Music Composition MUS-15 Jan Dreier July 6, 2015 1 Introduction The goal of algorithmic music composition is to automate the process of creating music. One wants to create pleasant music without

More information

Automatic Piano Music Transcription

Automatic Piano Music Transcription Automatic Piano Music Transcription Jianyu Fan Qiuhan Wang Xin Li Jianyu.Fan.Gr@dartmouth.edu Qiuhan.Wang.Gr@dartmouth.edu Xi.Li.Gr@dartmouth.edu 1. Introduction Writing down the score while listening

More information

Melodic Outline Extraction Method for Non-note-level Melody Editing

Melodic Outline Extraction Method for Non-note-level Melody Editing Melodic Outline Extraction Method for Non-note-level Melody Editing Yuichi Tsuchiya Nihon University tsuchiya@kthrlab.jp Tetsuro Kitahara Nihon University kitahara@kthrlab.jp ABSTRACT In this paper, we

More information

Music Radar: A Web-based Query by Humming System

Music Radar: A Web-based Query by Humming System Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,

More information

Topic 10. Multi-pitch Analysis

Topic 10. Multi-pitch Analysis Topic 10 Multi-pitch Analysis What is pitch? Common elements of music are pitch, rhythm, dynamics, and the sonic qualities of timbre and texture. An auditory perceptual attribute in terms of which sounds

More information

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You Chris Lewis Stanford University cmslewis@stanford.edu Abstract In this project, I explore the effectiveness of the Naive Bayes Classifier

More information

... A Pseudo-Statistical Approach to Commercial Boundary Detection. Prasanna V Rangarajan Dept of Electrical Engineering Columbia University

... A Pseudo-Statistical Approach to Commercial Boundary Detection. Prasanna V Rangarajan Dept of Electrical Engineering Columbia University A Pseudo-Statistical Approach to Commercial Boundary Detection........ Prasanna V Rangarajan Dept of Electrical Engineering Columbia University pvr2001@columbia.edu 1. Introduction Searching and browsing

More information

BayesianBand: Jam Session System based on Mutual Prediction by User and System

BayesianBand: Jam Session System based on Mutual Prediction by User and System BayesianBand: Jam Session System based on Mutual Prediction by User and System Tetsuro Kitahara 12, Naoyuki Totani 1, Ryosuke Tokuami 1, and Haruhiro Katayose 12 1 School of Science and Technology, Kwansei

More information

Music Representations

Music Representations Lecture Music Processing Music Representations Meinard Müller International Audio Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de Book: Fundamentals of Music Processing Meinard Müller Fundamentals

More information

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION ULAŞ BAĞCI AND ENGIN ERZIN arxiv:0907.3220v1 [cs.sd] 18 Jul 2009 ABSTRACT. Music genre classification is an essential tool for

More information

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC Scientific Journal of Impact Factor (SJIF): 5.71 International Journal of Advance Engineering and Research Development Volume 5, Issue 04, April -2018 e-issn (O): 2348-4470 p-issn (P): 2348-6406 MUSICAL

More information

Week 14 Music Understanding and Classification

Week 14 Music Understanding and Classification Week 14 Music Understanding and Classification Roger B. Dannenberg Professor of Computer Science, Music & Art Overview n Music Style Classification n What s a classifier? n Naïve Bayesian Classifiers n

More information

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon A Study of Synchronization of Audio Data with Symbolic Data Music254 Project Report Spring 2007 SongHui Chon Abstract This paper provides an overview of the problem of audio and symbolic synchronization.

More information

A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS

A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS Mutian Fu 1 Guangyu Xia 2 Roger Dannenberg 2 Larry Wasserman 2 1 School of Music, Carnegie Mellon University, USA 2 School of Computer

More information

Building a Better Bach with Markov Chains

Building a Better Bach with Markov Chains Building a Better Bach with Markov Chains CS701 Implementation Project, Timothy Crocker December 18, 2015 1 Abstract For my implementation project, I explored the field of algorithmic music composition

More information

Subjective Similarity of Music: Data Collection for Individuality Analysis

Subjective Similarity of Music: Data Collection for Individuality Analysis Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp

More information

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 AN HMM BASED INVESTIGATION OF DIFFERENCES BETWEEN MUSICAL INSTRUMENTS OF THE SAME TYPE PACS: 43.75.-z Eichner, Matthias; Wolff, Matthias;

More information

The Human Features of Music.

The Human Features of Music. The Human Features of Music. Bachelor Thesis Artificial Intelligence, Social Studies, Radboud University Nijmegen Chris Kemper, s4359410 Supervisor: Makiko Sadakata Artificial Intelligence, Social Studies,

More information

Artificially intelligent accompaniment using Hidden Markov Models to model musical structure

Artificially intelligent accompaniment using Hidden Markov Models to model musical structure Artificially intelligent accompaniment using Hidden Markov Models to model musical structure Anna Jordanous Music Informatics, Department of Informatics, University of Sussex, UK a.k.jordanous at sussex.ac.uk

More information

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval DAY 1 Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval Jay LeBoeuf Imagine Research jay{at}imagine-research.com Rebecca

More information

Music Genre Classification and Variance Comparison on Number of Genres

Music Genre Classification and Variance Comparison on Number of Genres Music Genre Classification and Variance Comparison on Number of Genres Miguel Francisco, miguelf@stanford.edu Dong Myung Kim, dmk8265@stanford.edu 1 Abstract In this project we apply machine learning techniques

More information

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES 12th International Society for Music Information Retrieval Conference (ISMIR 2011) A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES Erdem Unal 1 Elaine Chew 2 Panayiotis Georgiou

More information

Singer Traits Identification using Deep Neural Network

Singer Traits Identification using Deep Neural Network Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic

More information

A NOVEL HMM APPROACH TO MELODY SPOTTING IN RAW AUDIO RECORDINGS

A NOVEL HMM APPROACH TO MELODY SPOTTING IN RAW AUDIO RECORDINGS A NOVEL HMM APPROACH TO MELODY SPOTTING IN RAW AUDIO RECORDINGS Aggelos Pikrakis and Sergios Theodoridis Dept. of Informatics and Telecommunications University of Athens Panepistimioupolis, TYPA Buildings

More information

PLANE TESSELATION WITH MUSICAL-SCALE TILES AND BIDIMENSIONAL AUTOMATIC COMPOSITION

PLANE TESSELATION WITH MUSICAL-SCALE TILES AND BIDIMENSIONAL AUTOMATIC COMPOSITION PLANE TESSELATION WITH MUSICAL-SCALE TILES AND BIDIMENSIONAL AUTOMATIC COMPOSITION ABSTRACT We present a method for arranging the notes of certain musical scales (pentatonic, heptatonic, Blues Minor and

More information

THE importance of music content analysis for musical

THE importance of music content analysis for musical IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 15, NO. 1, JANUARY 2007 333 Drum Sound Recognition for Polyphonic Audio Signals by Adaptation and Matching of Spectrogram Templates With

More information

Phone-based Plosive Detection

Phone-based Plosive Detection Phone-based Plosive Detection 1 Andreas Madsack, Grzegorz Dogil, Stefan Uhlich, Yugu Zeng and Bin Yang Abstract We compare two segmentation approaches to plosive detection: One aproach is using a uniform

More information

Automatic LP Digitalization Spring Group 6: Michael Sibley, Alexander Su, Daphne Tsatsoulis {msibley, ahs1,

Automatic LP Digitalization Spring Group 6: Michael Sibley, Alexander Su, Daphne Tsatsoulis {msibley, ahs1, Automatic LP Digitalization 18-551 Spring 2011 Group 6: Michael Sibley, Alexander Su, Daphne Tsatsoulis {msibley, ahs1, ptsatsou}@andrew.cmu.edu Introduction This project was originated from our interest

More information

6.UAP Project. FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System. Daryl Neubieser. May 12, 2016

6.UAP Project. FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System. Daryl Neubieser. May 12, 2016 6.UAP Project FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System Daryl Neubieser May 12, 2016 Abstract: This paper describes my implementation of a variable-speed accompaniment system that

More information

Polyphonic Audio Matching for Score Following and Intelligent Audio Editors

Polyphonic Audio Matching for Score Following and Intelligent Audio Editors Polyphonic Audio Matching for Score Following and Intelligent Audio Editors Roger B. Dannenberg and Ning Hu School of Computer Science, Carnegie Mellon University email: dannenberg@cs.cmu.edu, ninghu@cs.cmu.edu,

More information

A Learning-Based Jam Session System that Imitates a Player's Personality Model

A Learning-Based Jam Session System that Imitates a Player's Personality Model A Learning-Based Jam Session System that Imitates a Player's Personality Model Masatoshi Hamanaka 12, Masataka Goto 3) 2), Hideki Asoh 2) 2) 4), and Nobuyuki Otsu 1) Research Fellow of the Japan Society

More information

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur Module 8 VIDEO CODING STANDARDS Lesson 27 H.264 standard Lesson Objectives At the end of this lesson, the students should be able to: 1. State the broad objectives of the H.264 standard. 2. List the improved

More information

Research Article. ISSN (Print) *Corresponding author Shireen Fathima

Research Article. ISSN (Print) *Corresponding author Shireen Fathima Scholars Journal of Engineering and Technology (SJET) Sch. J. Eng. Tech., 2014; 2(4C):613-620 Scholars Academic and Scientific Publisher (An International Publisher for Academic and Scientific Resources)

More information

An Empirical Comparison of Tempo Trackers

An Empirical Comparison of Tempo Trackers An Empirical Comparison of Tempo Trackers Simon Dixon Austrian Research Institute for Artificial Intelligence Schottengasse 3, A-1010 Vienna, Austria simon@oefai.at An Empirical Comparison of Tempo Trackers

More information

Chord Representations for Probabilistic Models

Chord Representations for Probabilistic Models R E S E A R C H R E P O R T I D I A P Chord Representations for Probabilistic Models Jean-François Paiement a Douglas Eck b Samy Bengio a IDIAP RR 05-58 September 2005 soumis à publication a b IDIAP Research

More information

UTILITY SYSTEM FOR CONSTRUCTING DATABASE OF PERFORMANCE DEVIATIONS

UTILITY SYSTEM FOR CONSTRUCTING DATABASE OF PERFORMANCE DEVIATIONS UTILITY SYSTEM FOR CONSTRUCTING DATABASE OF PERFORMANCE DEVIATIONS Ken ichi Toyoda, Kenzi Noike, Haruhiro Katayose Kwansei Gakuin University Gakuen, Sanda, 669-1337 JAPAN {toyoda, noike, katayose}@ksc.kwansei.ac.jp

More information

QUALITY OF COMPUTER MUSIC USING MIDI LANGUAGE FOR DIGITAL MUSIC ARRANGEMENT

QUALITY OF COMPUTER MUSIC USING MIDI LANGUAGE FOR DIGITAL MUSIC ARRANGEMENT QUALITY OF COMPUTER MUSIC USING MIDI LANGUAGE FOR DIGITAL MUSIC ARRANGEMENT Pandan Pareanom Purwacandra 1, Ferry Wahyu Wibowo 2 Informatics Engineering, STMIK AMIKOM Yogyakarta 1 pandanharmony@gmail.com,

More information

Singer Recognition and Modeling Singer Error

Singer Recognition and Modeling Singer Error Singer Recognition and Modeling Singer Error Johan Ismael Stanford University jismael@stanford.edu Nicholas McGee Stanford University ndmcgee@stanford.edu 1. Abstract We propose a system for recognizing

More information

Machine Learning of Expressive Microtiming in Brazilian and Reggae Drumming Matt Wright (Music) and Edgar Berdahl (EE), CS229, 16 December 2005

Machine Learning of Expressive Microtiming in Brazilian and Reggae Drumming Matt Wright (Music) and Edgar Berdahl (EE), CS229, 16 December 2005 Machine Learning of Expressive Microtiming in Brazilian and Reggae Drumming Matt Wright (Music) and Edgar Berdahl (EE), CS229, 16 December 2005 Abstract We have used supervised machine learning to apply

More information

Lyrics Classification using Naive Bayes

Lyrics Classification using Naive Bayes Lyrics Classification using Naive Bayes Dalibor Bužić *, Jasminka Dobša ** * College for Information Technologies, Klaićeva 7, Zagreb, Croatia ** Faculty of Organization and Informatics, Pavlinska 2, Varaždin,

More information

Enhancing Music Maps

Enhancing Music Maps Enhancing Music Maps Jakob Frank Vienna University of Technology, Vienna, Austria http://www.ifs.tuwien.ac.at/mir frank@ifs.tuwien.ac.at Abstract. Private as well as commercial music collections keep growing

More information

2. AN INTROSPECTION OF THE MORPHING PROCESS

2. AN INTROSPECTION OF THE MORPHING PROCESS 1. INTRODUCTION Voice morphing means the transition of one speech signal into another. Like image morphing, speech morphing aims to preserve the shared characteristics of the starting and final signals,

More information

A prototype system for rule-based expressive modifications of audio recordings

A prototype system for rule-based expressive modifications of audio recordings International Symposium on Performance Science ISBN 0-00-000000-0 / 000-0-00-000000-0 The Author 2007, Published by the AEC All rights reserved A prototype system for rule-based expressive modifications

More information

Extracting Significant Patterns from Musical Strings: Some Interesting Problems.

Extracting Significant Patterns from Musical Strings: Some Interesting Problems. Extracting Significant Patterns from Musical Strings: Some Interesting Problems. Emilios Cambouropoulos Austrian Research Institute for Artificial Intelligence Vienna, Austria emilios@ai.univie.ac.at Abstract

More information

Creating a Feature Vector to Identify Similarity between MIDI Files

Creating a Feature Vector to Identify Similarity between MIDI Files Creating a Feature Vector to Identify Similarity between MIDI Files Joseph Stroud 2017 Honors Thesis Advised by Sergio Alvarez Computer Science Department, Boston College 1 Abstract Today there are many

More information

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the

More information

Neural Network Predicating Movie Box Office Performance

Neural Network Predicating Movie Box Office Performance Neural Network Predicating Movie Box Office Performance Alex Larson ECE 539 Fall 2013 Abstract The movie industry is a large part of modern day culture. With the rise of websites like Netflix, where people

More information

Adaptive decoding of convolutional codes

Adaptive decoding of convolutional codes Adv. Radio Sci., 5, 29 214, 27 www.adv-radio-sci.net/5/29/27/ Author(s) 27. This work is licensed under a Creative Commons License. Advances in Radio Science Adaptive decoding of convolutional codes K.

More information

Piano Transcription MUMT611 Presentation III 1 March, Hankinson, 1/15

Piano Transcription MUMT611 Presentation III 1 March, Hankinson, 1/15 Piano Transcription MUMT611 Presentation III 1 March, 2007 Hankinson, 1/15 Outline Introduction Techniques Comb Filtering & Autocorrelation HMMs Blackboard Systems & Fuzzy Logic Neural Networks Examples

More information

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis Fengyan Wu fengyanyy@163.com Shutao Sun stsun@cuc.edu.cn Weiyao Xue Wyxue_std@163.com Abstract Automatic extraction of

More information

Machine Learning Term Project Write-up Creating Models of Performers of Chopin Mazurkas

Machine Learning Term Project Write-up Creating Models of Performers of Chopin Mazurkas Machine Learning Term Project Write-up Creating Models of Performers of Chopin Mazurkas Marcello Herreshoff In collaboration with Craig Sapp (craig@ccrma.stanford.edu) 1 Motivation We want to generative

More information

Brain-Computer Interface (BCI)

Brain-Computer Interface (BCI) Brain-Computer Interface (BCI) Christoph Guger, Günter Edlinger, g.tec Guger Technologies OEG Herbersteinstr. 60, 8020 Graz, Austria, guger@gtec.at This tutorial shows HOW-TO find and extract proper signal

More information

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University Week 14 Query-by-Humming and Music Fingerprinting Roger B. Dannenberg Professor of Computer Science, Art and Music Overview n Melody-Based Retrieval n Audio-Score Alignment n Music Fingerprinting 2 Metadata-based

More information

CPU Bach: An Automatic Chorale Harmonization System

CPU Bach: An Automatic Chorale Harmonization System CPU Bach: An Automatic Chorale Harmonization System Matt Hanlon mhanlon@fas Tim Ledlie ledlie@fas January 15, 2002 Abstract We present an automated system for the harmonization of fourpart chorales in

More information

TOWARDS CHARACTERISATION OF MUSIC VIA RHYTHMIC PATTERNS

TOWARDS CHARACTERISATION OF MUSIC VIA RHYTHMIC PATTERNS TOWARDS CHARACTERISATION OF MUSIC VIA RHYTHMIC PATTERNS Simon Dixon Austrian Research Institute for AI Vienna, Austria Fabien Gouyon Universitat Pompeu Fabra Barcelona, Spain Gerhard Widmer Medical University

More information

Analysis of local and global timing and pitch change in ordinary

Analysis of local and global timing and pitch change in ordinary Alma Mater Studiorum University of Bologna, August -6 6 Analysis of local and global timing and pitch change in ordinary melodies Roger Watt Dept. of Psychology, University of Stirling, Scotland r.j.watt@stirling.ac.uk

More information

LSTM Neural Style Transfer in Music Using Computational Musicology

LSTM Neural Style Transfer in Music Using Computational Musicology LSTM Neural Style Transfer in Music Using Computational Musicology Jett Oristaglio Dartmouth College, June 4 2017 1. Introduction In the 2016 paper A Neural Algorithm of Artistic Style, Gatys et al. discovered

More information

Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals

Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals Eita Nakamura and Shinji Takaki National Institute of Informatics, Tokyo 101-8430, Japan eita.nakamura@gmail.com, takaki@nii.ac.jp

More information

CLASSIFICATION OF MUSICAL METRE WITH AUTOCORRELATION AND DISCRIMINANT FUNCTIONS

CLASSIFICATION OF MUSICAL METRE WITH AUTOCORRELATION AND DISCRIMINANT FUNCTIONS CLASSIFICATION OF MUSICAL METRE WITH AUTOCORRELATION AND DISCRIMINANT FUNCTIONS Petri Toiviainen Department of Music University of Jyväskylä Finland ptoiviai@campus.jyu.fi Tuomas Eerola Department of Music

More information

Detection of Panoramic Takes in Soccer Videos Using Phase Correlation and Boosting

Detection of Panoramic Takes in Soccer Videos Using Phase Correlation and Boosting Detection of Panoramic Takes in Soccer Videos Using Phase Correlation and Boosting Luiz G. L. B. M. de Vasconcelos Research & Development Department Globo TV Network Email: luiz.vasconcelos@tvglobo.com.br

More information

A Beat Tracking System for Audio Signals

A Beat Tracking System for Audio Signals A Beat Tracking System for Audio Signals Simon Dixon Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria. simon@ai.univie.ac.at April 7, 2000 Abstract We present

More information

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS Andrew N. Robertson, Mark D. Plumbley Centre for Digital Music

More information

Rhythm together with melody is one of the basic elements in music. According to Longuet-Higgins

Rhythm together with melody is one of the basic elements in music. According to Longuet-Higgins 5 Quantisation Rhythm together with melody is one of the basic elements in music. According to Longuet-Higgins ([LH76]) human listeners are much more sensitive to the perception of rhythm than to the perception

More information

Evaluating Melodic Encodings for Use in Cover Song Identification

Evaluating Melodic Encodings for Use in Cover Song Identification Evaluating Melodic Encodings for Use in Cover Song Identification David D. Wickland wickland@uoguelph.ca David A. Calvert dcalvert@uoguelph.ca James Harley jharley@uoguelph.ca ABSTRACT Cover song identification

More information

TRACKING THE ODD : METER INFERENCE IN A CULTURALLY DIVERSE MUSIC CORPUS

TRACKING THE ODD : METER INFERENCE IN A CULTURALLY DIVERSE MUSIC CORPUS TRACKING THE ODD : METER INFERENCE IN A CULTURALLY DIVERSE MUSIC CORPUS Andre Holzapfel New York University Abu Dhabi andre@rhythmos.org Florian Krebs Johannes Kepler University Florian.Krebs@jku.at Ajay

More information

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods Kazuyoshi Yoshii, Masataka Goto and Hiroshi G. Okuno Department of Intelligence Science and Technology National

More information

The dangers of parsimony in query-by-humming applications

The dangers of parsimony in query-by-humming applications The dangers of parsimony in query-by-humming applications Colin Meek University of Michigan Beal Avenue Ann Arbor MI 489 USA meek@umich.edu William P. Birmingham University of Michigan Beal Avenue Ann

More information

Implementation of a turbo codes test bed in the Simulink environment

Implementation of a turbo codes test bed in the Simulink environment University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 2005 Implementation of a turbo codes test bed in the Simulink environment

More information

Sudhanshu Gautam *1, Sarita Soni 2. M-Tech Computer Science, BBAU Central University, Lucknow, Uttar Pradesh, India

Sudhanshu Gautam *1, Sarita Soni 2. M-Tech Computer Science, BBAU Central University, Lucknow, Uttar Pradesh, India International Journal of Scientific Research in Computer Science, Engineering and Information Technology 2018 IJSRCSEIT Volume 3 Issue 3 ISSN : 2456-3307 Artificial Intelligence Techniques for Music Composition

More information

SMART VEHICLE SCREENING SYSTEM USING ARTIFICIAL INTELLIGENCE METHODS

SMART VEHICLE SCREENING SYSTEM USING ARTIFICIAL INTELLIGENCE METHODS 1 TERNOPIL ACADEMY OF NATIONAL ECONOMY INSTITUTE OF COMPUTER INFORMATION TECHNOLOGIES SMART VEHICLE SCREENING SYSTEM USING ARTIFICIAL INTELLIGENCE METHODS Presenters: Volodymyr Turchenko Vasyl Koval The

More information

Analysis and Clustering of Musical Compositions using Melody-based Features

Analysis and Clustering of Musical Compositions using Melody-based Features Analysis and Clustering of Musical Compositions using Melody-based Features Isaac Caswell Erika Ji December 13, 2013 Abstract This paper demonstrates that melodic structure fundamentally differentiates

More information

Various Artificial Intelligence Techniques For Automated Melody Generation

Various Artificial Intelligence Techniques For Automated Melody Generation Various Artificial Intelligence Techniques For Automated Melody Generation Nikahat Kazi Computer Engineering Department, Thadomal Shahani Engineering College, Mumbai, India Shalini Bhatia Assistant Professor,

More information

Feature-Based Analysis of Haydn String Quartets

Feature-Based Analysis of Haydn String Quartets Feature-Based Analysis of Haydn String Quartets Lawson Wong 5/5/2 Introduction When listening to multi-movement works, amateur listeners have almost certainly asked the following situation : Am I still

More information

A New Method for Calculating Music Similarity

A New Method for Calculating Music Similarity A New Method for Calculating Music Similarity Eric Battenberg and Vijay Ullal December 12, 2006 Abstract We introduce a new technique for calculating the perceived similarity of two songs based on their

More information

Multiple instrument tracking based on reconstruction error, pitch continuity and instrument activity

Multiple instrument tracking based on reconstruction error, pitch continuity and instrument activity Multiple instrument tracking based on reconstruction error, pitch continuity and instrument activity Holger Kirchhoff 1, Simon Dixon 1, and Anssi Klapuri 2 1 Centre for Digital Music, Queen Mary University

More information