Exploring Melodic Features for the Classification and Retrieval of Traditional Music in the Context of Cultural Source

Size: px
Start display at page:

Download "Exploring Melodic Features for the Classification and Retrieval of Traditional Music in the Context of Cultural Source"

Transcription

1 Exploring Melodic Features for the Classification and Retrieval of Traditional Music in the Context of Cultural Source Jan Miles Co Ateneo de Manila University Quezon City, Philippines Andrei Coronel Ateneo de Manila University Quezon City, Philippines ABSTRACT The major change in the preference of music downloading over CD sales has created a major shift in the distribution of music. Music industries are focusing on online products and services over the sales of physical media formats. As the number of available musical recordings rapidly increase, websites face the problem of music organization. Music Information Retrieval (MIR) is a relatively new field with the main goal of developing systems that allow users to retrieve music with the use of music content as opposed to text-based searching. This study attempts to create an accurate MIR system that uses melodic features and similarity in terms of cultural source as a means for retrieving MIDI files. The experiments for the identification of feature sets for the MIR system have been conducted. The feature sets were implemented in a Proposed MIR System. The various feature sets are presented as results together with tests concerning the consistency and quality of the Proposed MIR System. Categories and Subject Descriptors H.3 [Information Storage and Retrieval]: Information Search and Retrieval retrieval models, search process, selection process; H.5 [Information Interfaces and Presentation]: Sound and Music Computing methodologies and techniques, modeling, Systems; General Terms Management, Performance, Design, Experimentation. Keywords Music, MIDI, Retrieval, Similarity. 1. INTRODUCTION Several research have been conducted in the representation of culture in music due to its impact in creation, performance and interpretation. Cultural background affects the preference of people for music: A group of people with the same cultural background often prefers a similar set of music. Therefore, apart from genre, culture can be used for music organization, especially in an era where most people have access to international music[1]. Music Information Retrieval (MIR) is a relatively new field with the main goal of developing systems that allow users to retrieve music with the use of music content as opposed to text-based search. A community-based framework called Music Information Retrieval Evaluation exchange (MIREX) was inaugurated in 2005 for the formal evaluation of MIR systems and algorithms. Activities in MIREX reflect the main concerns and specific interests of the community with regard to MIR research [3]. The tasks lists for MIREX 2005, 2006, and 2007 include: Audio Train/Test Tasks with the following subtasks: Audio Artist Identification, Audio Genre Classification, Audio Music Mood Classification, and Audio Classical Composer Identification; Symbolic Genre Classification; Audio Onset Detection; Audio Key Detection; Symbolic Key Detection; Audio Tag Classification; Audio Cover Song Identification; Real-time Audio to Score Alignment (a.k.a. Score Following); Query by Singing/Humming; Multiple Fundamental Frequency Estimation and Tracking; Audio Chord Estimation; Audio Melody Extraction; Query by Tapping; Audio Beat Tracking; Audio Music Similarity and Retrieval; Symbolic Melodic Similarity; Structural Segmentation; Audio Drum Detection; and Audio Tempo Extraction [3]. Of the 19 tasks listed, only 3 involve processing of symbolic files. Mainly because many researchers in MIR have backgrounds involved with signal processing such as electrical engineering, acoustics, speech processing, etc.[2]. This study attempts to create a new method for an accurate retrieval of symbolic files, where cultural source will be used as a similarity measure for the retrieval of songs. Section 2 gives a brief description of the techniques adapted from previous literature. Section 3 introduces the experimental flow. Section 4 presents results of the various experiments. Finally, Sections 5 and 6 concludes the study. 2. REVIEW OF RELATED LITERATURE [4] used audio signals to classify music according to cultural source. Timbre features, wavelet-based features, and musicologybased features were extracted from music recordings and were used to classify music into six cultural styles: Western classical music, Chinese traditional music, Japanese traditional music, Indian classical music, Arabic folk music, and African folk music. A dataset of 1300 music recordings, characterized by roughly the same number of recordings per category were used. 900 pieces were used for training and 400 where used for testing. Three supervised classifiers were used: decision tree, KNN, and SVM. An overall accuracy of 86% was achieved by using a combination of the features and SVM. For Western and Oriental (Chinese and Japanese) music, an accuracy of 94% was achieved. It was determined that timbre features are the most effective for cultural classification and can be slightly improved by using additional features. The research suggested the exploration of pair-wise style classification, which will be adapted in this study. [5] studied 21 musical features in an attempt to create a Genetic Algorithm (GA) fitness function. The 21 features reflect pitch, tonality, melodic contour, rhythm and repeated pattern or motifs of a melody. The 21 features were extracted from 36 melodies that were composed of 18 classical melodies, 10 pre-classical compositions, 6 traditional nursery rhymes, and 2 melodies from popular tunes. The features with high standard deviation were determined to be the features that can be used to determine melodic style. These features are pitch range, non-scale notes, step movement, note density, repeated duration, repeated pitch patterns of 3 notes, repeated pitch patterns of 4 notes, repeated duration patterns of 3 notes, and repeated patterns of 4 notes. The 21

2 features identified by [5] will be used in this study as a feature set for cultural classification. [6] mapped the 21 features used by [5] to 13 features available in jsymbolic, a java-based tool that extracts 160 features from MIDI files [7]. The 13 features were used for genre classification. Among the 160 features that jsymbolic can extract from MIDI files, [6] also identified 54 features that pertain to melodic features alone. These features were used to determine a subset of features that are useful for classification of music according to genre using C4.5 decision tree. In this study, the same technique for genre classification will be used for classification of traditional music according to cultural source. The 13 features of [5] mapped by [6] and the 54 melodic features identified by [6] will be used as a feature subset for cultural classification. The method of using C4.5 decision tree as a tool for feature selection will also be used to reduce feature subsets for cultural classification. [8] performed feature selection in order to improve automatic genre classification of traditional Malay music. 10 traditional Malay genres were used for classification: Dikir Barat, Etnik Sabah, Gamelan, Ghazal, Inang, Joget, Keroncong, Tumbuk Kalang, Wayang Kulit, and Zapin. A dataset of 191 songs were converted into WAV files and loaded into MARSYAS-0.2.2, a tool that extracts timbral texture, rhythmic content, and pitch content features. The number of features that were extracted is 73, which is composed of STFT, MFCC, and Beat Features. 7 feature selection methods were used: Correlation-based Feature Selection (CFS) with Best First Search, CFS with Genetic Search Strategy, CFS with Greedy Stepwise Search Strategy, Principal Component Analysis, Chi-square Feature Evaluation, Gain Ration Feature Evaluation, and SVM-Based Feature Evaluation. To test the effectiveness of the feature sets, 18 classification algorithms were used. The most effective feature selection algorithms for Naive Bayes are CFS with Best First Search Strategy and CFS with Greedy Stepwise Search Strategy, whereas Chi-square Feature Evaluation is most effective for SVM. No threshold was indicated for the ranking in Chi-square Feature Evaluation. As a result, CFS with Best First Search Strategy and CFS with Greedy Stepwise Search will be adapted in this study as feature section techniques for cultural classification with the use of Naive Bayes and SVM. 3. METHODOLOGY 3.1 Creation of Feature Sets 400 MIDI files were collected from Musica International [9], composed of 100 Scottish, 100 English, 100 American, and 100 French traditional songs. Four classification challenges were created: Challenge 1: UK VS NonUK; Challenge 2: Scottish VS English; Challenge 3: American VS French; Challenge 4: Scottish, English, American, and French. The corresponding datasets were prepared as follows: 1. Dataset 1: 100 Scottish and 100 English labeled as UK and 100 American and 100 French labeled as NonUK consisting of 400 songs. 2. Dataset 2: 100 Scottish labeled as Scottish and 100 English labeled as English consisting of 200 songs. 3. Dataset 3: 100 American labeled as American and 100 French labeled as French consisting of 200 songs. 4. Dataset 4: 100 Scottish labeled as Scottish, 100 English labeled as English, 100 American labeled as American, and 100 French labeled as French consisting of 400 songs. Each of the four datasets was loaded in jsymbolic for feature extraction. Only the 54 melodic features identified by [6] were selected. Each of the four datasets was loaded in WEKA 3, an open source software with a collection of different machine learning algorithm implementations in JAVA [10]. C4.5 decision tree, or J48 (JAVA implementation) was used to identify the subset of features for classification. Each of the four datasets was loaded in jsymbolic for feature extraction. Only the subsets of features identified by C4.5 were selected. 1. Dataset 1: 25 features. 2. Dataset 2: 21 features. 3. Dataset 3: 9 features. 4. Dataset 4: 32 features. Each of the four datasets was loaded in WEKA 3 for classification. Two classification algorithms were used for each dataset: Naive Bayes and SVM with 10-fold cross-validation. Each of the four datasets was loaded in jsymbolic for feature extraction. Only the 13 features mapped by [6] were selected. After feature extraction, each of the four datasets was loaded in WEKA 3 for classification using Naive Bayes and SVM with 10- fold cross-validation. Table 1. Result of Classification Naïve Bayes SVM Towsey C4.5 Towsey C4.5 UK-NonUK 65% 69% 72% 76.75% Scottish- English 64% 80% 71% 83.5% American- French 84.5% 80% 86% 83.5% Scottish-English- American-French 54.75% 53.75% 60.50% 65.75% According to the results, a higher accuracy can be achieved by using the C4.5 feature sets for challenges 1 and 2, and the Towsey feature set for challenge 3 when classifying with Naive Bayes and SVM. For challenge 4, the Towsey feature set performs better with Naive Bayes while the C4.5 feature set performs better with SVM. The winning feature set for classification challenge was selected for feature selection. The following winning feature sets were chosen for feature selection: C4.5 feature set for challenge 1, C4.5 feature set for challenge 2, Towsey feature set for challenges 3 and 4, and C4.5 feature set for challenge 4. 4 feature selection algorithms were used: Multiple Linear Regression (MLR), Correlation-based Feature Selector (CFS) with Best First Search (BFS), CFS with Greedy Stepwise, and CFS with Genetic Search. The datasets with each of the five feature sets was loaded in R for MLR. The formula used to create the linear models is: model1 <- lm(category f1+f2+f3+f4+.fn,dataset1) where the category corresponds to 0 or 1 (for challenges 1 to 3) and 0 to 3 (for challenge 4) and where f1 to fn represents each of the features.

3 The feature sets were reduced according to the coefficient of the features. If two or more coefficients are similar, only one of those will be selected and the other/s will be disregarded. The dataset with each of the five feature sets was loaded in WEKA 3 for feature selection with Correlation-based Feature Selector (CFS) with Best First Search (BFS), CFS with Greedy Stepwise, and CFS with Genetic Search. The datasets were used as full training sets. The datasets were loaded in jsymbolic for feature extraction. The features extracted are the reduced winning feature sets in preparation for classification. The datasets with the reduced feature sets were loaded WEKA 3 for classification. Two classification algorithms were used for each dataset: Naive Bayes and SVM with 10-fold cross-validation. The 4 datasets with the 54 jsymbolic melodic features were loaded in Weka 3 for feature selection with Correlation-based Feature Selector (CFS) with Best First Search (BFS), CFS with Greedy Stepwise, and CFS with Genetic Search. The datasets were used as full training sets. The four datasets were loaded in jsymbolic for feature extraction. The features extracted are the reduced melodic feature sets in preparation for classification. The datasets with the reduced feature sets were loaded WEKA 3 for classification. The dataset with the melodic features were also loaded in WEKA 3 for classification. Two classification algorithms were used for each dataset: Naive Bayes and SVM with 10-fold cross-validation. 3.2 Creation of Proposed MIR System A typical MIR System works according to the following steps: 1. Input a query song (the actual song file). 2. Trigger the software to calculate for similar songs. 3. Software computes for similar songs. 4. Software returns top 10 most similar songs. In the typical setup mentioned above, feature extraction and distance computations are done in real time. Generally, this affects the turn-around time of the software, which relies heavily on the complications of feature extraction as well as the number of songs on its database. Since this research is mainly focused on the implementation of the various feature sets and its effectiveness in the retrieval of similar music, feature extraction and distance computation was precomputed using jsymbolic, for feature extraction, and R for distance computation. The similarity matrices replace the necessary computation that must be performed by the MIR System in order to retrieve the most similar songs. The Proposed MIR System was developed using NetBeans and Java version 1.6.0_45 on OS X The User Interface (UI) was also developed using NetBeans IDE. Figure 1. UI of Proposed MIR System The Proposed MIR System was designed to accept a song name as the query. The system will then use the various matrices to: 1. Identify the song's category A distance matrix containing the distances of each song from the four representative songs: one Scottish, one English, one American, and one French song, was computed using feature set 4 (for classification between Scottish, English, America, and French songs). The smallest distance will be identified as the song's category. 2. Reduce the candidate songs to half its size In this case, the dataset contains 400 MIDI files. The candidate songs will be reduced to 200. A distance matrix containing the distances of each song from all of the songs was computed using feature set 1 (for classification between UK and NonUK songs). The idea is to retrieve similar songs of the same larger class: searching from either only UK or NonUK songs. 3. Reduce the candidate songs to 10 Two distance matrices were created for this part. A distance matrix containing the distances of each song from all of the songs was computed using feature set 2 (for classification between Scottish and English songs) and feature set 3 (for classification between American and French songs). If the identified category is Scottish or English, the first distance matrix will be used, otherwise, the 2nd matrix will be used. The value of the 200 songs reduced from earlier will be retrieved from either the 2nd or the 3rd matrix, the songs with the top 10 smallest values will be retrieved as the top 10 most similar songs. The features of the system, those that are represented by the additional buttons: Create Baseline File, Compute Baseline, Clear output.csv, and Activate Genre Method was developed for the automatic computation involved with the testing of the system, which is discussed in the next section.

4 4. RESULTS AND ANALYSIS 4.1 Feature Sets Results Table 2. Feature Sets with the Highest Classification Accuracy UK and NonUK Classification C4.5 SVM % C4.5 Reduced by MLR Naïve Bayes % Scottish and English Classification C4.5 SVM % C4.5 Naïve Bayes 21 80% American and French Classification 54 jsymbolic Melodic SVM % Towsey Naïve Bayes % Towsey Reduced by CFS Greedy Stepwise Naïve Bayes % Scottish, English, American, and French Classification 54 jsymbolic Melodic Towsey Reduced by CFS BFS, CFS Genetic Search, and CFS Greedy Stepwise SVM 54 68% Naïve Bayes % The columns represent the following: the feature set, the classifier used, the number of features, and the classification accuracy. For UK and NonUK classification, the highest accuracy can be achieved when the C4.5 reduced by MLR feature set is used with Naive Bayes (70.25%) and the C4.5 feature set is used with SVM (76.75%). For Scottish and English classification, the highest accuracy can be achieved when the C4.5 feature set is used with Naive Bayes (80%) and SVM (83.5%). For American and French classification, the highest accuracy can be achieved when the Towsey feature set is used with Naive Bayes (84.5%) or when the Towsey reduced by CFS Greedy Stepwise feature set is used with Naive Bayes (84.5%). For SVM, the highest accuracy can be achieved when the 54 jsymbolic melodic feature set is used (86.5%). For Scottish, English, American, and French classification, the highest accuracy can be achieved when the Towsey reduced by CFS BFS, CFS Genetic Search, and CFS Greedy Stepwise feature set is used with Naive Bayes (57.5%) and the 54 jsymbolic melodic feature set is used with SVM (68%) Analysis For classification challenges 1 and 2: UK and NonUK; Scottish and English, C4.5 decision tree serves as a good feature selection technique that can improve cultural classification of traditional music in the context of cultural style. For classification challenge 3: American and French classification, the 13 Towsey feature set performs similarly with the Towsey reduced by CFS Greedy Stepwise feature set that only has 2 features when used with Naive Bayes. It can be concluded that only the two features: Pitch Variety and Range are needed for classification, which implies that the other features are irrelevant for classification when Naive Bayes is used. The other features may either be strongly correlated with each other or may not be correlated with the classes. The Towsey feature set only works well with classification challenges 3 and 4, where American and French classification are involved. The reason for this can be traced back to the study done by Towsey which was done using 36 melodies: 18 classical melodies by Bach, Mozart, Beethoven and Tchaikovsky; 10 preclassical compositions by Du Fay, Gesualdo, Gibbons, Montiverdi and Palestrina; 6 traditional nursery rhymes and two melodies from popular tunes [5]. The melodic style captured by the feature set that was identified might only apply with these types of songs. An accuracy of at least 70% can only be achieved in the first 3 classification challenges. It is more difficult to classify music into 4 classes specifically in: Scottish, English, American, and French music. In most of the cases, feature selection techniques are needed in order to achieve the highest accuracy. In classification challenge 1, the C4.5 feature set reduced from the 54 jsymbolic melodic feature set is needed to achieve the highest accuracy with SVM. MLR is needed to reduce the C4.5 feature set to achieve the highest accuracy with Naive Bayes. In classification challenge 2, the C4.5 feature set reduced from the 54 jsymbolic melodic feature set is also needed to achieve the highest accuracy both for Naive Bayes and SVM. In classification challenge 3: the Towsey feature set is enough to achieve the highest accuracy with Naive Bayes while classification with SVM does not require any feature selection. In classification challenge 4, the Towsey feature set must be reduced by CFS with either of the three search strategies: BFS, Genetic Search, and Greedy Stepwise for classification with Naive Bayes. No feature selection is needed to achieve the highest accuracy for classification with SVM. Among the eight cases, only three cases do not require feature selection: challenge 3 with the use of the Towsey feature set with Naive Bayes, challenge 3 with the use of the 54 jsymbolic melodic feature set with SVM, and challenge 4 with the use of the 54 jsymbolic melodic feature set with SVM. For some of the classification challenges that require feature selection for a higher accuracy, various feature selection techniques must be used. Different challenges require different feature selection techniques. In terms of classification algorithm, SVM performs better in the cultural classification of traditional music in the context of cultural style. 4.2 Proposed MIR System Dataset Two datasets were used to test the Proposed MIR System. The first dataset used was the original dataset that was used to train the feature sets. This dataset is composed of 400 MIDI files: 100 Scottish, 100 English, 100 American, and 100 French traditional songs collected from [9]. The second dataset consists of the songs of the first dataset together with 20 new songs: 5 Scottish, 5 English, 5 American, and 5 French traditional songs also collected from [9]. The second dataset is composed of 420 traditional songs. Only 5 songs were collected for each category, due to lack of new traditional MIDIs from [9] Consistency Using the original dataset, the leave-one-out method was used to determine the behavior of the system in terms of retrieving the top

5 10 similar songs from a query song. The method can be described by the following steps: 1. Input a query song. 2. Record the top 10 retrieved songs. 3. Input each of the recorded songs as a query song and record the retrieved songs. 4. Compare the first recorded list of songs from each of the recorded set of songs. 5. Score each of the lists. 100% for 9 similar songs. (9 is the highest number since one songs was used as the query for the list.) The consistency can be described by this formula: Consistency = (Number of Similar Songs/9)*100 One song was chosen for each of the categories: one Scottish, one English, one American, and one French song. The songs were chosen at random: 79thfare for Scottish, abidewit for English, afamousr for American, and adieusui for French. After the first experiment, it was observed that some of the songs had similar names. In order to remove this ambiguity, the songs were renamed. The category of the song was added to its song name. Example: 79thfare-scottish. After renaming the songs, the experiment was repeated. For visualization, the values were plotted using a line graph. The x-axis corresponds to the rank of the query song. The y-axis corresponds to the consistency (%) of the song used. Figure 2. Consistency of Proposed MIR System Using 4 Songs A decreasing trend was observed. Due to the interesting observation, the experiment was redesigned. In order to get results that are more descriptive of the entire dataset, 20 songs (20% of the dataset) were selected for each category. A total of 80 songs were used for initial querying. The mean of the consistencies (%) were then plotted on a graph with the same structure as the previous plot. Figure 3. Consistency of Proposed MIR System Using 80 Songs By getting the mean of 20 songs per category, the plot presents results that are more descriptive of the entire dataset. The lines are smoother compared to the original plot, possibly because outliers were tamed by getting the average of 20 songs. The same decreasing trend can be observed. The decreasing trend is true for all of the categories. It can be inferred that the decreasing trend is caused by the level of similarities of the songs used as a query. When the songs on the lower part of the recorded list are used, it returns a lower number of songs that are similar to the recorded songs since it is less similar from the original query. Thus, the decreasing trend can be expected from a typical MIR System. Note that the overlapping of the lines is irrelevant in terms of displaying a relationship between the categories. The y-axis simply corresponds to the level of accuracy. Therefore, the overlaps are insignificant Quality Genre/Class Method In order to gauge the ability of the Proposed MIR System to retrieve melodically similar songs, the Genre Method was used. The same method was used by [11] to test the quality of his proposed music similarity algorithm. The principle behind the method proposes that if the genre of the query and the most similar song match, it is probable that they are the same. For this particular experiment, instead of using genre, cultural source was used as the category. The original dataset was used. The method can be described by the following steps: 1. Input a query song. 2. Retrieve the class of the top most similar song. 3. Record the class of the query song and class of the top most similar song. 4. Repeat until all songs in the Music Similarity Software have been used as a query. 5. Compute the accuracy of prediction. In a way, the Proposed MIR System is being used to predict the classification of the query by assigning the category of the top most similar song to the query. A confusion matrix was used to compute the accuracy of prediction. An overall accuracy of 52.75% was achieved.

6 Figure 4. Confusion Matrix Produced by Genre/Class Method In the first experiment, it was established that some of the songs were present on two or more categories. This implies that fuzzy boundaries exist between the categories and this can be used to interpret the values in the confusion matrix that was created from the Genre/Class Experiment. The table below represents a summary of the songs that belong to two or more categories. The 1 s indicate the song s category. Figure 5. Songs Belonging to Two or More Categories It can be noticed that the categories Scottish, English, and American are closely related. This can be used to explain why a high number of Scottish, English, and American songs have been labeled either as Scottish, English, or American. Interestingly, the category French is more defined/distinct from the other categories. This is consistent with what can be observed in the confusion matrix. Only a few Scottish, English, and American songs have been falsely labeled as French and only a few French songs have been falsely labeled as Scottish, English, and French. Folk songs are commonly shared by oral tradition. As a result, geographic location may be used to describe the fuzzy boundaries between the categories. The succeeding images present the location of Scotland, England, America, and France. Figure 6. World Map from [12] Figure 7. UK Map from [13] The fuzzy boundary between Scottish and English songs can be supported by its geographic location. The countries are relatively near to each other. On the contrary, the fuzzy boundary between Scottish, English, and American songs seems peculiar. While France is located relatively nearer to Scotland and England, the boundary between Scottish, English, and American songs are fuzzier compared to what is expected. Walter Wiora can explain the odd relationship between French, Scottish, and English songs. In his study of German folk songs, he was able to identify seven types of variation a folk song can transform into: changes in contour, changes in tonality, changes in rhythm, insertion and deletion of parts, changes in form, changes in expression, and demolition of the melody. The difficulty of understanding melodic variation caused by oral transmission was also stated by Bertrand Bronson [14]. Taking this into consideration, geography does not guarantee that countries relatively close to each other will have more similar songs New Songs Method It can be observed that the previous experiment excludes 9 out of the 10 retrieved songs. In order to take all of the retrieved songs into account, another experiment was performed, which involves all of the retrieved songs. Also, the 2nd dataset was used where the new songs were used for querying. The method can be described by the following steps: 1. Input a new song (song that does not belong to the training set) as a query. 2. Record the category of the retrieved songs. 3. Compute the accuracy of prediction for each possible category. In this experiment, the classification of the query song does not only rely on the first song but also on the other retrieved songs. The accuracy is computed by using this formula: Accuracy = (Number of Correct Classification/10)*100

7 Figure 8. New Songs Method for Scottish Songs The close relationship of Scottish, English, and American is evident when using the new Scottish songs as a query. In the case of the 5th new Scottish songs, all retrieved songs were Scottish. Figure 9. New Songs Method for English Songs The close relationship of Scottish, English, and American is also evident when using the new English songs as a query. Mostly, English songs were retrieved except in the case of the 2nd song, no English songs were retrieved. Figure 10. New Songs Method for American Songs For all of the new American songs, except the 3rd song, majority of the retrieved songs are American songs. The close relationship of Scottish, English, and American is also evident when using the new American songs as a query. Figure 11. New Songs Method for French Songs The 1st and 2nd songs retrieved mostly French songs. However, in the case of the 3rd and 4th song, no French songs were retrieved, English and Scottish songs were mostly retrieved. The 5th song mostly retrieved American songs Subjective Evaluation According to the previous experiment, the ideal case would be for all of the retrieved songs to have categories similar to that of the query. The worst case would be for none of the retrieved songs to have a similar category as to that of the query. Two sets of MIDI files were played and were briefly listened to. One lists represented the ideal case and one list represented the worst case. Upon listening to the first list, one can easily conclude that the songs are alike in terms of style, specifically, songs having a flute like characteristic. On the contrary, upon listening to the second list, one can conclude that the songs have nothing in common. The results are consistent with the Genre/Class Method that, indeed, songs of the same class are most likely similar. Consequently, it can be deduced that the Proposed MIR System, to a certain extent, is able to retrieve similar songs given a query. 5. CONCLUSION In this study, various feature sets were created for the cultural classification of traditional music in the context of cultural source. Existing feature subsets such as that of Towsey and the 54 jsymbolic melodic feature set were used, as well as new feature subsets created with C4.5 decision tree, Multiple Linear Regression (MLR) and Correlation-based Feature Selector (CFS) with three search strategies: Best First Search (BFS), Genetic Search, and Greedy Stepwise. Naive Bayes and SVM were used for classification. An accuracy of at least 70% was achieved when classifying between two classes: UK and NonUK, Scottish and English, American and French. It was determined that existing feature sets such as that of Towsey and the 54 jsymbolic melodic feature set can be used to achieve the highest accuracy in some of the classification challenges. It was observed that most cases however, require feature selection techniques to achieve the highest accuracy and various feature selection techniques must to used to improve feature sets. It was also discovered that SVM performs better than Naive Bayes in the cultural classification of traditional music in the context of cultural source. The various feature sets identified were used in a Propose MIR System. The Proposed MIR System uses the various feature sets as similarity measures when computing the similarity of a query to the songs that are available. Simple Euclidean distance was used to compute the similarities. While examining the behavior of the Proposed MIR System, it was determined that training set contained songs that belong to two or more categories. As a result,

8 it was established that a fuzzy boundary exists between Scottish, English, and American songs. This finding was supported by the results of using the Genre/Class Method and the New Songs Method that was used for testing the ability of the system to retrieve similar songs. The results also reflected Walter Wiora's concept that folk songs transmitted orally undergo multiple transformations and may change its form given that countries relatively near each other do not have fuzzy boundaries. Lastly, when subjective evaluation was used, it was identified that, to a certain extent, the Proposed MIR System has the ability to retrieve similar songs given a query. 6. FURTHER STUDY Going back to the results, only subjective evaluation can truly identify whether the Proposed MIR System is able to retrieve similar songs. Though the other methods used: the Genre/Class Method and the New Songs Method are conceptually accepted methods, it cannot replace evaluation done by a human listener. Based from the interesting results from the Genre/Class Method and the New Songs Method, this study suggests that a human listener must evaluate the dataset that was used, to identify whether a fuzzy boundary truly exists between the classes and also to identify whether the retrieved songs are in fact the most similar songs given a query. 7. ACKNOWLEDGMENTS Our thanks to DOST-ERDT for the funding of this research. 8. REFERENCES [1] Yuxiang Liu, Qiaoliang Xiang, Ye Wang, and Lianhong Cai. In Acoustics, Speech and Signal Processing, ICASSP IEEE International Conference on, pages 57 60, April. [2] Stephen J. Downie. The music information retrieval evaluation exchange ( ): A window into music information retrieval research. Acoustical Science and Technology, 29(4): , [3] MIREX Home Page, [4] Jiawei Han and Micheline Kamber. Data Mining : Concepts and Techniques. Elsevier, [5] Michael Towsey, Andrew Brown, Susan Wright, and Joachim Diederich. Towards melodic extension using genetic algorithms. Educational Technology and Society, [6] Andrei Coronel, Ariel Maguyon, and Proceso Fernandez. Specifying features for classical and non-classical melody evaluation. PCSC, [7] Cory Mckay and Ichiro Fujinaga. jsymbolic: A feature extractor for midi files. In Int. Computer Music Conf, pages , [8] Shyamala Doraisamy, Shahram Golzari, Noris Mohd Norowi, Md. Nasir, B Sulaiman, and Nur Izura Udzir. ISMIR 2008 Session 3A Content-Based Retrieval, Categorization and Similarity, A Study on Feature Selection and Classification Techniques for Automatic Genre Classification of Traditional Malay Music. [9] Musica International, January [10] Weka 3: Data Mining Software in Java, February [11] D. Schnitzer. Mirage High-Performance Music Similarity Computation and Automatic Playlist Generation. Master s thesis, Vienna University of Technology, [12] K483295worldmap, April [13] The United Kingdom of Great Britain and Northern Ireland, March [14] Peter Van Kranenburg, Jo rg Garbers, Anja Volk, Frans Wiering, Louis P. Grijp, and Remco C. Veltkamp. Collaboration perspectives for folk song research and music information retrieval: The indispensable role of computational musicology.

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

Music Emotion Recognition. Jaesung Lee. Chung-Ang University Music Emotion Recognition Jaesung Lee Chung-Ang University Introduction Searching Music in Music Information Retrieval Some information about target music is available Query by Text: Title, Artist, or

More information

CALCULATING SIMILARITY OF FOLK SONG VARIANTS WITH MELODY-BASED FEATURES

CALCULATING SIMILARITY OF FOLK SONG VARIANTS WITH MELODY-BASED FEATURES CALCULATING SIMILARITY OF FOLK SONG VARIANTS WITH MELODY-BASED FEATURES Ciril Bohak, Matija Marolt Faculty of Computer and Information Science University of Ljubljana, Slovenia {ciril.bohak, matija.marolt}@fri.uni-lj.si

More information

Outline. Why do we classify? Audio Classification

Outline. Why do we classify? Audio Classification Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification Implementation Future Work Why do we classify

More information

Specifying Features for Classical and Non-Classical Melody Evaluation

Specifying Features for Classical and Non-Classical Melody Evaluation Specifying Features for Classical and Non-Classical Melody Evaluation Andrei D. Coronel Ateneo de Manila University acoronel@ateneo.edu Ariel A. Maguyon Ateneo de Manila University amaguyon@ateneo.edu

More information

jsymbolic 2: New Developments and Research Opportunities

jsymbolic 2: New Developments and Research Opportunities jsymbolic 2: New Developments and Research Opportunities Cory McKay Marianopolis College and CIRMMT Montreal, Canada 2 / 30 Topics Introduction to features (from a machine learning perspective) And how

More information

MUSI-6201 Computational Music Analysis

MUSI-6201 Computational Music Analysis MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)

More information

Music Genre Classification and Variance Comparison on Number of Genres

Music Genre Classification and Variance Comparison on Number of Genres Music Genre Classification and Variance Comparison on Number of Genres Miguel Francisco, miguelf@stanford.edu Dong Myung Kim, dmk8265@stanford.edu 1 Abstract In this project we apply machine learning techniques

More information

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval DAY 1 Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval Jay LeBoeuf Imagine Research jay{at}imagine-research.com Rebecca

More information

Subjective Similarity of Music: Data Collection for Individuality Analysis

Subjective Similarity of Music: Data Collection for Individuality Analysis Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp

More information

FACTORS AFFECTING AUTOMATIC GENRE CLASSIFICATION: AN INVESTIGATION INCORPORATING NON-WESTERN MUSICAL FORMS

FACTORS AFFECTING AUTOMATIC GENRE CLASSIFICATION: AN INVESTIGATION INCORPORATING NON-WESTERN MUSICAL FORMS FACTORS AFFECTING AUTOMATIC GENRE CLASSIFICATION: AN INVESTIGATION INCORPORATING NON-WESTERN MUSICAL FORMS Noris Mohd Norowi, Shyamala Doraisamy, Rahmita Wirza Faculty of Computer Science and Information

More information

jsymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada

jsymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada jsymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada What is jsymbolic? Software that extracts statistical descriptors (called features ) from symbolic music files Can read: MIDI MEI (soon)

More information

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION ULAŞ BAĞCI AND ENGIN ERZIN arxiv:0907.3220v1 [cs.sd] 18 Jul 2009 ABSTRACT. Music genre classification is an essential tool for

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Aric Bartle (abartle@stanford.edu) December 14, 2012 1 Background The field of composer recognition has

More information

Singer Traits Identification using Deep Neural Network

Singer Traits Identification using Deep Neural Network Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic

More information

Supervised Learning in Genre Classification

Supervised Learning in Genre Classification Supervised Learning in Genre Classification Introduction & Motivation Mohit Rajani and Luke Ekkizogloy {i.mohit,luke.ekkizogloy}@gmail.com Stanford University, CS229: Machine Learning, 2009 Now that music

More information

Audio Feature Extraction for Corpus Analysis

Audio Feature Extraction for Corpus Analysis Audio Feature Extraction for Corpus Analysis Anja Volk Sound and Music Technology 5 Dec 2017 1 Corpus analysis What is corpus analysis study a large corpus of music for gaining insights on general trends

More information

Lyrics Classification using Naive Bayes

Lyrics Classification using Naive Bayes Lyrics Classification using Naive Bayes Dalibor Bužić *, Jasminka Dobša ** * College for Information Technologies, Klaićeva 7, Zagreb, Croatia ** Faculty of Organization and Informatics, Pavlinska 2, Varaždin,

More information

Introductions to Music Information Retrieval

Introductions to Music Information Retrieval Introductions to Music Information Retrieval ECE 272/472 Audio Signal Processing Bochen Li University of Rochester Wish List For music learners/performers While I play the piano, turn the page for me Tell

More information

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? NICHOLAS BORG AND GEORGE HOKKANEN Abstract. The possibility of a hit song prediction algorithm is both academically interesting and industry motivated.

More information

Automatic Music Clustering using Audio Attributes

Automatic Music Clustering using Audio Attributes Automatic Music Clustering using Audio Attributes Abhishek Sen BTech (Electronics) Veermata Jijabai Technological Institute (VJTI), Mumbai, India abhishekpsen@gmail.com Abstract Music brings people together,

More information

Automatic Piano Music Transcription

Automatic Piano Music Transcription Automatic Piano Music Transcription Jianyu Fan Qiuhan Wang Xin Li Jianyu.Fan.Gr@dartmouth.edu Qiuhan.Wang.Gr@dartmouth.edu Xi.Li.Gr@dartmouth.edu 1. Introduction Writing down the score while listening

More information

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu

More information

Release Year Prediction for Songs

Release Year Prediction for Songs Release Year Prediction for Songs [CSE 258 Assignment 2] Ruyu Tan University of California San Diego PID: A53099216 rut003@ucsd.edu Jiaying Liu University of California San Diego PID: A53107720 jil672@ucsd.edu

More information

Feature-Based Analysis of Haydn String Quartets

Feature-Based Analysis of Haydn String Quartets Feature-Based Analysis of Haydn String Quartets Lawson Wong 5/5/2 Introduction When listening to multi-movement works, amateur listeners have almost certainly asked the following situation : Am I still

More information

TOWARDS STRUCTURAL ALIGNMENT OF FOLK SONGS

TOWARDS STRUCTURAL ALIGNMENT OF FOLK SONGS TOWARDS STRUCTURAL ALIGNMENT OF FOLK SONGS Jörg Garbers and Frans Wiering Utrecht University Department of Information and Computing Sciences {garbers,frans.wiering}@cs.uu.nl ABSTRACT We describe an alignment-based

More information

MELODY ANALYSIS FOR PREDICTION OF THE EMOTIONS CONVEYED BY SINHALA SONGS

MELODY ANALYSIS FOR PREDICTION OF THE EMOTIONS CONVEYED BY SINHALA SONGS MELODY ANALYSIS FOR PREDICTION OF THE EMOTIONS CONVEYED BY SINHALA SONGS M.G.W. Lakshitha, K.L. Jayaratne University of Colombo School of Computing, Sri Lanka. ABSTRACT: This paper describes our attempt

More information

Creating a Feature Vector to Identify Similarity between MIDI Files

Creating a Feature Vector to Identify Similarity between MIDI Files Creating a Feature Vector to Identify Similarity between MIDI Files Joseph Stroud 2017 Honors Thesis Advised by Sergio Alvarez Computer Science Department, Boston College 1 Abstract Today there are many

More information

Detecting Musical Key with Supervised Learning

Detecting Musical Key with Supervised Learning Detecting Musical Key with Supervised Learning Robert Mahieu Department of Electrical Engineering Stanford University rmahieu@stanford.edu Abstract This paper proposes and tests performance of two different

More information

Evaluating Melodic Encodings for Use in Cover Song Identification

Evaluating Melodic Encodings for Use in Cover Song Identification Evaluating Melodic Encodings for Use in Cover Song Identification David D. Wickland wickland@uoguelph.ca David A. Calvert dcalvert@uoguelph.ca James Harley jharley@uoguelph.ca ABSTRACT Cover song identification

More information

The Million Song Dataset

The Million Song Dataset The Million Song Dataset AUDIO FEATURES The Million Song Dataset There is no data like more data Bob Mercer of IBM (1985). T. Bertin-Mahieux, D.P.W. Ellis, B. Whitman, P. Lamere, The Million Song Dataset,

More information

Automatic Music Genre Classification

Automatic Music Genre Classification Automatic Music Genre Classification Nathan YongHoon Kwon, SUNY Binghamton Ingrid Tchakoua, Jackson State University Matthew Pietrosanu, University of Alberta Freya Fu, Colorado State University Yue Wang,

More information

Composer Style Attribution

Composer Style Attribution Composer Style Attribution Jacqueline Speiser, Vishesh Gupta Introduction Josquin des Prez (1450 1521) is one of the most famous composers of the Renaissance. Despite his fame, there exists a significant

More information

Chord Classification of an Audio Signal using Artificial Neural Network

Chord Classification of an Audio Signal using Artificial Neural Network Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------

More information

Modeling memory for melodies

Modeling memory for melodies Modeling memory for melodies Daniel Müllensiefen 1 and Christian Hennig 2 1 Musikwissenschaftliches Institut, Universität Hamburg, 20354 Hamburg, Germany 2 Department of Statistical Science, University

More information

A MANUAL ANNOTATION METHOD FOR MELODIC SIMILARITY AND THE STUDY OF MELODY FEATURE SETS

A MANUAL ANNOTATION METHOD FOR MELODIC SIMILARITY AND THE STUDY OF MELODY FEATURE SETS A MANUAL ANNOTATION METHOD FOR MELODIC SIMILARITY AND THE STUDY OF MELODY FEATURE SETS Anja Volk, Peter van Kranenburg, Jörg Garbers, Frans Wiering, Remco C. Veltkamp, Louis P. Grijp* Department of Information

More information

Analysis and Clustering of Musical Compositions using Melody-based Features

Analysis and Clustering of Musical Compositions using Melody-based Features Analysis and Clustering of Musical Compositions using Melody-based Features Isaac Caswell Erika Ji December 13, 2013 Abstract This paper demonstrates that melodic structure fundamentally differentiates

More information

STRING QUARTET CLASSIFICATION WITH MONOPHONIC MODELS

STRING QUARTET CLASSIFICATION WITH MONOPHONIC MODELS STRING QUARTET CLASSIFICATION WITH MONOPHONIC Ruben Hillewaere and Bernard Manderick Computational Modeling Lab Department of Computing Vrije Universiteit Brussel Brussels, Belgium {rhillewa,bmanderi}@vub.ac.be

More information

Effects of acoustic degradations on cover song recognition

Effects of acoustic degradations on cover song recognition Signal Processing in Acoustics: Paper 68 Effects of acoustic degradations on cover song recognition Julien Osmalskyj (a), Jean-Jacques Embrechts (b) (a) University of Liège, Belgium, josmalsky@ulg.ac.be

More information

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS Andrew N. Robertson, Mark D. Plumbley Centre for Digital Music

More information

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC Scientific Journal of Impact Factor (SJIF): 5.71 International Journal of Advance Engineering and Research Development Volume 5, Issue 04, April -2018 e-issn (O): 2348-4470 p-issn (P): 2348-6406 MUSICAL

More information

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene Beat Extraction from Expressive Musical Performances Simon Dixon, Werner Goebl and Emilios Cambouropoulos Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria.

More information

SIMSSA DB: A Database for Computational Musicological Research

SIMSSA DB: A Database for Computational Musicological Research SIMSSA DB: A Database for Computational Musicological Research Cory McKay Marianopolis College 2018 International Association of Music Libraries, Archives and Documentation Centres International Congress,

More information

CS229 Project Report Polyphonic Piano Transcription

CS229 Project Report Polyphonic Piano Transcription CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project

More information

Dimensional Music Emotion Recognition: Combining Standard and Melodic Audio Features

Dimensional Music Emotion Recognition: Combining Standard and Melodic Audio Features Dimensional Music Emotion Recognition: Combining Standard and Melodic Audio Features R. Panda 1, B. Rocha 1 and R. P. Paiva 1, 1 CISUC Centre for Informatics and Systems of the University of Coimbra, Portugal

More information

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the

More information

Music Radar: A Web-based Query by Humming System

Music Radar: A Web-based Query by Humming System Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,

More information

Melody classification using patterns

Melody classification using patterns Melody classification using patterns Darrell Conklin Department of Computing City University London United Kingdom conklin@city.ac.uk Abstract. A new method for symbolic music classification is proposed,

More information

Music Mood Classification - an SVM based approach. Sebastian Napiorkowski

Music Mood Classification - an SVM based approach. Sebastian Napiorkowski Music Mood Classification - an SVM based approach Sebastian Napiorkowski Topics on Computer Music (Seminar Report) HPAC - RWTH - SS2015 Contents 1. Motivation 2. Quantification and Definition of Mood 3.

More information

A wavelet-based approach to the discovery of themes and sections in monophonic melodies Velarde, Gissel; Meredith, David

A wavelet-based approach to the discovery of themes and sections in monophonic melodies Velarde, Gissel; Meredith, David Aalborg Universitet A wavelet-based approach to the discovery of themes and sections in monophonic melodies Velarde, Gissel; Meredith, David Publication date: 2014 Document Version Accepted author manuscript,

More information

Music Information Retrieval

Music Information Retrieval CTP 431 Music and Audio Computing Music Information Retrieval Graduate School of Culture Technology (GSCT) Juhan Nam 1 Introduction ü Instrument: Piano ü Composer: Chopin ü Key: E-minor ü Melody - ELO

More information

Analysing Musical Pieces Using harmony-analyser.org Tools

Analysing Musical Pieces Using harmony-analyser.org Tools Analysing Musical Pieces Using harmony-analyser.org Tools Ladislav Maršík Dept. of Software Engineering, Faculty of Mathematics and Physics Charles University, Malostranské nám. 25, 118 00 Prague 1, Czech

More information

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM A QUER B EAMPLE MUSIC RETRIEVAL ALGORITHM H. HARB AND L. CHEN Maths-Info department, Ecole Centrale de Lyon. 36, av. Guy de Collongue, 69134, Ecully, France, EUROPE E-mail: {hadi.harb, liming.chen}@ec-lyon.fr

More information

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION Halfdan Rump, Shigeki Miyabe, Emiru Tsunoo, Nobukata Ono, Shigeki Sagama The University of Tokyo, Graduate

More information

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University Week 14 Query-by-Humming and Music Fingerprinting Roger B. Dannenberg Professor of Computer Science, Art and Music Overview n Melody-Based Retrieval n Audio-Score Alignment n Music Fingerprinting 2 Metadata-based

More information

Computational Modelling of Harmony

Computational Modelling of Harmony Computational Modelling of Harmony Simon Dixon Centre for Digital Music, Queen Mary University of London, Mile End Rd, London E1 4NS, UK simon.dixon@elec.qmul.ac.uk http://www.elec.qmul.ac.uk/people/simond

More information

Music Information Retrieval with Temporal Features and Timbre

Music Information Retrieval with Temporal Features and Timbre Music Information Retrieval with Temporal Features and Timbre Angelina A. Tzacheva and Keith J. Bell University of South Carolina Upstate, Department of Informatics 800 University Way, Spartanburg, SC

More information

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset Ricardo Malheiro, Renato Panda, Paulo Gomes, Rui Paiva CISUC Centre for Informatics and Systems of the University of Coimbra {rsmal,

More information

Singer Recognition and Modeling Singer Error

Singer Recognition and Modeling Singer Error Singer Recognition and Modeling Singer Error Johan Ismael Stanford University jismael@stanford.edu Nicholas McGee Stanford University ndmcgee@stanford.edu 1. Abstract We propose a system for recognizing

More information

Automatic Rhythmic Notation from Single Voice Audio Sources

Automatic Rhythmic Notation from Single Voice Audio Sources Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung

More information

CSC475 Music Information Retrieval

CSC475 Music Information Retrieval CSC475 Music Information Retrieval Symbolic Music Representations George Tzanetakis University of Victoria 2014 G. Tzanetakis 1 / 30 Table of Contents I 1 Western Common Music Notation 2 Digital Formats

More information

ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC

ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC Vaiva Imbrasaitė, Peter Robinson Computer Laboratory, University of Cambridge, UK Vaiva.Imbrasaite@cl.cam.ac.uk

More information

Content-based music retrieval

Content-based music retrieval Music retrieval 1 Music retrieval 2 Content-based music retrieval Music information retrieval (MIR) is currently an active research area See proceedings of ISMIR conference and annual MIREX evaluations

More information

arxiv: v1 [cs.ir] 16 Jan 2019

arxiv: v1 [cs.ir] 16 Jan 2019 It s Only Words And Words Are All I Have Manash Pratim Barman 1, Kavish Dahekar 2, Abhinav Anshuman 3, and Amit Awekar 4 1 Indian Institute of Information Technology, Guwahati 2 SAP Labs, Bengaluru 3 Dell

More information

Semi-supervised Musical Instrument Recognition

Semi-supervised Musical Instrument Recognition Semi-supervised Musical Instrument Recognition Master s Thesis Presentation Aleksandr Diment 1 1 Tampere niversity of Technology, Finland Supervisors: Adj.Prof. Tuomas Virtanen, MSc Toni Heittola 17 May

More information

Tempo and Beat Analysis

Tempo and Beat Analysis Advanced Course Computer Science Music Processing Summer Term 2010 Meinard Müller, Peter Grosche Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Tempo and Beat Analysis Musical Properties:

More information

Transcription of the Singing Melody in Polyphonic Music

Transcription of the Singing Melody in Polyphonic Music Transcription of the Singing Melody in Polyphonic Music Matti Ryynänen and Anssi Klapuri Institute of Signal Processing, Tampere University Of Technology P.O.Box 553, FI-33101 Tampere, Finland {matti.ryynanen,

More information

Music Information Retrieval

Music Information Retrieval Music Information Retrieval Informative Experiences in Computation and the Archive David De Roure @dder David De Roure @dder Four quadrants Big Data Scientific Computing Machine Learning Automation More

More information

Week 14 Music Understanding and Classification

Week 14 Music Understanding and Classification Week 14 Music Understanding and Classification Roger B. Dannenberg Professor of Computer Science, Music & Art Overview n Music Style Classification n What s a classifier? n Naïve Bayesian Classifiers n

More information

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC Vishweshwara Rao, Sachin Pant, Madhumita Bhaskar and Preeti Rao Department of Electrical Engineering, IIT Bombay {vishu, sachinp,

More information

A Computational Model for Discriminating Music Performers

A Computational Model for Discriminating Music Performers A Computational Model for Discriminating Music Performers Efstathios Stamatatos Austrian Research Institute for Artificial Intelligence Schottengasse 3, A-1010 Vienna stathis@ai.univie.ac.at Abstract In

More information

Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections

Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections 1/23 Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections Rudolf Mayer, Andreas Rauber Vienna University of Technology {mayer,rauber}@ifs.tuwien.ac.at Robert Neumayer

More information

A Basis for Characterizing Musical Genres

A Basis for Characterizing Musical Genres A Basis for Characterizing Musical Genres Roelof A. Ruis 6285287 Bachelor thesis Credits: 18 EC Bachelor Artificial Intelligence University of Amsterdam Faculty of Science Science Park 904 1098 XH Amsterdam

More information

MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES

MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES Jun Wu, Yu Kitano, Stanislaw Andrzej Raczynski, Shigeki Miyabe, Takuya Nishimoto, Nobutaka Ono and Shigeki Sagayama The Graduate

More information

Multi-Modal Music Emotion Recognition: A New Dataset, Methodology and Comparative Analysis

Multi-Modal Music Emotion Recognition: A New Dataset, Methodology and Comparative Analysis Multi-Modal Music Emotion Recognition: A New Dataset, Methodology and Comparative Analysis R. Panda 1, R. Malheiro 1, B. Rocha 1, A. Oliveira 1 and R. P. Paiva 1, 1 CISUC Centre for Informatics and Systems

More information

Automatic Labelling of tabla signals

Automatic Labelling of tabla signals ISMIR 2003 Oct. 27th 30th 2003 Baltimore (USA) Automatic Labelling of tabla signals Olivier K. GILLET, Gaël RICHARD Introduction Exponential growth of available digital information need for Indexing and

More information

Analysis of local and global timing and pitch change in ordinary

Analysis of local and global timing and pitch change in ordinary Alma Mater Studiorum University of Bologna, August -6 6 Analysis of local and global timing and pitch change in ordinary melodies Roger Watt Dept. of Psychology, University of Stirling, Scotland r.j.watt@stirling.ac.uk

More information

Automated extraction of motivic patterns and application to the analysis of Debussy s Syrinx

Automated extraction of motivic patterns and application to the analysis of Debussy s Syrinx Automated extraction of motivic patterns and application to the analysis of Debussy s Syrinx Olivier Lartillot University of Jyväskylä, Finland lartillo@campus.jyu.fi 1. General Framework 1.1. Motivic

More information

Lyric-Based Music Mood Recognition

Lyric-Based Music Mood Recognition Lyric-Based Music Mood Recognition Emil Ian V. Ascalon, Rafael Cabredo De La Salle University Manila, Philippines emil.ascalon@yahoo.com, rafael.cabredo@dlsu.edu.ph Abstract: In psychology, emotion is

More information

Contextual music information retrieval and recommendation: State of the art and challenges

Contextual music information retrieval and recommendation: State of the art and challenges C O M P U T E R S C I E N C E R E V I E W ( ) Available online at www.sciencedirect.com journal homepage: www.elsevier.com/locate/cosrev Survey Contextual music information retrieval and recommendation:

More information

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis Automatic characterization of ornamentation from bassoon recordings for expressive synthesis Montserrat Puiggròs, Emilia Gómez, Rafael Ramírez, Xavier Serra Music technology Group Universitat Pompeu Fabra

More information

Exploring Relationships between Audio Features and Emotion in Music

Exploring Relationships between Audio Features and Emotion in Music Exploring Relationships between Audio Features and Emotion in Music Cyril Laurier, *1 Olivier Lartillot, #2 Tuomas Eerola #3, Petri Toiviainen #4 * Music Technology Group, Universitat Pompeu Fabra, Barcelona,

More information

Efficient Vocal Melody Extraction from Polyphonic Music Signals

Efficient Vocal Melody Extraction from Polyphonic Music Signals http://dx.doi.org/1.5755/j1.eee.19.6.4575 ELEKTRONIKA IR ELEKTROTECHNIKA, ISSN 1392-1215, VOL. 19, NO. 6, 213 Efficient Vocal Melody Extraction from Polyphonic Music Signals G. Yao 1,2, Y. Zheng 1,2, L.

More information

Exploring the Design Space of Symbolic Music Genre Classification Using Data Mining Techniques Ortiz-Arroyo, Daniel; Kofod, Christian

Exploring the Design Space of Symbolic Music Genre Classification Using Data Mining Techniques Ortiz-Arroyo, Daniel; Kofod, Christian Aalborg Universitet Exploring the Design Space of Symbolic Music Genre Classification Using Data Mining Techniques Ortiz-Arroyo, Daniel; Kofod, Christian Published in: International Conference on Computational

More information

Music Genre Classification

Music Genre Classification Music Genre Classification chunya25 Fall 2017 1 Introduction A genre is defined as a category of artistic composition, characterized by similarities in form, style, or subject matter. [1] Some researchers

More information

Methods for the automatic structural analysis of music. Jordan B. L. Smith CIRMMT Workshop on Structural Analysis of Music 26 March 2010

Methods for the automatic structural analysis of music. Jordan B. L. Smith CIRMMT Workshop on Structural Analysis of Music 26 March 2010 1 Methods for the automatic structural analysis of music Jordan B. L. Smith CIRMMT Workshop on Structural Analysis of Music 26 March 2010 2 The problem Going from sound to structure 2 The problem Going

More information

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES 12th International Society for Music Information Retrieval Conference (ISMIR 2011) A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES Erdem Unal 1 Elaine Chew 2 Panayiotis Georgiou

More information

Topic 10. Multi-pitch Analysis

Topic 10. Multi-pitch Analysis Topic 10 Multi-pitch Analysis What is pitch? Common elements of music are pitch, rhythm, dynamics, and the sonic qualities of timbre and texture. An auditory perceptual attribute in terms of which sounds

More information

AN EMOTION MODEL FOR MUSIC USING BRAIN WAVES

AN EMOTION MODEL FOR MUSIC USING BRAIN WAVES AN EMOTION MODEL FOR MUSIC USING BRAIN WAVES Rafael Cabredo 1,2, Roberto Legaspi 1, Paul Salvador Inventado 1,2, and Masayuki Numao 1 1 Institute of Scientific and Industrial Research, Osaka University,

More information

A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION

A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION Graham E. Poliner and Daniel P.W. Ellis LabROSA, Dept. of Electrical Engineering Columbia University, New York NY 127 USA {graham,dpwe}@ee.columbia.edu

More information

Hidden Markov Model based dance recognition

Hidden Markov Model based dance recognition Hidden Markov Model based dance recognition Dragutin Hrenek, Nenad Mikša, Robert Perica, Pavle Prentašić and Boris Trubić University of Zagreb, Faculty of Electrical Engineering and Computing Unska 3,

More information

Multimodal Music Mood Classification Framework for Christian Kokborok Music

Multimodal Music Mood Classification Framework for Christian Kokborok Music Journal of Engineering Technology (ISSN. 0747-9964) Volume 8, Issue 1, Jan. 2019, PP.506-515 Multimodal Music Mood Classification Framework for Christian Kokborok Music Sanchali Das 1*, Sambit Satpathy

More information

Improving Frame Based Automatic Laughter Detection

Improving Frame Based Automatic Laughter Detection Improving Frame Based Automatic Laughter Detection Mary Knox EE225D Class Project knoxm@eecs.berkeley.edu December 13, 2007 Abstract Laughter recognition is an underexplored area of research. My goal for

More information

Music Similarity and Cover Song Identification: The Case of Jazz

Music Similarity and Cover Song Identification: The Case of Jazz Music Similarity and Cover Song Identification: The Case of Jazz Simon Dixon and Peter Foster s.e.dixon@qmul.ac.uk Centre for Digital Music School of Electronic Engineering and Computer Science Queen Mary

More information

GRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM

GRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM 19th European Signal Processing Conference (EUSIPCO 2011) Barcelona, Spain, August 29 - September 2, 2011 GRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM Tomoko Matsui

More information

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST) Computational Models of Music Similarity 1 Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST) Abstract The perceived similarity of two pieces of music is multi-dimensional,

More information

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring 2009 Week 6 Class Notes Pitch Perception Introduction Pitch may be described as that attribute of auditory sensation in terms

More information

The song remains the same: identifying versions of the same piece using tonal descriptors

The song remains the same: identifying versions of the same piece using tonal descriptors The song remains the same: identifying versions of the same piece using tonal descriptors Emilia Gómez Music Technology Group, Universitat Pompeu Fabra Ocata, 83, Barcelona emilia.gomez@iua.upf.edu Abstract

More information

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t MPEG-7 FOR CONTENT-BASED MUSIC PROCESSING Λ Emilia GÓMEZ, Fabien GOUYON, Perfecto HERRERA and Xavier AMATRIAIN Music Technology Group, Universitat Pompeu Fabra, Barcelona, SPAIN http://www.iua.upf.es/mtg

More information

Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals

Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals Eita Nakamura and Shinji Takaki National Institute of Informatics, Tokyo 101-8430, Japan eita.nakamura@gmail.com, takaki@nii.ac.jp

More information