A Basis for Characterizing Musical Genres

Size: px
Start display at page:

Download "A Basis for Characterizing Musical Genres"

Transcription

1 A Basis for Characterizing Musical Genres Roelof A. Ruis Bachelor thesis Credits: 18 EC Bachelor Artificial Intelligence University of Amsterdam Faculty of Science Science Park XH Amsterdam dr. A.K. Honingh Institute for Logic, Language and Computation Faculty of Science University of Amsterdam Science Park XG Amsterdam Supervisors dr. M.W. van Someren Informatics Institute Faculty of Science University of Amsterdam Science Park XG Amsterdam July 4th, 2014

2 Abstract By exploratory means, a method is presented for providing musical genre characterizations using songs from two MIDI corpora. This characterization is first based on global features and later refined using interval sequences derived from melodies. In the first part of the research a top 10 of characteristic global features per genre is established through correlation and clustering. In the second part this correlation is again used to determine a top 10 of characteristic interval sequences. Analysis is then presented as to show meaningful interpretation of the data, resulting in small genre characterizations. 2

3 Contents 1 Introduction 4 2 Related Work 4 3 The Corpora 5 4 Part 1 - Global Features Method Principal Component Analysis Unsupervised Clustering Point-Biserial Correlation Interpretation BOD main BOD popular BOD jazz BOD classical BAL Part 2 - Interval Sequences Method Preprocessing Interval Sequences Sorting on Occurrence Interpretation Conclusion 16 7 Discussion Part Part General A JSymbolic Features 20 B Results Part 1 21 C Results Part

4 1 Introduction When humans talk about music, they tend to group songs with comparable characteristics as belonging to the same genre. Often people can recall some of the musical aspects on which they base this grouping but only tend to name superficial features like instrumentation. It would thus be interesting to see if a more exact genre description could be constructed. Musicologists have already tried to describe genres using literature on music and a good ear, but nowadays it is possible to use computational methods to analyze larger corpora and find more refined musical features. In light of this, musical genre classifiers have been built which obtain high accuracy on mimicking genre grouping behavior. They do however not provide information on genre-specific features, which are exactly what musicologists are eager to discover. It is therefore desirable to aid the musicologist with a method to extract musicologically clearly interpretable features related to genre, or better still, a computationally extracted genre description. Of course the difficulty in extracting such a description should not be underestimated; it is difficult to define beforehand the features by which a genre should be described. Furthermore, it could be that some genres may be defined more by meta-aspects of songs that can not be extracted from the MIDI context such as artist, year of release and geographical origin of a song. Because this research will work with MIDI files which contain no meta-data, such features will be left out and might be explored in future research. It is sensible to look at both global and surface features. Global features can be seen as aspects of songs to which an overall value can be given which is applicable to the whole song, such as electric guitar fraction or average time between attacks. In contrast, surface features focus on the actual melodic content and try to describe certain patterns therein. Either these global or surface aspects might contribute to more accurate genre descriptions and due to the difference in analysis for both, the research is split into two parts. During the first part, global features provided by a feature extractor will be examined using principal component analysis and various other machine learning techniques, working towards a high level genre description. The second part of the research will focus on finding genre specific melodic structures, which can be used as detailed genre descriptors. It is to be expected that these melodic structures themselves can then not only be used descriptively but also to improve the accuracy of existing classification methods. The interpretation of the results will incorporate results from part one as to check their communalities. Because other researchers have conducted very little research into finding explanatory features, the research presented here will be mainly exploratory, providing extensive documentation of the investigated paths and decisions that have been made. Furthermore, no statistical tests will be used at this stage, so if interesting results surface, more meticulous further research will be required. This documentation, therefore serves two goals. Firstly, trying to provide features, expressed in terms of global features and later refined and completed with melodic structures, gained from a corpus of genre specified MIDI songs, capable of meaningfully characterizing the different genres in the corpus. Secondly, it hopes to provide material for supplementary studies, which might be suitable for other than musicological research. For instance social network analysis, where meaningfull features play a role in characterization of groups of people. 2 Related Work Much research has been done into the machine based recognition of musical genres using only MIDI data [8] [3], a combination of MIDI and audio features [2] or features purely based on audio 4

5 [10]. All of the systems presented in these papers first apply some form of feature extraction whereafter they perform a classification based on those features. The classification accuracy is then evaluated and improvements or new ways of classifying music are claimed. These systems, however, do not provide any means of understanding what characterizes a musical genre but only how well it can be classified based on arbitrary features. Sturm [11] argues that because there are more independent variables changing between particular genres in a dataset, classification is unreliable and other types of experimental designs are required to understand how a genre can be characterized. He mentions, among others, inspecting features and answering the question At what is the system looking to identify the genres used by music? (p. 376). A further motivation for this research is the final remark of McKay and Fujinaga [8] in a small genre classification research stating that Further study of which features were selected by which specialist classifier ensembles could also be of great musicological interest. (p. 530). This further emphasizes the need for a thorough look into feature meaning and selection. The high level features required in the first part of the research will have to be extracted by a program capable of handling MIDI songs. The program used in this paper, created by McKay [7] is called JSymbolic and is capable of extracting up to 111 features which meanings are explained in the paper. McKay also notes The library of features used in this thesis should be seen as a work in progress that can continually be expanded and refined [...] (p. 62), stressing that the features extracted by JSymbolic could used some refinement; they do not provide the musicologist with much deep insight into song structure and melodic and rhythmic sequences but instead define a coarse value for, for instance, average note duration or average melodic interval. The second part of the research exploring the use of small melodic structures in genre characterisation, will use an idea presented by Conklin [5]. The paper presents a method for combining a series of notes from a melody into a segment and assigning these segments to a class, based on certain rules about their content. These melodic segment classes are then used for style discrimination of different melodies and results well above random chance are achieved. Honingh et al. [6] also use the aforementioned concept of melodic segment classes to distinguish between tonal and atonal music and develop a model of pitch class progression. Both these studies thus show that inspecting musical features at this scale might not only lead to determining genre distinctions but also provide more insight into which information is hidden in the structure of the music on a very detailed level. 3 The Corpora For this research, two genre annotated MIDI corpora were used. Firstly the Ballroom Dance Corpus [4], from here on indicated with BAL, containing 6 types of ballroom dances which structure can be observed in table 1. Secondly the Bodidharma MIDI Corpus [7] is used, indicated with BOD. The large version of the BOD corpus consists of 38 genres with a hierarchically defined structure with a maximum depth of four. There is, however, also a small subset of this large corpus consisting of 9 sub genres divided into three main genres that will be used here, but the method presented in this paper could be extended to cover the large version of the BOD corpus. The structure of the reduced BOD corpus as used in this research can be seen in table 2. 5

6 Genre Nr. of songs Bossa Nova 36 Mambo 15 Merengue 24 Rumba 20 Salsa 7 Tango 26 Table 1: Structure of the BAL corpus with song counts Because both corpora come prelabeled and are used in other researches, it is assumed that this labeling is correct. Main genre Subgenre Nr. of songs Popular 75 Hardcore Rap 25 Punk 25 Trad. Country 25 Jazz 75 Bebop 25 Jazz Soul 25 Swing 25 Classical 70 Baroque 25 Modern Classical 25 Romantic 20 Table 2: Structure of the BOD corpus with song counts 4 Part 1 - Global Features 4.1 Method Principal Component Analysis First, features are generated for all songs in both the BAL and BOD corpus using the JSymbolic extractor, producing a csv file with data for each separate genre. A complete list of the feature names is included as listing 1 in appendix A. To get an intuition about the high dimensional structure of the data and the possible overlap of genres, principal component analysis (PCA) is used to map the high dimensional data to a 2D space. This also provides means to visually detect outliers or anomalies and a first possible step to finding characteristic features. Visualization The BAL corpus is inspected first. Figure 1 shows all songs in this corpus mapped to a 2D principal component (PC) space using the two most dominant principal components as axes. All features were used in the calculation of these components. It can clearly be observed that most genres form quite compact clusters but that there is still a decent amount of overlap. Because all features were used it is to be expected that the clusters will get better when a decent subset of features is selected. Principal Component Bossa Nova Mambo Merengue Rumba Salsa Tango Principal Component 1 Figure 1: BAL corpus in 2D PC space based on all features. A few other things can be observed from this visualization. At first, the salsa songs are not forming a coherent cluster, which is also due to the fact that very few salsa songs were provided by the BAL corpus. For this reason the salsa genre is dropped for the rest of part 1 of the research. Secondly, the tango songs seem to form two distinct groups. Listening to the songs indicated that the group in the upper right corner consists of tango songs played on solo piano while the other tango songs are performed with a multiple instrument setup. Because it is likely that this distinction will affect the precision of the genre characterization, the tangos performed on piano solo are also removed from the data. The new BAL corpus with salsa and piano tango songs removed is shown in figure 2. 6

7 Principal Component Bossa Nova Mambo Merengue Rumba Tango Contribution Figure 3: Top level of BOD corpus in PC space based on all features PC vector 1 PC vector 2 PC vector Principal Component 1 Figure 2: BAL corpus in PC space without salsa and tango piano based on all features. The BOD corpus shows various degrees of such cluster formation. Clear cluster separation with a bit of overlap can be observed for the main level (figure 3) and the popular sub level (figure 8 in appendix B). Furthermore a clear distinction can be seen between bebop and swing in the jazz sub-genre (figure 9 in appendix B), while jazz soul does have a lot of overlap with bebop. The classical sub level (figure 10 in appendix B) shows the most overlap of all, indicating that the features used might not be working too well for separation of classical music or that a smaller selection of features might be required to get good separation. No other anomalies or strange outliers can be seen so removing any songs from the BOD corpus beforehand is not necessary. Principal Component Popular Jazz Classical Principal Component Dimension Figure 4: Contribution to dimensions of the first 3 PCA vectors for the BAL corpus. Inspecting Dimensions The PC axes are formed by a linear combination of the feature dimensions and it might be that they can explain the forming and position of the clusters fairly well. Results indicate however that too many features are contributing to one PC axis to be of any value. Figure 4 and 5 show the dimensions for the first 3 PCA vectors plotted against their individual dimensional contribution. If only a low number of features would contribute to the PC axis it is to be expected to get a few high contribution values for the very first dimensions and thereafter a steep decline with only very small contributions for the other dimensions. As the figures show, this is not the case for there is no steep decline and no large contribution by the first couple of feature values. Visual Aid While the actual PC axes are thus difficult to interpret by looking at the feature decomposition, it might be possible for a trained musicologist to use a PC plot as an aid for deriving genre characteristics by ear. Looking at figure 2, one could select a range of songs by varying the value of principal component 1 7

8 Contribution while keeping the value of principal component 2 at a fixed value. Listening to this range of songs might then provide insights into aspects of the music that change throughout. Concluding we have seen that while PCA is a method suited for detecting outliers and anomalies in the corpus and visualizing the forming of clusters it can not be used for detecting characteristic features, although it could be used as a visual aid PC vector 1 PC vector 2 PC vector 3 from the others) such an improvement is to be expected. Evaluating clustering Unsupervised clustering is finding clusters of songs which are close together in feature space. It can however be that the found clusters do not represent the original genre groups. To find out how well the found clusters match the original genres the following measure is devised: It is checked to which genre the majority of songs in a cluster belongs (scaled by the total amount of songs in the dataset for that genre), which is then assumed to be the correct class for that cluster. This is done for every cluster. Hereafter, Accuracy and recall are measured and the F1-score is derived for each genre individually. Because the cluster centroids are randomly initialized on each run, 1000 cycles of the clustering algorithm are ran to get a reliable mean. There are as many cluster centroids initialized as there are different genres in the measured corpus Dimension Figure 5: Contribution to dimensions of the first 3 PCA vectors for the top level of the BOD corpus Unsupervised Clustering To discover if the JSymbolic features can really be used for recognizing (and therefore characterizing) genres a solid measure is needed. Although clustering of genres can already be observed in the PCA plots, it provides no exact measure for the amount of separation between the different genres. It is although possible to use unsupervised k-means clustering to detect how well clusters form in the same way as the genre labels indicate for different feature sets. This way it is possible to compare clustering of different feature sets and see if the clustering improves. For characteristic features (features that are in particular suited for separating one genre Bossa Nova Mambo Merengue Rumba Tango Table 3: F1-scores BAL using all features. Clustering with all features Clustering the BAL corpus yields the F1-scores displayed in table 3. A score of 1 means complete separation of a genre into its own cluster and a score of 0 means that all songs for a genre got mixed up in other genre clusters. The values of this clustering correspond to the visual results obtained from the PCA analysis as shown in figure 2. Looking at that figure, the low value for mambo can be explained from the overlap of almost all of the mambo songs with merengue or rumba. These genres take precedence and thus mambo songs are almost always classified under a differ- 8

9 ent genre. Popular Jazz Classical Table 4: F1-scores BOD main using all features. Hardcore Rap Punk Trad. Country Table 5: F1-scores BOD popular using all features. Bebop Jazz Soul Swing Table 6: F1-scores BOD jazz using all features. Genre F1 Baroque Modern Classical Romantic Table 7: F1-scores BOD classical using all features. Within the BOD corpus, the top level as well as the three sub levels can be clustered and scored individually in the same way. Tables 4, 5, 6 and 7 show the F1-scores for clustering these genres. It can be observed that most of the genres already separate very decently on their own except for modern classical music. The fact that some genres are not well separable means that the features used here cannot fully explain the difference between them or that too many features are used Point-Biserial Correlation To get a better understanding of which features predict well for which genre and thus work towards finding actual characteristic features, correlation between the features and the genres can be calculated. Because the features have continuous values and the songs can be regarded as being dichotomous (a song either belongs or does not belong to the measured genre) point-biserial correlation can be used. For each individual genre this correlation is calculated, where all songs within the measured genre are said to be in class 1, and all other songs in class 0. Measuring the point-biserial correlation for all features and all genres in the BOD corpus yields a top 10 correlating features for each genre. Table 8 shows this top 10 for bossa nova together with the correlation scores. Feature 65. Rel. Note Density of Highest Line* Avg. Note Duration* Avg. Time Between Attacks* Rhythmic Variability* Repeated Notes** Importance of Bass Register Melodic Thirds** Avg. Nr. of Independent Voices Str. of Strongest Rhythm. Pulse* Comb. Str. of Two Str. Rhythmic Pulses* Table 8: Top 10 highest point-biserial correlation scores for the bossa nova genre (in BAL). The features in the table are highly correlated with the bossa nova genre when checked against all other genres in the BAL corpus, meaning that these features are likely to characterize bossa nova based on differences it has with the other genres in the corpus. Using the information provided in the research by McKay, a detailed description of the individual features can be given. As an example the first feature Relative Note Density of Highest Line will now be discussed. A musical piece often contains different lines or voices which have an average pitch. The highest line 9

10 refers to the voice with the highest average pitch. The relative note density of the highest line now is the number of notes in the voice with the highest average pitch divided by the average number of notes in all other voices. The high correlation is indicating that, relative to the other genres, bossa nova has a large amount of notes in its high lines compared to its lower lines. Section 4.2 will elucidate more on the interpretation of the other features and other genres. Furthermore some first remarks on melody and rhythm can already be made. The features marked with one asterisk are related to rhythm and the features marked with a double asterisk are related to melody. This shows that bossa nova has high correlations with features relating to both rhythm and melody, but which could use much refinement. For instance if not much melodic thirds are observed one might want to know which intervals then áre most likely to occur. This stresses the need for part 2 of the research. Verifying correlation results To show that the correlating features can indeed be regarded as being characteristic for their corresponding genre, two methods might be used. First the F1- score of the unsupervised clustering using only the selected features listed above can be computed. The results are shown in table 9. Bossa Nova Mambo Merengue Rumba Tango Table 9: F1-scores BAL using Bossa Nova correlating features. When comparing these results to the values given in table 3 it can be observed that the difference between bossa nova and the second best scoring genre has more than quadrupled. This means that the split between the bossa nova genre and the other genres is now rather significant. The fact that the F1 score for bossa nova is somewhat lower than in table 3 is understandable because bossa nova already got separated decently when all features were used. Because PCA could be used as a method of visualizing the data it can be used as a check on how the correlated features have split the data. Figure 6 shows the BAL corpus in PC space using only the features from table 8, and a strong separation between bossa nova and the other genres can clearly be observed. This and the foregoing check using F1 values are a strong indication that this particular feature subset characterizes just bossa nova. Because the goal is to find characteristics of each individual genre, this is a very appealing result. Principal Component Bossa Nova Mambo Merengue Rumba Tango Principal Component 1 Figure 6: BAL corpus in PC space with Bossa Nova correlating features. The top 10 features for each genre together with their F1 scores are included in appendix B and the reference scores for clusters using all features are in tables 3, 4, 5 and 6, 7. Results shown here for bossa nova in the BAL corpus generalize to almost all other genres in the other corpora used: The F1-score of the selected genre increases relative to the other genres when its correlating features are used when compared with clusters in which all features were used. Some genres (such as tango, table B) show 10

11 a strong relative increase while with others such as merengue (table B), this is less obvious, in the case of merengue also showing a big increase for mambo. Another anomaly is observed in table B, where the features correlating with bebop actually provide a more distinct clustering of swing. In these cases that correlating features fail to improve clustering results, these features might not provide a decent genre characterization and are therefore to be interpreted with care. 4.2 Interpretation Although quite some definitions of genres are published online in the Grove Music Online Dictionary [1], these texts do contain only a very small amount of information about specific genre characteristics and instead focus more on the social and cultural backgrounds. It is thus very valuable to work towards genre characterizations as to provide additional information on the understanding of genres. The following paragraphs provide a concise interpretation of the results of some genres; a full fledged interpretation is left to musicologists. The discussed results can be found in table 8 for bossa nova and the leftmost tables of appendix B for the other genres BOD main While the results hopefully present new knowledge to be used for more precise genre descriptions, findings matches with existing beliefs strengthen the validity of the method presented. Especially in the BOD main corpus, such familiar features are observed, likely because these broad genres capture global characteristics. Popular Table B shows the features correlating with the popular music in relation to jazz and classical music. Almost all features match with an intuitive description of popular music. Firstly the use of electric instruments and electric guitar in particular can be observed. Secondly popular music has a strong tonal focus and is likely to use tones only present in a certain scale, which is indicated by positive correlation with Nr. of common pitches, Most common pitch prevalence and negative correlation with Pitch variety. Furthermore a negative correlation with the range of highest line can be observed, which in pop music is often the melodic line which tends to be restricted in range. Jazz In jazz music, a significant correlation with saxophone and brass fraction can be observed. The presence of melodic tritones together with a high pitch variety is an indication of the more complex melodies often found in jazz. Classical Characteristics for the classical genre show the absence of electric instruments and percussion, and importance of high register, absence of importance of bass register and high value for primary register all three indicate that relative to pop and jazz, much is going on in the high lines BOD popular Hardcore Rap The hardcore rap genre is characterized by relatively long songs, where much movement in minor and major seconds is observed ( chromatic motion and stepwise motion ). Both negative note density and average time between attacks indicate that these songs are relatively slow paced. Punk Punk songs have a strong negative correlation with Duration meaning that they are often short songs. Furthermore they use a lot of repeated notes and arpeggiation. An interesting negative correlation that might not be so easily explained is that with Melodic thirds which seems to indicate that punk music lacks melodic thirds as opposed to hardcore rap and traditional country. Traditional Country Traditional country has the highest difference in melodic material of 11

12 all, indicated by pitch variety and pitch class variety. Furthermore a strong correlation with the use of melodic thirds can be observed BOD jazz The BOD jazz corpus provides the most difficult to interpret features and, as shown before, because the bebop features seemed to be unreliable, interpreting the results for the jazz corpus is left to others BOD classical Baroque Within the classical corpus, baroque stands out for its restricted range, restricted pitch variety and restricted melodic tritones. Together with most common melodic interval prevalence, indicating that there is a certain melodic interval that is often reoccurring, one can conclude that this genre is focused around tonality. Unfortunately no features concerning rhythm show up. Modern Classical Although for modern classical change of meter has a quite strong correlation, most other features show very low values. It is therefore difficult likely that these features are not very useful. Romantic The romantic genre seems to be characterized by rhythmic looseness, not only indicated by the feature with the same name but furthermore because of the lack of a strong second strongest rhythmic pulse BAL Bossa Nova As discussed earlier, bossa nova has a relative high amount of notes in the high lines. Contrasting though, there is also a high correlation with Importance of Bass Register. A strong positive correlation with average note duration and average time between attacks might be indicative of relatively slow music when compared to the other genres in the corpus. Furthermore a lack of melodic thirds shows up, which is a feature that might be verified by the second part of the research. Mambo The F1 score for mambo, although increased compared to table 3 is an indication that features for mambo will not be very reliable. Though, the occurrence of melodic fifths is something that again can be examined in part 2. Merengue The first feature for merengue voice equality - number of notes is described as Standard deviation of the total number of notes in each channel that contains at least one note [7] (p.66). A high positive correlation means that some instruments will play much notes while others will play very few. The overall note density is high, and the average note duration is negatively correlated, indicating many short notes. Rumba Rumba has a high correlation with both pitch variety and pitch class variety, indicating that it is likely that its either non-tonal or often uses modulation within a song. Furthermore it has quite a strong correlation with melodic thirds which will be inspected in part two. Tango A lack of unpitched instruments and percussion in general is observed. The use of polyrhythms is notable, together with the high pitch variety and chromatic motion, creating the expectation of interesting complex melodies. Both the use of chromatic motion as well as the lack of arpeggiation can be verified during part two. 5 Part 2 - Interval Sequences 5.1 Method Now that it has been shown that characteristic features can be extracted on a global level, 12

13 the attention in this part shifts to characteristic melodies obtained on the local level. It is known that rhythm plays a significant role at least in characterization of ball room dances [4] and it would be interesting to discover if they can be characterized by melody as well. Most songs however consist of multiple voices of which only one plays the melody. The reason that only melody voices will be analyzed is because the melody plays a dominant role and is often recalled the best, increasing the chance that genres can best be characterized by information extracted from melody. Furthermore multiple notes playing at the same time are rare in melodies, which eases determining by which notes it is formed. A melody consists of consecutive notes which have intervals in between them. Though, different melodies might map to the same interval (Both C to G and A to E have an interval of 7 in between them) so using these interval sequences captures the shape of the melodies instead of the exact pitches and key. This is similar to the way a derivative captures the movement of the original function while disregarding its vertical position Preprocessing Preprocessing the Corpus Because there was no corpus with prelabeled melodic lines that would work well in MATLAB it required some preprocessing of the existing corpora. This preprocessing was rather time consuming so melody extraction was only done for the BAL corpus and the BOD corpus was dropped. Selecting melodic lines was done through listening to the MIDI files and extracting the correct tracks by hand. In the case of melodies with more than one note playing at the same time, the upper voice was chosen because it is in general perceived as the main melody. For each genre, this time including salsa, 10 to 20 melodic lines were selected, depending on the available files. It might be possible to automate this process of melodic track selection, for instance with a method presented by Rizo [9]. Interval extraction From the previously extracted melodic lines, the intervals in semitones were calculated. Figure 7 shows such an example extraction for a random melody: the numbers indicate the intervals relative to the preceding note. Notice that an interval sequence of length n corresponds to a melody of length n+1. During this procedure, the rhythmic information about the melody is lost. This interval series is the data used for calculating the counts of the different micro-melodies. Interval series will be written between curly brackets like so {1, 2, -3, 1, 5, -1, -2} for the example sequence Figure 7: Melodic excerpt with relative pitch intervals The feature vector consists of interval sequences of length 1, 2 and 3. Taking any longer sequences is likely to provide very sparse results while increasing extraction time. In the feature vector all intervals from {-12} up to {12}, {-12,- 12} up to {12,12} and {-7,-7,-7} up to {7,7,7} are included. The triple interval sequences are only measuring a maximum distance of seven semitones (a perfect fifth) because computation would otherwise take rather long. All of the sequences can now be scored by counting their occurrences in the series of pitch intervals for each melody Interval Sequences As well as in the first section, it is possible to find interval sequences that explain a genre particularly well by calculating the point-biserial correlation. This is done in the same way as shown in part 1, grouping the melodies from one genre in one class

14 It is although possible that a particular sequence scores a high correlation because it only exists in one certain melody. Such a sequence should nevertheless not be seen as characteristic because it does not occur in multiple melodies pertaining to one genre. To account for this effect, besides a column indicating the sequence and correlation scores, a third column is added to the results. This column shows the percentage of occurrence of the sequence among all melodies of that genre. If now sequences can be found that score high on both measures, it is a strong indication that these sequences are characteristic sequences illustrative for the genre as a whole. It should be noted that the foregoing is also valid for sequences with a negative correlation and a low occurrence: negative correlation implies that a genre stands out from the others by lacking a particular sequence. Sequence Occurrence {-2,2,-2} % {-2,2} % {2,-2,2} % {2,0,-2} % {-1,-1,-3} % {-7} % {0,-2,2} % {-2,2,0} % {1} % {7,-2,2} % Table 10: Top 10 characteristic melodic sequences for bossa nova in BAL sorted on correlation value. Table 10 shows the 10 sequences with the highest correlation and occurrence percentages for bossa nova, the tables for the other genres are included as the odd numbered tables in appendix C. Section 5.2 provides an interpretation of these results Sorting on Occurrence Sorting the data on correlation stressed the melodic properties that are in particular not shared between genres. While this serves to find very characteristic sequences, results showed that for many sequences occurrence was low. That is why the same set of sequences is now sorted on occurrence, with the condition that the corresponding correlation has to be significant. Table 11 shows the occurrence sorted results for bossa nova. In the case of a sequences with negative correlation, an occurrence of 0% is the optimal result, which is why in the table the first 9 negatively correlating sequences score higher than the tenth (positively correlating) sequence. In appendix C the even numbered tables provide all occurrence sorted results. Sequence Occurrence {-6} % {-3, -3} % {9, -1} % {-2, 0, -3} % {8} % {-8} % {0, 4} % {0, -2, 0} % {-9} % {1} % Table 11: Top 10 characteristic melodic sequences for bossa nova in BAL sorted on highest occurrence. 5.2 Interpretation For each genre in the BAL corpus, the results as included in table 10 and 11 and appendix C will be discussed. The correlation sorted results show rather low scores in the occurrence column for most of the sequences, so albeit that these sequences have a strong correlation with a genre, they are not representative for most melodies of that genre. Therefore definitive conclusions about melodies in genres based on this data should be drawn with caution. Bossa Nova No other genre scores so many sequences with negative correlation in the occurrence sorted results as bossa nova. This is interesting because it indicates that melodic lines in bossa nova can be partly explained by what does 14

15 not happen in the music. Looking at the correlation sorted results, a lot of movement in major seconds ({-2}, {2}) is observed, especially in context of arpeggiation ({-2} followed by {2} or the other way around). Because these sequences of major seconds score rather high, the absence of sequence {0}, {-2}, {0} is remarkable. It indicates that a double note is rarely followed by a double note a minor second lower, strengthening the assumption that the major second is used extensively in arpeggiation. Furthermore there is a complete lack of descending motion in augmented fourths ({-6}, {- 3,-3}) and only a very small amount of movement in minor sixths ({-8}, {8}). The only result that shows up in both tables is the rising minor second ({1}), indicating that this interval is used a lot more in this genre than in the other genres, except for tango. Tango then again stands out because it incorporates chromatic movement, which does not seem to be a bossa nova characteristic. The results obtained in part one indicated a lack of melodic thirds ({3} and {4}) which is in complete accordance with part two: only a single {-3} can be found in the correlation sorted results, and only a single {4} can be found in the occurrence sorted results. Mambo Both tables show very low occurrence on all features, and correlations are rather low compared to other genres as well. The sequences also lack interesting recurring interval patterns. Therefore no reliable conclusions can be drawn regarding characteristic melodic lines. It can however be argued that the mambo melodies might contain the most diverse range of pitch sequences of all the ballroom dance melodies. Part one presented a mambo feature indicating melodic fifths {7}, while also noting that it would likely be unreliable. While the found interval sequences do show some ±{7} intervals, their occurrence is indeed too low to be reliable. Merengue This genre stands out by repeating notes more than the other genres: sequences of two and three repeating notes ({0}, {0, 0}) score high on both occurrence and correlation. These repeating notes are often accompanied by movement in major thirds ({4} or {-4}) and also on their own, the major thirds seem to stand out. The perfect fifth ({7}) also shows up in both tables which together with the major thirds might be indicative of movement over major chords. Finally there is a complete lack of downward chromatic motion ({-1,-1}), indicating that the melodic lines in merengue are likely to move in a more tonal way: minor and major scales have no consecutive minor seconds. Rumba As well as with mambo, occurrence and correlation are rather low for rumba. Though, a large amount of movement in ascending minor thirds ({3}) can be observed, as well as patterns indicative of modulation in minor seconds ({-4, 3}). This is in accordance with the results found in part one which indicated movement in melodic thirds ({3} and {4}). Because occurrence is so low for all sequences, absence of other characteristics is likely. Salsa The correlation sorted results for salsa are scoring too low to be really reliable. Some sequences are however showing some interesting movement: {-4, -3, -4} is a downwards movement over a major seventh chord, {-4, 2, 5} is moving over a major add two chord. Incorporating the occurrence sorted results, strong indication of movement in minor thirds ({-3}) can be observed. Besides this movement, there is a sequence with length three with also quite high correlation ({-1, -4, 2}) which moves over a piece of the standard major or minor scale. Opposing these quite strong characteristics are other frequently occurring sequences which are more difficult to explain in terms of logic melodic movement ({-2, 5}, {1, -3}). This could again be an indication that salsa is not so much characterized by specific pitch sequences. 15

16 Tango Of all ballroom dances, tango stands out for its extensive use of both ascending and descending minor seconds ({1}, {-1}) indicative of chromatic motion, a feature also found in part one of the research. Besides, almost all sequences use only movement in (minor and major) seconds, showing that tango lacks large melodic jumps. When played, most of the sequences sound like they are in a minor key and have a strong tendency to lead to a certain note. Especially the sequence {-2, -2, -1} stands out as a characteristic tango melody, with a downwards motion over a minor scale, over which the patterns {- 1, -2, 7} and {-4, 0, -1} are also moving. This might be an indication that tango melodies, at least the ones presented in this corpus, stand out by being written in minor keys. Another result presented in part one was the absence of arpeggiation. As arpeggiation is movement back and forth between two notes, it can be represented by two interval sequences {n, n} and { n, n} with n any number between 0 and 12. Out of both tables, only three sequences incorporate {-1, 1}. It is thus plausible that very few arpeggiation is occurring in tango songs. 6 Conclusion In this research, both global features and a selected set of small melodic sequences have been examined in order to set up a basis for understanding musical genre through computational means. Using PCA proved useful in visualization of the data and detecting outliers, while it could not be used for actual feature selection. Hereafter a measure for determining the amount of genre clustering and mix up between genres was devised using unsupervised k-means clustering which could then be used to verify correlation between genre and features. Part two used correlation between genres and interval sequences together with occurrence of these sequences among songs to find genre specific melodic structures. Both the global features and small melodic sequences proved to be useful in characterization of a diverse range of musical genres. Some genres, like merengue and mambo in part 1, showed much overlap with each other and seemed to be difficult to separate or characterize. This was though to be expected as some genres would be more likely characterized by other features (such as meta information). The extensive amounts of data provided through both parts of the research require careful further analysis by musicologists or other researchers with a decent musical background. The small amount of analysis performed in this report however served to show that the results indeed have an understandable and clearly interpretable meaning, that eventually might be used to expand understanding of musical genres and define them in a more formal way. Furthermore, as shown in the interpretation of part two, many of the global features related to melody found in part one could be seen again in part two on a more detailed level. This serves to show that global features can be used in exploration of interesting characteristics while methods such as interval sequences can then be used to refine the found global features. The methods presented can thus be used in future work for exploring and refining the global features related to rhythm. 7 Discussion Because the approach taken in this research was based on exploratory means of finding the characteristic features, there is much room for discussing the used methods and alternative approaches. This section will provide an overview of the alternatives for each part of the research as well as some general remarks, on which future work might be based. 16

17 7.1 Part 1 Inspecting PCA components In the principal component analysis presented in the first part of the research, the principal component axes are made up of a combination of all features. It might be interesting to see which features are the most important within these axes and to see what happens to the position of the clusters when important features are removed. Though contribution of the individual features to the PC axes was shown to be small, inspecting a combination of important features might help in getting a better understanding of characteristic features within one corpus. 7.2 Part 2 The use of counts of melodic interval sequences as features yields understandable information which, as was shown, can be used to describe a genre in more detail. There are however more possibilities for extracting features while making sure that these features remain clearly interpretable. Conditional chance Incorporating theory from the area of language processing, the individual pitch distances can be seen as an alphabet and sequences of these can be seen as n-grams. Because the sequences have been counted it is now possible to create an n-gram melodic sequence model from the data and gather data about conditional interval chance. Combining this knowledge with the results obtained in this research, on can, within a genre, further inspect high scoring sequences on which way they are likely to continue. These chance features might also be used in classification, but presumably only 2-grams can be used in this way because a single melodic line lacks enough information to give statistical accurate results for higher n-grams. The second moment In cryptography the second moment of the frequency distribution of letters in a piece of enciphered information can be used to determine in which language it was likely written. As above, the individual pitch distances can be seen as the alphabet (in the case of this research the alphabet incorporates all pitch distances from {-12} up to {12} but different values can be used). Now the second moment of the multinomial distributed intervals can be calculated using equation 1. Here f i is the frequency of the pitch interval {i} and N is the sum over all frequencies of all sequences. S 2 = {12} i={ 12} f i(f i 1) N(N 1) (1) The outcome of this formula is describing the way pitches are distributed within melodies and thus within genres. Whenever the pitches are distributed more equal, one expects to find a lower value for S 2 and whenever pitches are distributed unevenly, for instance in pentatonic scales where only 5 out of 12 pitches are frequently used, one expects to find a high value for S 2. Such a measure might thus be indicative of the amount of tonality and notes in a scale, and could also be used for classification purposes where the initial second moments for the genres should be calculated using a large training corpus. Multiple normal distributions Though certain melodic sequences can be seen as genre characteristics because they occur more in that particular genre, it is not likely that the other genres completely lack them. It is more likely that each particular melodic sequence count (each feature) has a different mean and standard deviation for each genre. Though much data is needed for accurate calculation, these figures then give a much clearer view of the numbers of melodic sequences in each genre. Furthermore, because these distributions also return a chance of a melody belonging to a genre given its melodic sequence counts, they might also be used for classification. 17

18 7.3 General One Versus One Correlation In both parts of the research, the classes used for calculating the correlation of the features with a particular genre were based on a one versus all genre grouping. The songs in the measured genre were said to be in class 1 and all the other songs belonged to class 0. While this serves to find strong characteristics between one genre and all other genres it might be that other interesting genre differences arise through calculation of the correlation between just one genre with one other genre, or one genre with some but not all of the other genres. Exploring alternative musical aspects The methods used in the second part of the research could easily be used to evaluate other feature sets derived from different musical aspects. One of the very obvious aspects is using counts of different sequences of note lengths or even complete rhythmic patterns of melodic lines. Another possibility would be using chord sequences which might be extracted through combining pitch information from multiple voices. On a somewhat larger scale, one could try to base counts or features on the overall structure of the song, a verse-chorus structure for example. This might also provide features describing the compressability of a song; in popular music there is much repetition, so the repeating parts can be compressed, while a modern classical piece is often uncompressable because it lacks repeating parts. References [1] Grove music online. oxfordmusiconline.com, June [2] Z. Cataltepe, Y. Yaslan, and A. Sonmez. Music genre classification using midi and audio features. EURASIP Journal on Applied Signal Processing, 1: , january [3] W. Chai and B. Vercoe. Folk music classification using hidden markov models. In Proc. of International Conference on Artificial Intelligence, [4] E. Chew, A. Volk, and C.-Y. Lee. Dance music classification using inner metric analysis. In B. Golden, S. Raghavan, and E. Wasil, editors, The Next Wave in Computing, Optimization, and Decision Technologies, volume 29 of Operations Research/- Computer Science Interfaces Series, pages Springer US, [5] D. Conklin. Melodic analysis with segment classes. Machine Learning, 65: , [6] A. Honingh, T. Weyde, and D. Conklin. Sequential association rules in atonal music. In Proc. of Mathematics and Computation in Music (MCM2009), [7] C. McKay. Automatic genre classification of midi recordings. Master s thesis, McGill University, Montreal, [8] C. Mckay and I. Fujinaga. Automatic genre classification using large high-level musical feature sets. In Int. Conf. on Music Information Retrieval, ISMIR, pages , [9] D. Rizo, P. J. P. D. Len, C. Prez-sancho, A. Pertusa, and J. M. Iesta. A pattern recognition approach for melody track selection in midi files. In Int. Conf. on Music Information Retrieval, ISMIR, pages 61 66, [10] X. Shao, C. Xu, and M. S. Kankanhalli. Unsupervised classification of music genre using hidden markov model. In IEEE International Conference on Multimedia and Expo, pages , [11] B. L. Sturm. Classification accuracy is not enough - on the evaluation of music genre 18

jsymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada

jsymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada jsymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada What is jsymbolic? Software that extracts statistical descriptors (called features ) from symbolic music files Can read: MIDI MEI (soon)

More information

jsymbolic 2: New Developments and Research Opportunities

jsymbolic 2: New Developments and Research Opportunities jsymbolic 2: New Developments and Research Opportunities Cory McKay Marianopolis College and CIRMMT Montreal, Canada 2 / 30 Topics Introduction to features (from a machine learning perspective) And how

More information

Analysis and Clustering of Musical Compositions using Melody-based Features

Analysis and Clustering of Musical Compositions using Melody-based Features Analysis and Clustering of Musical Compositions using Melody-based Features Isaac Caswell Erika Ji December 13, 2013 Abstract This paper demonstrates that melodic structure fundamentally differentiates

More information

Sequential Association Rules in Atonal Music

Sequential Association Rules in Atonal Music Sequential Association Rules in Atonal Music Aline Honingh, Tillman Weyde, and Darrell Conklin Music Informatics research group Department of Computing City University London Abstract. This paper describes

More information

Sequential Association Rules in Atonal Music

Sequential Association Rules in Atonal Music Sequential Association Rules in Atonal Music Aline Honingh, Tillman Weyde and Darrell Conklin Music Informatics research group Department of Computing City University London Abstract. This paper describes

More information

UvA-DARE (Digital Academic Repository) Clustering and classification of music using interval categories Honingh, A.K.; Bod, L.W.M.

UvA-DARE (Digital Academic Repository) Clustering and classification of music using interval categories Honingh, A.K.; Bod, L.W.M. UvA-DARE (Digital Academic Repository) Clustering and classification of music using interval categories Honingh, A.K.; Bod, L.W.M. Published in: Mathematics and Computation in Music DOI:.07/978-3-642-21590-2_

More information

Audio Feature Extraction for Corpus Analysis

Audio Feature Extraction for Corpus Analysis Audio Feature Extraction for Corpus Analysis Anja Volk Sound and Music Technology 5 Dec 2017 1 Corpus analysis What is corpus analysis study a large corpus of music for gaining insights on general trends

More information

Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals

Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals Eita Nakamura and Shinji Takaki National Institute of Informatics, Tokyo 101-8430, Japan eita.nakamura@gmail.com, takaki@nii.ac.jp

More information

Creating a Feature Vector to Identify Similarity between MIDI Files

Creating a Feature Vector to Identify Similarity between MIDI Files Creating a Feature Vector to Identify Similarity between MIDI Files Joseph Stroud 2017 Honors Thesis Advised by Sergio Alvarez Computer Science Department, Boston College 1 Abstract Today there are many

More information

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You Chris Lewis Stanford University cmslewis@stanford.edu Abstract In this project, I explore the effectiveness of the Naive Bayes Classifier

More information

Outline. Why do we classify? Audio Classification

Outline. Why do we classify? Audio Classification Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification Implementation Future Work Why do we classify

More information

Specifying Features for Classical and Non-Classical Melody Evaluation

Specifying Features for Classical and Non-Classical Melody Evaluation Specifying Features for Classical and Non-Classical Melody Evaluation Andrei D. Coronel Ateneo de Manila University acoronel@ateneo.edu Ariel A. Maguyon Ateneo de Manila University amaguyon@ateneo.edu

More information

CSC475 Music Information Retrieval

CSC475 Music Information Retrieval CSC475 Music Information Retrieval Symbolic Music Representations George Tzanetakis University of Victoria 2014 G. Tzanetakis 1 / 30 Table of Contents I 1 Western Common Music Notation 2 Digital Formats

More information

CS229 Project Report Polyphonic Piano Transcription

CS229 Project Report Polyphonic Piano Transcription CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project

More information

Computational Modelling of Harmony

Computational Modelling of Harmony Computational Modelling of Harmony Simon Dixon Centre for Digital Music, Queen Mary University of London, Mile End Rd, London E1 4NS, UK simon.dixon@elec.qmul.ac.uk http://www.elec.qmul.ac.uk/people/simond

More information

Hidden Markov Model based dance recognition

Hidden Markov Model based dance recognition Hidden Markov Model based dance recognition Dragutin Hrenek, Nenad Mikša, Robert Perica, Pavle Prentašić and Boris Trubić University of Zagreb, Faculty of Electrical Engineering and Computing Unska 3,

More information

Composer Style Attribution

Composer Style Attribution Composer Style Attribution Jacqueline Speiser, Vishesh Gupta Introduction Josquin des Prez (1450 1521) is one of the most famous composers of the Renaissance. Despite his fame, there exists a significant

More information

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the

More information

The Human Features of Music.

The Human Features of Music. The Human Features of Music. Bachelor Thesis Artificial Intelligence, Social Studies, Radboud University Nijmegen Chris Kemper, s4359410 Supervisor: Makiko Sadakata Artificial Intelligence, Social Studies,

More information

Feature-Based Analysis of Haydn String Quartets

Feature-Based Analysis of Haydn String Quartets Feature-Based Analysis of Haydn String Quartets Lawson Wong 5/5/2 Introduction When listening to multi-movement works, amateur listeners have almost certainly asked the following situation : Am I still

More information

Varying Degrees of Difficulty in Melodic Dictation Examples According to Intervallic Content

Varying Degrees of Difficulty in Melodic Dictation Examples According to Intervallic Content University of Tennessee, Knoxville Trace: Tennessee Research and Creative Exchange Masters Theses Graduate School 8-2012 Varying Degrees of Difficulty in Melodic Dictation Examples According to Intervallic

More information

MUSI-6201 Computational Music Analysis

MUSI-6201 Computational Music Analysis MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)

More information

Enhancing Music Maps

Enhancing Music Maps Enhancing Music Maps Jakob Frank Vienna University of Technology, Vienna, Austria http://www.ifs.tuwien.ac.at/mir frank@ifs.tuwien.ac.at Abstract. Private as well as commercial music collections keep growing

More information

Modeling memory for melodies

Modeling memory for melodies Modeling memory for melodies Daniel Müllensiefen 1 and Christian Hennig 2 1 Musikwissenschaftliches Institut, Universität Hamburg, 20354 Hamburg, Germany 2 Department of Statistical Science, University

More information

Music Radar: A Web-based Query by Humming System

Music Radar: A Web-based Query by Humming System Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,

More information

Articulation Clarity and distinct rendition in musical performance.

Articulation Clarity and distinct rendition in musical performance. Maryland State Department of Education MUSIC GLOSSARY A hyperlink to Voluntary State Curricula ABA Often referenced as song form, musical structure with a beginning section, followed by a contrasting section,

More information

Detecting Musical Key with Supervised Learning

Detecting Musical Key with Supervised Learning Detecting Musical Key with Supervised Learning Robert Mahieu Department of Electrical Engineering Stanford University rmahieu@stanford.edu Abstract This paper proposes and tests performance of two different

More information

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES 12th International Society for Music Information Retrieval Conference (ISMIR 2011) A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES Erdem Unal 1 Elaine Chew 2 Panayiotis Georgiou

More information

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? NICHOLAS BORG AND GEORGE HOKKANEN Abstract. The possibility of a hit song prediction algorithm is both academically interesting and industry motivated.

More information

HUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH

HUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH Proc. of the th Int. Conference on Digital Audio Effects (DAFx-), Hamburg, Germany, September -8, HUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH George Tzanetakis, Georg Essl Computer

More information

SIMSSA DB: A Database for Computational Musicological Research

SIMSSA DB: A Database for Computational Musicological Research SIMSSA DB: A Database for Computational Musicological Research Cory McKay Marianopolis College 2018 International Association of Music Libraries, Archives and Documentation Centres International Congress,

More information

PITCH CLASS SET CATEGORIES AS ANALYSIS TOOLS FOR DEGREES OF TONALITY

PITCH CLASS SET CATEGORIES AS ANALYSIS TOOLS FOR DEGREES OF TONALITY PITCH CLASS SET CATEGORIES AS ANALYSIS TOOLS FOR DEGREES OF TONALITY Aline Honingh Rens Bod Institute for Logic, Language and Computation University of Amsterdam {A.K.Honingh,Rens.Bod}@uva.nl ABSTRACT

More information

Melody classification using patterns

Melody classification using patterns Melody classification using patterns Darrell Conklin Department of Computing City University London United Kingdom conklin@city.ac.uk Abstract. A new method for symbolic music classification is proposed,

More information

LESSON 1 PITCH NOTATION AND INTERVALS

LESSON 1 PITCH NOTATION AND INTERVALS FUNDAMENTALS I 1 Fundamentals I UNIT-I LESSON 1 PITCH NOTATION AND INTERVALS Sounds that we perceive as being musical have four basic elements; pitch, loudness, timbre, and duration. Pitch is the relative

More information

Automatic Piano Music Transcription

Automatic Piano Music Transcription Automatic Piano Music Transcription Jianyu Fan Qiuhan Wang Xin Li Jianyu.Fan.Gr@dartmouth.edu Qiuhan.Wang.Gr@dartmouth.edu Xi.Li.Gr@dartmouth.edu 1. Introduction Writing down the score while listening

More information

Automatic Music Clustering using Audio Attributes

Automatic Music Clustering using Audio Attributes Automatic Music Clustering using Audio Attributes Abhishek Sen BTech (Electronics) Veermata Jijabai Technological Institute (VJTI), Mumbai, India abhishekpsen@gmail.com Abstract Music brings people together,

More information

Exploring the Design Space of Symbolic Music Genre Classification Using Data Mining Techniques Ortiz-Arroyo, Daniel; Kofod, Christian

Exploring the Design Space of Symbolic Music Genre Classification Using Data Mining Techniques Ortiz-Arroyo, Daniel; Kofod, Christian Aalborg Universitet Exploring the Design Space of Symbolic Music Genre Classification Using Data Mining Techniques Ortiz-Arroyo, Daniel; Kofod, Christian Published in: International Conference on Computational

More information

METRICAL STRENGTH AND CONTRADICTION IN TURKISH MAKAM MUSIC

METRICAL STRENGTH AND CONTRADICTION IN TURKISH MAKAM MUSIC Proc. of the nd CompMusic Workshop (Istanbul, Turkey, July -, ) METRICAL STRENGTH AND CONTRADICTION IN TURKISH MAKAM MUSIC Andre Holzapfel Music Technology Group Universitat Pompeu Fabra Barcelona, Spain

More information

Chord Classification of an Audio Signal using Artificial Neural Network

Chord Classification of an Audio Signal using Artificial Neural Network Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------

More information

Melody Retrieval On The Web

Melody Retrieval On The Web Melody Retrieval On The Web Thesis proposal for the degree of Master of Science at the Massachusetts Institute of Technology M.I.T Media Laboratory Fall 2000 Thesis supervisor: Barry Vercoe Professor,

More information

A Pattern Recognition Approach for Melody Track Selection in MIDI Files

A Pattern Recognition Approach for Melody Track Selection in MIDI Files A Pattern Recognition Approach for Melody Track Selection in MIDI Files David Rizo, Pedro J. Ponce de León, Carlos Pérez-Sancho, Antonio Pertusa, José M. Iñesta Departamento de Lenguajes y Sistemas Informáticos

More information

Semi-supervised Musical Instrument Recognition

Semi-supervised Musical Instrument Recognition Semi-supervised Musical Instrument Recognition Master s Thesis Presentation Aleksandr Diment 1 1 Tampere niversity of Technology, Finland Supervisors: Adj.Prof. Tuomas Virtanen, MSc Toni Heittola 17 May

More information

GCSE MUSIC UNIT 3 APPRAISING. Mock Assessment Materials NOVEMBER hour approximately

GCSE MUSIC UNIT 3 APPRAISING. Mock Assessment Materials NOVEMBER hour approximately Candidate Name Centre Number Candidate Number GCSE MUSIC UNIT 3 APPRAISING Mock Assessment Materials NOVEMBER 2017 1 hour approximately Examiners Use Only Question Max Mark 1 9 2 9 3 9 4 9 5 9 6 9 7 9

More information

PLANE TESSELATION WITH MUSICAL-SCALE TILES AND BIDIMENSIONAL AUTOMATIC COMPOSITION

PLANE TESSELATION WITH MUSICAL-SCALE TILES AND BIDIMENSIONAL AUTOMATIC COMPOSITION PLANE TESSELATION WITH MUSICAL-SCALE TILES AND BIDIMENSIONAL AUTOMATIC COMPOSITION ABSTRACT We present a method for arranging the notes of certain musical scales (pentatonic, heptatonic, Blues Minor and

More information

CHAPTER 3. Melody Style Mining

CHAPTER 3. Melody Style Mining CHAPTER 3 Melody Style Mining 3.1 Rationale Three issues need to be considered for melody mining and classification. One is the feature extraction of melody. Another is the representation of the extracted

More information

Automatic Rhythmic Notation from Single Voice Audio Sources

Automatic Rhythmic Notation from Single Voice Audio Sources Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung

More information

Analysis of local and global timing and pitch change in ordinary

Analysis of local and global timing and pitch change in ordinary Alma Mater Studiorum University of Bologna, August -6 6 Analysis of local and global timing and pitch change in ordinary melodies Roger Watt Dept. of Psychology, University of Stirling, Scotland r.j.watt@stirling.ac.uk

More information

MELONET I: Neural Nets for Inventing Baroque-Style Chorale Variations

MELONET I: Neural Nets for Inventing Baroque-Style Chorale Variations MELONET I: Neural Nets for Inventing Baroque-Style Chorale Variations Dominik Hornel dominik@ira.uka.de Institut fur Logik, Komplexitat und Deduktionssysteme Universitat Fridericiana Karlsruhe (TH) Am

More information

2010 Music Solo Performance GA 3: Aural and written examination

2010 Music Solo Performance GA 3: Aural and written examination 2010 Music Solo Performance GA 3: Aural and written examination GENERAL COMMENTS The 2010 Music Solo Performance aural and written examination consisted of three sections and was worth 105 marks. All sections

More information

Robert Alexandru Dobre, Cristian Negrescu

Robert Alexandru Dobre, Cristian Negrescu ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Automatic Music Transcription Software Based on Constant Q

More information

Pitch Spelling Algorithms

Pitch Spelling Algorithms Pitch Spelling Algorithms David Meredith Centre for Computational Creativity Department of Computing City University, London dave@titanmusic.com www.titanmusic.com MaMuX Seminar IRCAM, Centre G. Pompidou,

More information

Supervised Learning in Genre Classification

Supervised Learning in Genre Classification Supervised Learning in Genre Classification Introduction & Motivation Mohit Rajani and Luke Ekkizogloy {i.mohit,luke.ekkizogloy}@gmail.com Stanford University, CS229: Machine Learning, 2009 Now that music

More information

An Integrated Music Chromaticism Model

An Integrated Music Chromaticism Model An Integrated Music Chromaticism Model DIONYSIOS POLITIS and DIMITRIOS MARGOUNAKIS Dept. of Informatics, School of Sciences Aristotle University of Thessaloniki University Campus, Thessaloniki, GR-541

More information

Automatic Laughter Detection

Automatic Laughter Detection Automatic Laughter Detection Mary Knox Final Project (EECS 94) knoxm@eecs.berkeley.edu December 1, 006 1 Introduction Laughter is a powerful cue in communication. It communicates to listeners the emotional

More information

Music Composition with RNN

Music Composition with RNN Music Composition with RNN Jason Wang Department of Statistics Stanford University zwang01@stanford.edu Abstract Music composition is an interesting problem that tests the creativity capacities of artificial

More information

Study Guide. Solutions to Selected Exercises. Foundations of Music and Musicianship with CD-ROM. 2nd Edition. David Damschroder

Study Guide. Solutions to Selected Exercises. Foundations of Music and Musicianship with CD-ROM. 2nd Edition. David Damschroder Study Guide Solutions to Selected Exercises Foundations of Music and Musicianship with CD-ROM 2nd Edition by David Damschroder Solutions to Selected Exercises 1 CHAPTER 1 P1-4 Do exercises a-c. Remember

More information

Singer Recognition and Modeling Singer Error

Singer Recognition and Modeling Singer Error Singer Recognition and Modeling Singer Error Johan Ismael Stanford University jismael@stanford.edu Nicholas McGee Stanford University ndmcgee@stanford.edu 1. Abstract We propose a system for recognizing

More information

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Róisín Loughran roisin.loughran@ul.ie Jacqueline Walker jacqueline.walker@ul.ie Michael O Neill University

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

Improving Frame Based Automatic Laughter Detection

Improving Frame Based Automatic Laughter Detection Improving Frame Based Automatic Laughter Detection Mary Knox EE225D Class Project knoxm@eecs.berkeley.edu December 13, 2007 Abstract Laughter recognition is an underexplored area of research. My goal for

More information

Harmonic syntax and high-level statistics of the songs of three early Classical composers

Harmonic syntax and high-level statistics of the songs of three early Classical composers Harmonic syntax and high-level statistics of the songs of three early Classical composers Wendy de Heer Electrical Engineering and Computer Sciences University of California at Berkeley Technical Report

More information

Music Segmentation Using Markov Chain Methods

Music Segmentation Using Markov Chain Methods Music Segmentation Using Markov Chain Methods Paul Finkelstein March 8, 2011 Abstract This paper will present just how far the use of Markov Chains has spread in the 21 st century. We will explain some

More information

Can Song Lyrics Predict Genre? Danny Diekroeger Stanford University

Can Song Lyrics Predict Genre? Danny Diekroeger Stanford University Can Song Lyrics Predict Genre? Danny Diekroeger Stanford University danny1@stanford.edu 1. Motivation and Goal Music has long been a way for people to express their emotions. And because we all have a

More information

Style-independent computer-assisted exploratory analysis of large music collections

Style-independent computer-assisted exploratory analysis of large music collections Style-independent computer-assisted exploratory analysis of large music collections Abstract Cory McKay Schulich School of Music McGill University Montreal, Quebec, Canada cory.mckay@mail.mcgill.ca The

More information

MUSIC CURRICULM MAP: KEY STAGE THREE:

MUSIC CURRICULM MAP: KEY STAGE THREE: YEAR SEVEN MUSIC CURRICULM MAP: KEY STAGE THREE: 2013-2015 ONE TWO THREE FOUR FIVE Understanding the elements of music Understanding rhythm and : Performing Understanding rhythm and : Composing Understanding

More information

A repetition-based framework for lyric alignment in popular songs

A repetition-based framework for lyric alignment in popular songs A repetition-based framework for lyric alignment in popular songs ABSTRACT LUONG Minh Thang and KAN Min Yen Department of Computer Science, School of Computing, National University of Singapore We examine

More information

Music Complexity Descriptors. Matt Stabile June 6 th, 2008

Music Complexity Descriptors. Matt Stabile June 6 th, 2008 Music Complexity Descriptors Matt Stabile June 6 th, 2008 Musical Complexity as a Semantic Descriptor Modern digital audio collections need new criteria for categorization and searching. Applicable to:

More information

MUSIC (MUS) Music (MUS) 1

MUSIC (MUS) Music (MUS) 1 Music (MUS) 1 MUSIC (MUS) MUS 2 Music Theory 3 Units (Degree Applicable, CSU, UC, C-ID #: MUS 120) Corequisite: MUS 5A Preparation for the study of harmony and form as it is practiced in Western tonal

More information

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring 2009 Week 6 Class Notes Pitch Perception Introduction Pitch may be described as that attribute of auditory sensation in terms

More information

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION ULAŞ BAĞCI AND ENGIN ERZIN arxiv:0907.3220v1 [cs.sd] 18 Jul 2009 ABSTRACT. Music genre classification is an essential tool for

More information

Automatic Laughter Detection

Automatic Laughter Detection Automatic Laughter Detection Mary Knox 1803707 knoxm@eecs.berkeley.edu December 1, 006 Abstract We built a system to automatically detect laughter from acoustic features of audio. To implement the system,

More information

Machine Learning Term Project Write-up Creating Models of Performers of Chopin Mazurkas

Machine Learning Term Project Write-up Creating Models of Performers of Chopin Mazurkas Machine Learning Term Project Write-up Creating Models of Performers of Chopin Mazurkas Marcello Herreshoff In collaboration with Craig Sapp (craig@ccrma.stanford.edu) 1 Motivation We want to generative

More information

Predicting Variation of Folk Songs: A Corpus Analysis Study on the Memorability of Melodies Janssen, B.D.; Burgoyne, J.A.; Honing, H.J.

Predicting Variation of Folk Songs: A Corpus Analysis Study on the Memorability of Melodies Janssen, B.D.; Burgoyne, J.A.; Honing, H.J. UvA-DARE (Digital Academic Repository) Predicting Variation of Folk Songs: A Corpus Analysis Study on the Memorability of Melodies Janssen, B.D.; Burgoyne, J.A.; Honing, H.J. Published in: Frontiers in

More information

Music Genre Classification

Music Genre Classification Music Genre Classification chunya25 Fall 2017 1 Introduction A genre is defined as a category of artistic composition, characterized by similarities in form, style, or subject matter. [1] Some researchers

More information

2014 Music Style and Composition GA 3: Aural and written examination

2014 Music Style and Composition GA 3: Aural and written examination 2014 Music Style and Composition GA 3: Aural and written examination GENERAL COMMENTS The 2014 Music Style and Composition examination consisted of two sections, worth a total of 100 marks. Both sections

More information

Statistical Modeling and Retrieval of Polyphonic Music

Statistical Modeling and Retrieval of Polyphonic Music Statistical Modeling and Retrieval of Polyphonic Music Erdem Unal Panayiotis G. Georgiou and Shrikanth S. Narayanan Speech Analysis and Interpretation Laboratory University of Southern California Los Angeles,

More information

A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES

A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES Panayiotis Kokoras School of Music Studies Aristotle University of Thessaloniki email@panayiotiskokoras.com Abstract. This article proposes a theoretical

More information

Computer Coordination With Popular Music: A New Research Agenda 1

Computer Coordination With Popular Music: A New Research Agenda 1 Computer Coordination With Popular Music: A New Research Agenda 1 Roger B. Dannenberg roger.dannenberg@cs.cmu.edu http://www.cs.cmu.edu/~rbd School of Computer Science Carnegie Mellon University Pittsburgh,

More information

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t MPEG-7 FOR CONTENT-BASED MUSIC PROCESSING Λ Emilia GÓMEZ, Fabien GOUYON, Perfecto HERRERA and Xavier AMATRIAIN Music Technology Group, Universitat Pompeu Fabra, Barcelona, SPAIN http://www.iua.upf.es/mtg

More information

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 AN HMM BASED INVESTIGATION OF DIFFERENCES BETWEEN MUSICAL INSTRUMENTS OF THE SAME TYPE PACS: 43.75.-z Eichner, Matthias; Wolff, Matthias;

More information

arxiv: v1 [cs.sd] 8 Jun 2016

arxiv: v1 [cs.sd] 8 Jun 2016 Symbolic Music Data Version 1. arxiv:1.5v1 [cs.sd] 8 Jun 1 Christian Walder CSIRO Data1 7 London Circuit, Canberra,, Australia. christian.walder@data1.csiro.au June 9, 1 Abstract In this document, we introduce

More information

ILLINOIS LICENSURE TESTING SYSTEM

ILLINOIS LICENSURE TESTING SYSTEM ILLINOIS LICENSURE TESTING SYSTEM FIELD 212: MUSIC January 2017 Effective beginning September 3, 2018 ILLINOIS LICENSURE TESTING SYSTEM FIELD 212: MUSIC January 2017 Subarea Range of Objectives I. Responding:

More information

Subjective Similarity of Music: Data Collection for Individuality Analysis

Subjective Similarity of Music: Data Collection for Individuality Analysis Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp

More information

Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors *

Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors * Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors * David Ortega-Pacheco and Hiram Calvo Centro de Investigación en Computación, Instituto Politécnico Nacional, Av. Juan

More information

Music Theory. Fine Arts Curriculum Framework. Revised 2008

Music Theory. Fine Arts Curriculum Framework. Revised 2008 Music Theory Fine Arts Curriculum Framework Revised 2008 Course Title: Music Theory Course/Unit Credit: 1 Course Number: Teacher Licensure: Grades: 9-12 Music Theory Music Theory is a two-semester course

More information

Automated extraction of motivic patterns and application to the analysis of Debussy s Syrinx

Automated extraction of motivic patterns and application to the analysis of Debussy s Syrinx Automated extraction of motivic patterns and application to the analysis of Debussy s Syrinx Olivier Lartillot University of Jyväskylä, Finland lartillo@campus.jyu.fi 1. General Framework 1.1. Motivic

More information

Evaluating Melodic Encodings for Use in Cover Song Identification

Evaluating Melodic Encodings for Use in Cover Song Identification Evaluating Melodic Encodings for Use in Cover Song Identification David D. Wickland wickland@uoguelph.ca David A. Calvert dcalvert@uoguelph.ca James Harley jharley@uoguelph.ca ABSTRACT Cover song identification

More information

Supporting Information

Supporting Information Supporting Information I. DATA Discogs.com is a comprehensive, user-built music database with the aim to provide crossreferenced discographies of all labels and artists. As of April 14, more than 189,000

More information

Introduction to Set Theory by Stephen Taylor

Introduction to Set Theory by Stephen Taylor Introduction to Set Theory by Stephen Taylor http://composertools.com/tools/pcsets/setfinder.html 1. Pitch Class The 12 notes of the chromatic scale, independent of octaves. C is the same pitch class,

More information

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis Fengyan Wu fengyanyy@163.com Shutao Sun stsun@cuc.edu.cn Weiyao Xue Wyxue_std@163.com Abstract Automatic extraction of

More information

Building a Better Bach with Markov Chains

Building a Better Bach with Markov Chains Building a Better Bach with Markov Chains CS701 Implementation Project, Timothy Crocker December 18, 2015 1 Abstract For my implementation project, I explored the field of algorithmic music composition

More information

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu

More information

Some properties of non-octave-repeating scales, and why composers might care

Some properties of non-octave-repeating scales, and why composers might care Some properties of non-octave-repeating scales, and why composers might care Craig Weston How to cite this presentation If you make reference to this version of the manuscript, use the following information:

More information

Partimenti Pedagogy at the European American Musical Alliance, Derek Remeš

Partimenti Pedagogy at the European American Musical Alliance, Derek Remeš Partimenti Pedagogy at the European American Musical Alliance, 2009-2010 Derek Remeš The following document summarizes the method of teaching partimenti (basses et chants donnés) at the European American

More information

The purpose of this essay is to impart a basic vocabulary that you and your fellow

The purpose of this essay is to impart a basic vocabulary that you and your fellow Music Fundamentals By Benjamin DuPriest The purpose of this essay is to impart a basic vocabulary that you and your fellow students can draw on when discussing the sonic qualities of music. Excursions

More information

Influence of timbre, presence/absence of tonal hierarchy and musical training on the perception of musical tension and relaxation schemas

Influence of timbre, presence/absence of tonal hierarchy and musical training on the perception of musical tension and relaxation schemas Influence of timbre, presence/absence of tonal hierarchy and musical training on the perception of musical and schemas Stella Paraskeva (,) Stephen McAdams (,) () Institut de Recherche et de Coordination

More information

EE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function

EE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function EE391 Special Report (Spring 25) Automatic Chord Recognition Using A Summary Autocorrelation Function Advisor: Professor Julius Smith Kyogu Lee Center for Computer Research in Music and Acoustics (CCRMA)

More information

Student Performance Q&A:

Student Performance Q&A: Student Performance Q&A: 2004 AP Music Theory Free-Response Questions The following comments on the 2004 free-response questions for AP Music Theory were written by the Chief Reader, Jo Anne F. Caputo

More information

10 Lessons In Jazz Improvisation By Mike Steinel University of North Texas

10 Lessons In Jazz Improvisation By Mike Steinel University of North Texas 10 Lessons In Jazz Improvisation By Mike Steinel University of North Texas Michael.steinel@unt.edu Sponsored by Hal Leonard Corporation And Yamaha Musical Instruments 10 Basic Lessons #1 - You Gotta Love

More information

6.UAP Project. FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System. Daryl Neubieser. May 12, 2016

6.UAP Project. FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System. Daryl Neubieser. May 12, 2016 6.UAP Project FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System Daryl Neubieser May 12, 2016 Abstract: This paper describes my implementation of a variable-speed accompaniment system that

More information