Autocorrelation in meter induction: The role of accent structure a)

Size: px
Start display at page:

Download "Autocorrelation in meter induction: The role of accent structure a)"

Transcription

1 Autocorrelation in meter induction: The role of accent structure a) Petri Toiviainen and Tuomas Eerola Department of Music, P.O. Box 35(M), University of Jyväskylä, Jyväskylä, Finland Received 16 April 2005; revised 14 October 2005; accepted 7 November 2005 The performance of autocorrelation-based meter induction was tested with two large collections of folk melodies, consisting of approximately melodies for which the correct meters were available. The performance was measured by the proportion of melodies whose meter was correctly classified by a discriminant function. Furthermore, it was examined whether including different melodic accent types would improve the classification performance. By determining the components of the autocorrelation functions that were significant in the classification it was found that periodicity in note onset locations was the most important cue for the determination of meter. Of the melodic accents included, Thomassen s melodic accent was found to provide the most reliable cues for the determination of meter. The discriminant function analyses suggested that periodicities longer than one measure may provide cues for meter determination that are more reliable than shorter periodicities. Overall, the method predicted notated meter with an accuracy reaching 96% for binary classification and 75% for classification into nine categories of meter Acoustical Society of America. DOI: / PACS number s : Cd DD Pages: I. INTRODUCTION Most music is organized to contain temporal periodicities that evoke a percept of regularly occurring pulses or beats. The period of the most salient pulse is typically within the range of 400 to 900 ms Fraisse, 1982; Parncutt, 1994; van Noorden and Moelants, The perceived pulses are often hierarchically organized and consist of at least two simultaneous levels whose periods have an integer ratio. This gives rise to a percept of regularly alternating strong and weak beats, a phenomenon referred to as meter Cooper and Meyer, 1960; Lerdahl and Jackendoff, In Western music, the ratio of the pulse lengths is usually limited to 1:2 duple meter and 1:3 triple meter. Meter in which each beat has three subdivisions, such as 6/8 or 9/8, is referred to as compound meter. A number of computational models have been developed for the extraction of the basic pulse from music. Modeling of meter perception has, however, received less attention. Large and Kolen 1994 presented a model of meter perception based on resonating oscillators. Toiviainen 1997 presented a model of competing subharmonic oscillators for determining the meter duple versus triple from an acoustical representation of music. Brown 1993 proposed a method for determining the meter of musical scores by applying autocorrelation to a temporal function consisting of impulses at each tone onset whose heights are weighted by the respective tone durations. A shortcoming of Brown s 1993 study is a Portions of this work were presented in The role of accent periodicities in meter induction: a classification study, Proceedings of the 8th International Conference on Music Perception and Cognition, Evanston, IL, August 2004; and Classification of musical metre with autocorrelation and discriminant functions, Proceedings of the 6th International Conference on Music Information Retrieval, London, that she does not provide any explicit criteria for the determination of meter from the autocorrelation functions. Although there is evidence that the pitch information present in music may affect the perception of pulse and meter Dawe et al., 1993; Thomassen, 1982; Hannon et al., 2004, most models of pulse finding developed to date rely only on note onset times and durations. Dixon and Cambouropoulos 2000, however, proposed a multi-agent model for beat tracking that makes use of pitch and amplitude information. They found that including this information when determining the salience of notes significantly improved the performance of their model. Vos et al., 1994 applied autocorrelation to the determination of meter in isochronous or almost isochronous music. They utilized a method similar to that proposed by Brown 1993, except for using the melodic intervals between subsequent tones instead of tone durations as the weighting factor in the autocorrelation analysis. Using a corpus of 30 compositions by J. S. Bach, they found that the maxima of the obtained autocorrelation functions matched the respective bar lengths as indicated in the musical score. As the majority of music is nonisochronous, reliance on mere melodic interval structure in meter induction is probably a special case. In a general case, it can be expected that meter determination be based on both temporal and pitch structure. To address the question of the relative importance of temporal and pitch structure in meter determination, it would be necessary to use a rhythmically more variable set of stimuli. According to a commonly adopted notion, meter can be inferred from phenomenal, structural, and metrical accents Lerdahl and Jackendoff, Phenomenal accents are the primary source of meter and are related to the surface structure of music and arise from changes in duration, pitch, timbre, and dynamics. Pitch-related phenomenal accents, more 1164 J. Acoust. Soc. Am , February /2006/119 2 /1164/7/$ Acoustical Society of America

2 commonly referred to as melodic accents, arise from changes in pitch height, pitch interval or pitch contour Huron and Royal, Using a correlational study, Huron and Royal 1996 investigated the extent to which different types of pitch-related accent correspond to respective metrical positions as indicated by musical notation. Of eight different types of melodic accent, they found that the empirically derived accent by Thomassen 1982 was the only one to correlate significantly with metric position. The lack of correlation of the other accent types however, does, not exclude the possibility that various metric accents exhibit periodic structure that may serve as an additional cue for meter induction. For instance, different types of accents may contain periodicities of equal length, while being phase shifted with respect to each other. This is the case, for instance, with the accents 1, 4, and 5 in Fig. 3. It must be noted that temporal structure and pitch information are not the sole determinants of meter, as it is also influenced by other features such as phrasing and lyrics as well as accents introduced by the performer. These aspects are, nonetheless, beyond the scope of the present study. A shortcoming of meter induction models presented to date is that they have not been evaluated with large sets of musical material. While Vos et al utilized a corpus of 30 compositions, Brown 1993 presented only a handful of short musical excerpts to visualize the performance of her model. The present study investigated the performance of autocorrelation-based meter induction with large collections of folk melodies consisting of thousands of items, for which the notated meters are available. The performance was assessed by the proportion of melodies that were correctly classified in terms of their meter. The components of the autocorrelation function that are significant in the classification were determined. Moreover, a number of different types of melodic accent and combinations thereof were applied to the classification to assess the significance of each of them in the induction of meter. Finally, confusions made by the algorithm between different types of meter were investigated in detail. II. AUTOCORRELATION AND METER Below, the method for constructing the autocorrelation function for meter induction is described. For the original description, see Brown Let the melody consist of N notes with onset times t i,i=1,2,...,n. Each note is associated with an accent value a i,i=1,2,...,n; in Brown 1993, a i equals the duration of the respective note. The onset impulse function f is a time series consisting of impulses of height a i located at each note onset position: N f n = a i i n, n = 0,1,2,..., 1 i=1 where = i n 1, n = t i/dt, 2 0, otherwise, where dt denotes the sampling interval and denotes rounding to the nearest integer. FIG. 1. Excerpt from a melody, its onset impulse function weighted by durational accents, f, and the corresponding autocorrelation function, F. The maximum of the autocorrelation function at the lag of 4/ 8 indicates duple meter. Autocorrelation refers to the correlation of two copies of a time series that are temporally shifted with respect to each other. For a given amount of shift or lag, a high value of autocorrelation suggests that the series contains a periodicity with length equaling the lag. In the present study, the autocorrelation function F was defined as F m = n f n f n m n f n 2, where m denotes the lag in units of sampling interval; the denominator normalizes the function to F 0 =1 irrespective of the length of the sequence. Often, the lag corresponding to the maximum of the autocorrelation function provides an estimate of the meter. This is the case for the melody depicted in Fig. 1. Sometimes the temporal structure alone is not sufficient for deducing the meter. This holds, for example, for isochronous and temporally highly aperiodic melodies. In such cases, melodic structure may provide cues for the determination of meter. This is the case, for instance, with the melody depicted in Fig. 2. With this isochronous melody, the autocorrelation function obtained from the duration-weighted onset impulse function fails to exhibit any peaks, thus making it impossible to determine the meter. Including information about pitch content in the onset impulse function leads, however, to an autocorrelation function with clearly discernible peaks. III. STUDY 1 A. Material The material used in the first study consisted of folk melodies in MIDI file format taken from two collections: the Essen collection Schaffrath, 1995, consisting of mainly European folk melodies, and the Digital Archive of Finnish Folk Tunes Eerola and Toiviainen, 2004a, subsequently referred to as the Finnish collection. For the present study, all melodies in either duple 2/4, 4/4, 4/8, etc.; 2 n eighth notes per measure or triple/compound 3/8, 3/4, 6/8, 9/8, 12/8, etc.; 3n eighth notes per measure meter were chosen. Consequently, a total of 5507 melodies in the Essen collection were used in the study, of which % were in duple and % were in triple/compound meter. From the 3 J. Acoust. Soc. Am., Vol. 119, No. 2, February 2006 P. Toiviainen and T. Eerola: Autocorrelation in meter induction 1165

3 FIG. 3. a Onset impulse functions constructed from a melodic excerpt using the six accent types described in the text; b the respective autocorrelation functions. As can be seen, the melodic accents frequently fail to co-occur either with each other or with the durational accents. All the autocorrelation functions, however, have maxima at lags of either 6/8 or 12/8, indicating triple or compound meter. FIG. 2. Excerpt from an isochronous melody. a Onset impulse function weighted by durational accents, f, and the corresponding autocorrelation function, F, showing no discernible peaks. b Onset impulse function weighted by interval size, f, and the corresponding autocorrelation function, F. The maximum of the autocorrelation function at the lag of 12/ 8 indicates triple or compound meter. Finnish collection, 6861 melodies were used, of which % were in duple and % were in triple/ compound meter. B. Method For each of the melodies in the two collections, we constructed a set of onset impulse functions weighted by various accent types Eqs. 1 and 2. In each case the sampling interval was set to 1/16 note. The accents consisted of 1 durational accent a i equals tone duration, 2 Thomassen s 1982 melodic accent, 3 interval size in semitones between previous and current tone see, e.g., Vos et al., 1994, 4 pivotal accent a i =1 if melody changes direction, a i =0 otherwise, and 5 gross contour accent a i =1 for ascending interval, a i = 1 for descending interval, a i =0 otherwise. Since the note onset times alone, without regard to any accent structure, provide information about metrical structure, we further included 6 constant accent a i =1. The analysis was carried out using the MIDI Toolbox for Matlab Eerola and Toiviainen, 2004b. For each melody, each of the onset impulse functions was subjected to autocorrelation. The components of the obtained autocorrelation functions corresponding to lags of 1, 2,, 16 eighth notes were included in the subsequent analyses. Figure 3 depicts the onset impulse functions and the respective autocorrelation functions constructed from a melodic excerpt using each of the accent types described above. The extent to which these autocorrelation functions could predict the meter of each melody was assessed by means of stepwise discriminant function analyses, in which various subsets of autocorrelation functions were used as independent variables and the meter duple versus triple/ compound as the dependent variable. The leave-one-out cross-validation scheme Lachenbruch and Mickey, 1968 was utilized. The measures observed in the analyses were the percentage of correctly classified cases, the order in which variables entered into the discriminant function, and the discriminant function coefficients. C. Results The analysis was started by considering the autocorrelation function based on durational accents. First, only the components corresponding to lags of 4 and 6 eighth notes were used as independent variables. This yielded a correct classification rate of 80.9% for the Essen collection and 84.6% for the Finnish collection. Thus, a significant proportion of the melodies was misclassified, suggesting that periodicities of 4/8 and 6/8 in durational accent are not sufficient for reliable classification. This can be clearly seen in Fig. 4, where the values of these components are displayed as scatter plots, showing a significant overlap between melodies representing the two types of meter. Next, all the components of the durational-accent-based autocorrelation function were entered into the analysis. This yielded a correct classification rate of 90.5% for the Essen collection and 93.1% for the Finnish collection. Inclusion of all 16 components as independent variables thus considerably improved classification performance. The first components to enter into the discriminant function were lags of 8/8, 12/8, and 16/8 in this order for the Essen collection, and 4/8, 12/8, and 16/8 for the Finnish collection. This suggests that periodicities longer than one bar may provide cues for meter determination that are more reliable than shorter periodicities. Subsequently, discriminant function analyses were carried out with the autocorrelation functions obtained using each of the remaining accents, one at a time. The results are summarized in Table I. As can be seen, a significant proportion of the components of the autocorrelation functions that entered first in the stepwise analysis correspond to relatively long time lags. In particular, for all accent types lag 12/8 is among the three most important components. For lag 16/ 8 the same holds true for all but one accent type. This again suggests that periodicities exceeding the span of one bar 1166 J. Acoust. Soc. Am., Vol. 119, No. 2, February 2006 P. Toiviainen and T. Eerola: Autocorrelation in meter induction

4 FIG. 4. Scatter plot of the values for lags 4/8 and 6/8 of the durationalaccent-based autocorrelation function for the melodies in the Essen collection and the Finnish collection that were used in the present study. Circles and crosses represent melodies in duple and triple meter, respectively. seem to offer highly important cues for meter induction. Somewhat unexpectedly, the autocorrelation function obtained by ignoring any accent structure constant accent yielded for both collections the highest classification rate. In light of this result, the onset function constructed using a constant accent may be slightly more efficient in meter induction than the function originally introduced by Brown In terms of correct classification rate, the constant accent was followed by durational accent and Thomassen s melodic accent, in this order. Finally, all autocorrelation functions were used together as independent variables. This yielded a correct classification rate of 95.3% and 96.4% for the Essen collection and the Finnish collection, respectively. The first variable to enter into the discriminant function was lag 8/8 with durational accent for the Essen collection and lag 4/8 with constant accent for the Finnish collection. The next three variables to enter were the same for both collections: lag 12/8 with constant accent, lag 16/8 with constant accent, and lag 6 with melodic accent, in this order. This analysis contained a total of 96 independent variables 16 for each of the 6 accent types ; due to the large number of cases, a large number of variables were entered into the stepwise analyses. To obtain a simpler model for meter classification, we performed a further discriminant function analysis in which the six most prominent components for two most prominent accent types from the previous analysis were used. These corresponded to the values of the autocorrelation functions at lags of 3/8, 4/8, 6/8, 8/8, 12/8, and 16/8 using constant accent and Thomassen s melodic accent. Using both collections together N=12 368, we obtained the following discriminant function, = F const 3/ F const 4/ F const 6/ F const 8/ F const 12/ F const 16/ F mel 3/ F mel 4/ F mel 6/ F mel 8/ F mel 12/ F mel 16/8, 4 where F const and F mel denote autocorrelation functions ob- TABLE I. Proportions of correctly classified melodies classification rate and components of the autocorrelation function that entered first in each analysis primary components; numbers refer to lags in units of one eighth note for both collections and each accent type. Accent type 1 Duration 2 Thomassen 3 Interval 4 Pivot 5 Contour 6 Constant Classification rate % Primary Components Classification rate % Primary Components Essen Collection N= , 12, 16 16, 12, 8 16, 12, 4 16, 12, 6 16, 12, 8 8, 12, 16 Finnish collection N= , 12, 16 8, 12, 4 8, 12, 16 6, 16, 12 12, 6, 16 4, 12, 16 J. Acoust. Soc. Am., Vol. 119, No. 2, February 2006 P. Toiviainen and T. Eerola: Autocorrelation in meter induction 1167

5 TABLE II. Confusion matrices between notated and predicted meters for both collections. Each element shows the number of melodies with the respective notated and predicted meter. For each meter, the recall, precision, and F score values are given. Essen Collection N=5592 Predicted meter 2/4 3/2 3/4 3/8 4/1 4/2 4/4 6/4 6/8 Total Recall Precision F Notated meter 2/ / / / / / / / / Total Finnish Collection N=7351 Predicted meter 2/ 4 3/ 2 3/ 4 3/ 8 4/ 4 5/ 2 5/ 4 6/ 4 6/ 8 Total Recall Precision F Notated meter 2/ / / / / / / / / Total tained with constant and Thomassen s melodic accent, respectively. With this discriminant function, the correct classification rates were 92.9% and 94.8% for the Essen and the Finnish collections, respectively. This simpler discriminant function thus yielded correct classification rates that were merely 2.4% and 1.6% lower than those obtained with 96 predictive variables. Variables that received the largest coefficients in the discriminant function were F const 12/8, F const 4/8, F const 16/8, F mel 12/8, and F mel 6/8, in this order. In concordance with the aforementioned results, this suggests that temporal structure above the bar level produces important cues for meter determination. Further, the fact that the most significant components of the melodic accent autocorrelation function correspond to multiples of 3/8 lags suggests that, especially for triple and compound meters, melodic accent structure provides additional cues. IV. STUDY 2 The aim of study 2 was to assess the capability of the autocorrelation-based meter induction method to carry out a more detailed classification. More specifically, instead of mere classification as duple versus triple, the dependent variable used in this study was the actual notated meter. In the analysis, special attention was paid to the pattern of confusion between meters. A. Material As in study 1, the material was taken from the Essen collection and the Digital Archive of Finnish Folk Tunes. From each collection, melodies that consisted of a single notated meter were included. Moreover, for each collection only meters that contained more than 30 exemplars were included. Consequently, a total of 5592 melodies in the Essen collection were used, representing nine different notated meters 2/4, 3/2, 3/4, 3/8, 4/1, 4/2, 4/4, 6/4, 6/8. From the Finnish collection, 7351 melodies were used, representing nine different notated meters 2/4, 3/2, 3/4, 3/8, 4/4, 5/2, 5/4, 6/4, 6/8. For each collection, the number of melodies representing each notated meter is shown in Table II. B. Methods The classification of meters was based on the discriminant function obtained using the autocorrelation functions obtained using all the accent types. The performance was assessed by means of a confusion matrix. Furthermore, for both collections the precision and recall values as well as the F score were calculated for each meter Salton and McGill, For a given meter, precision is defined as the number of melodies having the meter and being correctly classified divided by the total number of melodies being classified as representing the meter. A high value of precision thus indi J. Acoust. Soc. Am., Vol. 119, No. 2, February 2006 P. Toiviainen and T. Eerola: Autocorrelation in meter induction

6 cates that, of the melodies classified as being notated in a given meter, a large proportion is correctly classified. Similarly, for each meter, recall is defined as the number of melodies being notated in the meter and being correctly classified divided by the total number of melodies being notated in the meter. A high value of recall thus indicates that of the melodies being notated in a given meter, a large proportion is correctly classified. The F score is defined as the harmonic mean of precision and recall and is regarded as an overall measure of classification performance see, e.g., Salton and McGill, C. Results Overall, 83.2% of the melodies from the Essen collection and 68.0% of those from the Finnish collection were correctly classified. These proportions being lower than the corresponding figures in the first study is due to the larger number of classes used in this study nine than in the first two. The notably low correct classification rate for the Finnish collection can be mainly attributed to the fact that a large proportion 43.4% of the melodies representing 4/4 meter were classified as being 2/4 see below. To obtain a more detailed view of the classification performance, we calculated the confusion matrices for both collections. They are displayed in Table II. The table also shows the precision and recall values as well as the F scores for each meter. In terms of the F score, the most accurately classified meters were 4/4 and 2/4 for the Essen collection and 5/4 and 3/4 for the Finnish collection. Similarly, the least accurately classified meters were 6/4 and 3/8 for both collections. Table II reveals that the most frequent confusions were made within the groups of duple and triple/compound meters, as defined in study 1, whereas confusions across these groups where significantly less frequent. For both collections, meters 2/4 and 4/4 displayed the highest mutual confusion rate, followed by meters 3/4 and 6/4. A large proportion of these misclassifications can probably be attributed to the effect of tempo on the choice of notated meter cf. London, Certain confusions imply more severe misattributions by the algorithm. For instance, 11.7% of the melodies in the Essen collection notated in 3/4 meter were misclassified as representing binary meter 4/4 or 2/4, the corresponding figure for the Finnish collection being 12.6%. In general, duple meters were less frequently misclassified as representing triple/compound meter as vice versa. This asymmetry is in line with the results obtained in study 1. Further research would be needed to account for this phenomenon. As the confusion matrices contain an abundance of numbers, the relationship between meters may be difficult to see. Therefore we visualized the relations between meters by performing separate hierarchical cluster analyses for both collections. To this end, we calculated the distance between each meter from the confusion matrix according to the formula FIG. 5. Dendrograms obtained from the confusion matrix using the similarity measure of Eq. 5. The leftmost column displays the average note durations in quarter notes for the melodies representing each meter. d ij =1 c ij + c ji c ii + c jj, where d ij denotes the distance between meters i and j, and c ij is the number of cases where a melody in meter i has been classified as being in meter j. By definition, the larger the proportion of melodies confused between meters, c ij +c ji,to the number of melodies correctly classified for both meters, c ii +c jj, the smaller the distance d ij between the meters. Figure 5 displays the dendrograms obtained from the clustering algorithms. In the dendrograms of Fig. 5, the stage at which given meters cluster together reflects the algorithm s rate of confusion between the meters. For both collections, the meters to first cluster together are 3/8 and 6/8. For the Essen collection, this is followed by the clustering of the meters 3/4 and 6/4 as well as 2/4 and 4/4, in this order. Also for the Finnish collection these pairs of meters cluster next, albeit in reverse order, that is, the clustering of 2/4 and 4/4 precedes that of 3/4 and 6/4. A further similar feature between the two dendrograms is that the last clustering occurs between the cluster formed by the meters 3/8 and 6/8 and the cluster formed by all the other meters. This suggests that, in terms of the autocorrelation functions, meters 3/8 and 6/8 are most distinct from the other meters. One peculiar feature of the dendrogram for the Essen collection is the relatively late clustering of meters 4/1 and 4/2 with meters 2/4 and 4/4. In particular, the former two meters cluster with meter 3/2 before clustering with the latter two meters. A potential explanation for this is the difference in the average note durations between the meters, shown in the leftmost column of Fig. 5. More specifically, the average note durations for meters 4/1, 4/2, and 3/2 exceed those of meters 2/4 and 4/4 by a factor of 2. This anomaly, however, is not significant, as meters 4/1, 4/2, and 3/2 constitute merely a minor proportion of the whole collection. 5 J. Acoust. Soc. Am., Vol. 119, No. 2, February 2006 P. Toiviainen and T. Eerola: Autocorrelation in meter induction 1169

7 V. CONCLUSIONS We studied the classification performance of the autocorrelation-based meter induction model, originally introduced by Brown Using discriminant function analysis, we provided an explicit method for the classification. Furthermore, we applied the algorithm to investigate the role of melodic accent structure in meter induction. In conformance with the general view, we found that periodicity in note onset locations was the most important cue for the determination of meter. The results also imply that periodicities longer than one measure provide additional information for meter induction. A somewhat unexpected finding was that ignoring the accent structure in the onset impulse function yielded the best classification rate. This result is difficult to explain and calls for further investigation. Furthermore, the results suggest that periodicity in melodic accent structure may serve as an additional cue in the determination of meter. In particular, including Thomassen s 1982 melodic accent was found to improve classification performance to some extent. This, however, does not necessarily imply that melodic and durational accents need cooccur but rather that they exhibit similar periodic structure with eventual mutual phase shift. An additional finding that calls for further study was the significant difference between the correct classification rates for melodies in duple and triple/compound meter. More specifically, melodies in duple meter were more often correctly classified than melodies in triple/compound meter. A detailed investigation of misclassified melodies could provide insight into this question. The material used in the present study consisted of melodies predominantly in duple, triple, and compound meters, although study 2 utilized a few hundred Finnish melodies in either 5/4 or 5/2 meter. To obtain a deeper insight into the role of accent structure in meter induction, a study with nonregular meters such as those present in the folk music of the Balkan region could be carried out. An apparent limitation of the method presented in this article is its inability to deal with melodies that contain changes of meter. For a melody that, say, starts in 2/4 meter and changes to 3/ 4 meter, the algorithm gives unpredictable results. This is due to the fact that the algorithm considers the melody as a whole. The limitation may be overcome by applying a windowed analysis analogical to algorithms used in pitch estimation from acoustical signals, in which the autocorrelation is applied to short windowed segments of the melody, with the window moving gradually throughout the melody. The present study utilized melodies that were represented in symbolic, temporally quantized form. The choice of stimuli was mainly based on the availability of correct notated meters for the melodies in the collections. In principle the method could, however, be applied to performed music in acoustical form as well, at least with a monophonic input. This would require algorithms for onset detection e.g., Klapuri, 1999, pitch estimation e.g., Brown and Puckette, 1994; Klapuri, 2003, and beat tracking e.g., Dixon, 2001; Large and Kolen, 1994; Toiviainen, 1998, ACKNOWLEDGMENTS This work was supported by the Academy of Finland Project No Brown, J. C Determination of meter of musical scores by autocorrelation, J. Acoust. Soc. Am. 94, Brown, J. C., and Puckette, M. S A high resolution fundamental frequency determination based on phase changes of the Fourier transform, J. Acoust. Soc. Am. 94, Cooper, G., and Meyer, L. B The Rhythmic Structure of Music Univ. of Chicago, Chicago. Dawe, L. A., Platt, J. R., and Racine, R. J Harmonic accents in inference of metrical structure and perception of rhythm patterns, Percept. Psychophys. 54, Dixon, S Automatic extraction of tempo and beat from expressive performances, J. New Music Res. 30, Dixon, S., and Cambouropoulos, E Beat tracking with musical knowledge, in ECAI 2000: Proc. 14th European Conference on Artificial Intelligence IOS, Amsterdam, Netherlands, pp Eerola, T., and Toiviainen, P. 2004a. Digital Archive of Finnish Folk Tunes, University of Jyväskylä, Jyväskylä, Finland. Available at: Eerola, T., and Toiviainen, P. 2004b. MIDI Toolbox: MATLAB Tools for Music Research, University of Jyväskylä, Jyväskylä, Finland. Available at Fraisse, P Rhythm and tempo, in Psychology of Music, edited by D. Deutsch Academic, New York, pp Hannon, E., Snyder, J., Eerola, T., and Krumhansl, C. L The Role of melodic and temporal cues in perceiving musical meter, J. Exp. Psychol. Hum. Percept. Perform. 30, Huron, D., and Royal, M What is melodic accent? Converging evidence from musical practice, Music Percept. 13, Klapuri, A Sound Onset Detection by Applying Psychoacoustic Knowledge, Proc. IEEE Int. Conf. Acoustics Speech and Sig. Proc. IC- ASSP, Phoenix, AZ, pp Klapuri, A Multiple fundamental frequency estimation by harmonicity and spectral smoothness, IEEE Trans. Speech Audio Process. 11, Lachenbruch, P. A., and Mickey, M. R Estimation of error rates in discriminant analysis, Technometrics 10, Large, E. W., and Kolen, J. F Resonance and the perception of musical meter, Connection Sci. 6, Lerdahl, F., and Jackendoff, R A Generative Theory of Tonal Music MIT, Cambridge, MA. London, J Cognitive constraints on metric systems: some observations and hypotheses, Music Percept. 19, Parncutt, R A perceptual model of pulse salience and metrical accent in musical rhythms, Music Percept. 11, Salton, G., and McGill, M Introduction to Modern Information Retrieval McGraw Hill, New York. Schaffrath, H The Essen Folksong Collection in Kern Format computer database, edited by D. Huron Center for Computer Assisted Research in the Humanities, Menlo Park, CA. Thomassen, J. M Melodic accent: Experiments and a tentative model, J. Acoust. Soc. Am. 71, Toiviainen, P Modelling the perception of metre with competing subharmonic oscillators, Proc. Third Triennial ESCOM Conference, Uppsala, Uppsala University, pp Toiviainen, P An interactive MIDI accompanist, Comput. Music J. 22, Toiviainen, P Real-time recognition of improvisations with adaptive oscillators and a recursive Bayesian classifier, J. New Music Res. 30, van Noorden, L., and Moelants, D Resonance in the perception of musical pulse, J. New Music Res. 28, Vos, P. G., van Dijk, A., and Schomaker, L Melodic cues for metre, Perception 23, J. Acoust. Soc. Am., Vol. 119, No. 2, February 2006 P. Toiviainen and T. Eerola: Autocorrelation in meter induction

CLASSIFICATION OF MUSICAL METRE WITH AUTOCORRELATION AND DISCRIMINANT FUNCTIONS

CLASSIFICATION OF MUSICAL METRE WITH AUTOCORRELATION AND DISCRIMINANT FUNCTIONS CLASSIFICATION OF MUSICAL METRE WITH AUTOCORRELATION AND DISCRIMINANT FUNCTIONS Petri Toiviainen Department of Music University of Jyväskylä Finland ptoiviai@campus.jyu.fi Tuomas Eerola Department of Music

More information

Robert Alexandru Dobre, Cristian Negrescu

Robert Alexandru Dobre, Cristian Negrescu ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Automatic Music Transcription Software Based on Constant Q

More information

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene Beat Extraction from Expressive Musical Performances Simon Dixon, Werner Goebl and Emilios Cambouropoulos Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria.

More information

Meter and Autocorrelation

Meter and Autocorrelation Meter and Autocorrelation Douglas Eck University of Montreal Department of Computer Science CP 6128, Succ. Centre-Ville Montreal, Quebec H3C 3J7 CANADA eckdoug@iro.umontreal.ca Abstract This paper introduces

More information

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring 2009 Week 6 Class Notes Pitch Perception Introduction Pitch may be described as that attribute of auditory sensation in terms

More information

CALCULATING SIMILARITY OF FOLK SONG VARIANTS WITH MELODY-BASED FEATURES

CALCULATING SIMILARITY OF FOLK SONG VARIANTS WITH MELODY-BASED FEATURES CALCULATING SIMILARITY OF FOLK SONG VARIANTS WITH MELODY-BASED FEATURES Ciril Bohak, Matija Marolt Faculty of Computer and Information Science University of Ljubljana, Slovenia {ciril.bohak, matija.marolt}@fri.uni-lj.si

More information

Analysis of local and global timing and pitch change in ordinary

Analysis of local and global timing and pitch change in ordinary Alma Mater Studiorum University of Bologna, August -6 6 Analysis of local and global timing and pitch change in ordinary melodies Roger Watt Dept. of Psychology, University of Stirling, Scotland r.j.watt@stirling.ac.uk

More information

Human Preferences for Tempo Smoothness

Human Preferences for Tempo Smoothness In H. Lappalainen (Ed.), Proceedings of the VII International Symposium on Systematic and Comparative Musicology, III International Conference on Cognitive Musicology, August, 6 9, 200. Jyväskylä, Finland,

More information

An Empirical Comparison of Tempo Trackers

An Empirical Comparison of Tempo Trackers An Empirical Comparison of Tempo Trackers Simon Dixon Austrian Research Institute for Artificial Intelligence Schottengasse 3, A-1010 Vienna, Austria simon@oefai.at An Empirical Comparison of Tempo Trackers

More information

BEAT AND METER EXTRACTION USING GAUSSIFIED ONSETS

BEAT AND METER EXTRACTION USING GAUSSIFIED ONSETS B BEAT AND METER EXTRACTION USING GAUSSIFIED ONSETS Klaus Frieler University of Hamburg Department of Systematic Musicology kgfomniversumde ABSTRACT Rhythm, beat and meter are key concepts of music in

More information

METRICAL STRENGTH AND CONTRADICTION IN TURKISH MAKAM MUSIC

METRICAL STRENGTH AND CONTRADICTION IN TURKISH MAKAM MUSIC Proc. of the nd CompMusic Workshop (Istanbul, Turkey, July -, ) METRICAL STRENGTH AND CONTRADICTION IN TURKISH MAKAM MUSIC Andre Holzapfel Music Technology Group Universitat Pompeu Fabra Barcelona, Spain

More information

Finding Meter in Music Using an Autocorrelation Phase Matrix and Shannon Entropy

Finding Meter in Music Using an Autocorrelation Phase Matrix and Shannon Entropy Finding Meter in Music Using an Autocorrelation Phase Matrix and Shannon Entropy Douglas Eck University of Montreal Department of Computer Science CP 6128, Succ. Centre-Ville Montreal, Quebec H3C 3J7 CANADA

More information

Acoustic and musical foundations of the speech/song illusion

Acoustic and musical foundations of the speech/song illusion Acoustic and musical foundations of the speech/song illusion Adam Tierney, *1 Aniruddh Patel #2, Mara Breen^3 * Department of Psychological Sciences, Birkbeck, University of London, United Kingdom # Department

More information

THE INTERACTION BETWEEN MELODIC PITCH CONTENT AND RHYTHMIC PERCEPTION. Gideon Broshy, Leah Latterner and Kevin Sherwin

THE INTERACTION BETWEEN MELODIC PITCH CONTENT AND RHYTHMIC PERCEPTION. Gideon Broshy, Leah Latterner and Kevin Sherwin THE INTERACTION BETWEEN MELODIC PITCH CONTENT AND RHYTHMIC PERCEPTION. BACKGROUND AND AIMS [Leah Latterner]. Introduction Gideon Broshy, Leah Latterner and Kevin Sherwin Yale University, Cognition of Musical

More information

Quarterly Progress and Status Report. Perception of just noticeable time displacement of a tone presented in a metrical sequence at different tempos

Quarterly Progress and Status Report. Perception of just noticeable time displacement of a tone presented in a metrical sequence at different tempos Dept. for Speech, Music and Hearing Quarterly Progress and Status Report Perception of just noticeable time displacement of a tone presented in a metrical sequence at different tempos Friberg, A. and Sundberg,

More information

Automatic meter extraction from MIDI files (Extraction automatique de mètres à partir de fichiers MIDI)

Automatic meter extraction from MIDI files (Extraction automatique de mètres à partir de fichiers MIDI) Journées d'informatique Musicale, 9 e édition, Marseille, 9-1 mai 00 Automatic meter extraction from MIDI files (Extraction automatique de mètres à partir de fichiers MIDI) Benoit Meudic Ircam - Centre

More information

Supervised Learning in Genre Classification

Supervised Learning in Genre Classification Supervised Learning in Genre Classification Introduction & Motivation Mohit Rajani and Luke Ekkizogloy {i.mohit,luke.ekkizogloy}@gmail.com Stanford University, CS229: Machine Learning, 2009 Now that music

More information

Automatic music transcription

Automatic music transcription Music transcription 1 Music transcription 2 Automatic music transcription Sources: * Klapuri, Introduction to music transcription, 2006. www.cs.tut.fi/sgn/arg/klap/amt-intro.pdf * Klapuri, Eronen, Astola:

More information

NEW QUERY-BY-HUMMING MUSIC RETRIEVAL SYSTEM CONCEPTION AND EVALUATION BASED ON A QUERY NATURE STUDY

NEW QUERY-BY-HUMMING MUSIC RETRIEVAL SYSTEM CONCEPTION AND EVALUATION BASED ON A QUERY NATURE STUDY Proceedings of the COST G-6 Conference on Digital Audio Effects (DAFX-), Limerick, Ireland, December 6-8,2 NEW QUERY-BY-HUMMING MUSIC RETRIEVAL SYSTEM CONCEPTION AND EVALUATION BASED ON A QUERY NATURE

More information

Automatic Rhythmic Notation from Single Voice Audio Sources

Automatic Rhythmic Notation from Single Voice Audio Sources Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung

More information

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t MPEG-7 FOR CONTENT-BASED MUSIC PROCESSING Λ Emilia GÓMEZ, Fabien GOUYON, Perfecto HERRERA and Xavier AMATRIAIN Music Technology Group, Universitat Pompeu Fabra, Barcelona, SPAIN http://www.iua.upf.es/mtg

More information

Rhythm: patterns of events in time. HST 725 Lecture 13 Music Perception & Cognition

Rhythm: patterns of events in time. HST 725 Lecture 13 Music Perception & Cognition Harvard-MIT Division of Sciences and Technology HST.725: Music Perception and Cognition Prof. Peter Cariani Rhythm: patterns of events in time HST 725 Lecture 13 Music Perception & Cognition (Image removed

More information

Music Radar: A Web-based Query by Humming System

Music Radar: A Web-based Query by Humming System Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,

More information

Controlling Musical Tempo from Dance Movement in Real-Time: A Possible Approach

Controlling Musical Tempo from Dance Movement in Real-Time: A Possible Approach Controlling Musical Tempo from Dance Movement in Real-Time: A Possible Approach Carlos Guedes New York University email: carlos.guedes@nyu.edu Abstract In this paper, I present a possible approach for

More information

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu

More information

Audio Feature Extraction for Corpus Analysis

Audio Feature Extraction for Corpus Analysis Audio Feature Extraction for Corpus Analysis Anja Volk Sound and Music Technology 5 Dec 2017 1 Corpus analysis What is corpus analysis study a large corpus of music for gaining insights on general trends

More information

TRADITIONAL ASYMMETRIC RHYTHMS: A REFINED MODEL OF METER INDUCTION BASED ON ASYMMETRIC METER TEMPLATES

TRADITIONAL ASYMMETRIC RHYTHMS: A REFINED MODEL OF METER INDUCTION BASED ON ASYMMETRIC METER TEMPLATES TRADITIONAL ASYMMETRIC RHYTHMS: A REFINED MODEL OF METER INDUCTION BASED ON ASYMMETRIC METER TEMPLATES Thanos Fouloulis Aggelos Pikrakis Emilios Cambouropoulos Dept. of Music Studies, Aristotle Univ. of

More information

Smooth Rhythms as Probes of Entrainment. Music Perception 10 (1993): ABSTRACT

Smooth Rhythms as Probes of Entrainment. Music Perception 10 (1993): ABSTRACT Smooth Rhythms as Probes of Entrainment Music Perception 10 (1993): 503-508 ABSTRACT If one hypothesizes rhythmic perception as a process employing oscillatory circuits in the brain that entrain to low-frequency

More information

10 Visualization of Tonal Content in the Symbolic and Audio Domains

10 Visualization of Tonal Content in the Symbolic and Audio Domains 10 Visualization of Tonal Content in the Symbolic and Audio Domains Petri Toiviainen Department of Music PO Box 35 (M) 40014 University of Jyväskylä Finland ptoiviai@campus.jyu.fi Abstract Various computational

More information

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS Andrew N. Robertson, Mark D. Plumbley Centre for Digital Music

More information

A QUANTIFICATION OF THE RHYTHMIC QUALITIES OF SALIENCE AND KINESIS

A QUANTIFICATION OF THE RHYTHMIC QUALITIES OF SALIENCE AND KINESIS 10.2478/cris-2013-0006 A QUANTIFICATION OF THE RHYTHMIC QUALITIES OF SALIENCE AND KINESIS EDUARDO LOPES ANDRÉ GONÇALVES From a cognitive point of view, it is easily perceived that some music rhythmic structures

More information

The Generation of Metric Hierarchies using Inner Metric Analysis

The Generation of Metric Hierarchies using Inner Metric Analysis The Generation of Metric Hierarchies using Inner Metric Analysis Anja Volk Department of Information and Computing Sciences, Utrecht University Technical Report UU-CS-2008-006 www.cs.uu.nl ISSN: 0924-3275

More information

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY Eugene Mikyung Kim Department of Music Technology, Korea National University of Arts eugene@u.northwestern.edu ABSTRACT

More information

On time: the influence of tempo, structure and style on the timing of grace notes in skilled musical performance

On time: the influence of tempo, structure and style on the timing of grace notes in skilled musical performance RHYTHM IN MUSIC PERFORMANCE AND PERCEIVED STRUCTURE 1 On time: the influence of tempo, structure and style on the timing of grace notes in skilled musical performance W. Luke Windsor, Rinus Aarts, Peter

More information

Music Information Retrieval Using Audio Input

Music Information Retrieval Using Audio Input Music Information Retrieval Using Audio Input Lloyd A. Smith, Rodger J. McNab and Ian H. Witten Department of Computer Science University of Waikato Private Bag 35 Hamilton, New Zealand {las, rjmcnab,

More information

A Beat Tracking System for Audio Signals

A Beat Tracking System for Audio Signals A Beat Tracking System for Audio Signals Simon Dixon Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria. simon@ai.univie.ac.at April 7, 2000 Abstract We present

More information

Perceiving temporal regularity in music

Perceiving temporal regularity in music Cognitive Science 26 (2002) 1 37 http://www.elsevier.com/locate/cogsci Perceiving temporal regularity in music Edward W. Large a, *, Caroline Palmer b a Florida Atlantic University, Boca Raton, FL 33431-0991,

More information

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Róisín Loughran roisin.loughran@ul.ie Jacqueline Walker jacqueline.walker@ul.ie Michael O Neill University

More information

Hidden Markov Model based dance recognition

Hidden Markov Model based dance recognition Hidden Markov Model based dance recognition Dragutin Hrenek, Nenad Mikša, Robert Perica, Pavle Prentašić and Boris Trubić University of Zagreb, Faculty of Electrical Engineering and Computing Unska 3,

More information

Musicians Adjustment of Performance to Room Acoustics, Part III: Understanding the Variations in Musical Expressions

Musicians Adjustment of Performance to Room Acoustics, Part III: Understanding the Variations in Musical Expressions Musicians Adjustment of Performance to Room Acoustics, Part III: Understanding the Variations in Musical Expressions K. Kato a, K. Ueno b and K. Kawai c a Center for Advanced Science and Innovation, Osaka

More information

Syncopation and the Score

Syncopation and the Score Chunyang Song*, Andrew J. R. Simpson, Christopher A. Harte, Marcus T. Pearce, Mark B. Sandler Centre for Digital Music, Queen Mary University of London, London, United Kingdom Abstract The score is a symbolic

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

Influence of timbre, presence/absence of tonal hierarchy and musical training on the perception of musical tension and relaxation schemas

Influence of timbre, presence/absence of tonal hierarchy and musical training on the perception of musical tension and relaxation schemas Influence of timbre, presence/absence of tonal hierarchy and musical training on the perception of musical and schemas Stella Paraskeva (,) Stephen McAdams (,) () Institut de Recherche et de Coordination

More information

MODELING RHYTHM SIMILARITY FOR ELECTRONIC DANCE MUSIC

MODELING RHYTHM SIMILARITY FOR ELECTRONIC DANCE MUSIC MODELING RHYTHM SIMILARITY FOR ELECTRONIC DANCE MUSIC Maria Panteli University of Amsterdam, Amsterdam, Netherlands m.x.panteli@gmail.com Niels Bogaards Elephantcandy, Amsterdam, Netherlands niels@elephantcandy.com

More information

Tempo and Beat Analysis

Tempo and Beat Analysis Advanced Course Computer Science Music Processing Summer Term 2010 Meinard Müller, Peter Grosche Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Tempo and Beat Analysis Musical Properties:

More information

VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS. O. Javed, S. Khan, Z. Rasheed, M.Shah. {ojaved, khan, zrasheed,

VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS. O. Javed, S. Khan, Z. Rasheed, M.Shah. {ojaved, khan, zrasheed, VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS O. Javed, S. Khan, Z. Rasheed, M.Shah {ojaved, khan, zrasheed, shah}@cs.ucf.edu Computer Vision Lab School of Electrical Engineering and Computer

More information

Meter Detection in Symbolic Music Using a Lexicalized PCFG

Meter Detection in Symbolic Music Using a Lexicalized PCFG Meter Detection in Symbolic Music Using a Lexicalized PCFG Andrew McLeod University of Edinburgh A.McLeod-5@sms.ed.ac.uk Mark Steedman University of Edinburgh steedman@inf.ed.ac.uk ABSTRACT This work proposes

More information

Classification of Dance Music by Periodicity Patterns

Classification of Dance Music by Periodicity Patterns Classification of Dance Music by Periodicity Patterns Simon Dixon Austrian Research Institute for AI Freyung 6/6, Vienna 1010, Austria simon@oefai.at Elias Pampalk Austrian Research Institute for AI Freyung

More information

Multiple instrument tracking based on reconstruction error, pitch continuity and instrument activity

Multiple instrument tracking based on reconstruction error, pitch continuity and instrument activity Multiple instrument tracking based on reconstruction error, pitch continuity and instrument activity Holger Kirchhoff 1, Simon Dixon 1, and Anssi Klapuri 2 1 Centre for Digital Music, Queen Mary University

More information

AN APPROACH FOR MELODY EXTRACTION FROM POLYPHONIC AUDIO: USING PERCEPTUAL PRINCIPLES AND MELODIC SMOOTHNESS

AN APPROACH FOR MELODY EXTRACTION FROM POLYPHONIC AUDIO: USING PERCEPTUAL PRINCIPLES AND MELODIC SMOOTHNESS AN APPROACH FOR MELODY EXTRACTION FROM POLYPHONIC AUDIO: USING PERCEPTUAL PRINCIPLES AND MELODIC SMOOTHNESS Rui Pedro Paiva CISUC Centre for Informatics and Systems of the University of Coimbra Department

More information

Transcription of the Singing Melody in Polyphonic Music

Transcription of the Singing Melody in Polyphonic Music Transcription of the Singing Melody in Polyphonic Music Matti Ryynänen and Anssi Klapuri Institute of Signal Processing, Tampere University Of Technology P.O.Box 553, FI-33101 Tampere, Finland {matti.ryynanen,

More information

THE importance of music content analysis for musical

THE importance of music content analysis for musical IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 15, NO. 1, JANUARY 2007 333 Drum Sound Recognition for Polyphonic Audio Signals by Adaptation and Matching of Spectrogram Templates With

More information

HST 725 Music Perception & Cognition Assignment #1 =================================================================

HST 725 Music Perception & Cognition Assignment #1 ================================================================= HST.725 Music Perception and Cognition, Spring 2009 Harvard-MIT Division of Health Sciences and Technology Course Director: Dr. Peter Cariani HST 725 Music Perception & Cognition Assignment #1 =================================================================

More information

Tapping to Uneven Beats

Tapping to Uneven Beats Tapping to Uneven Beats Stephen Guerra, Julia Hosch, Peter Selinsky Yale University, Cognition of Musical Rhythm, Virtual Lab 1. BACKGROUND AND AIMS [Hosch] 1.1 Introduction One of the brain s most complex

More information

EXPLAINING AND PREDICTING THE PERCEPTION OF MUSICAL STRUCTURE

EXPLAINING AND PREDICTING THE PERCEPTION OF MUSICAL STRUCTURE JORDAN B. L. SMITH MATHEMUSICAL CONVERSATIONS STUDY DAY, 12 FEBRUARY 2015 RAFFLES INSTITUTION EXPLAINING AND PREDICTING THE PERCEPTION OF MUSICAL STRUCTURE OUTLINE What is musical structure? How do people

More information

Music. Last Updated: May 28, 2015, 11:49 am NORTH CAROLINA ESSENTIAL STANDARDS

Music. Last Updated: May 28, 2015, 11:49 am NORTH CAROLINA ESSENTIAL STANDARDS Grade: Kindergarten Course: al Literacy NCES.K.MU.ML.1 - Apply the elements of music and musical techniques in order to sing and play music with NCES.K.MU.ML.1.1 - Exemplify proper technique when singing

More information

Feature-Based Analysis of Haydn String Quartets

Feature-Based Analysis of Haydn String Quartets Feature-Based Analysis of Haydn String Quartets Lawson Wong 5/5/2 Introduction When listening to multi-movement works, amateur listeners have almost certainly asked the following situation : Am I still

More information

MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC

MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC Lena Quinto, William Forde Thompson, Felicity Louise Keating Psychology, Macquarie University, Australia lena.quinto@mq.edu.au Abstract Many

More information

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods Kazuyoshi Yoshii, Masataka Goto and Hiroshi G. Okuno Department of Intelligence Science and Technology National

More information

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You Chris Lewis Stanford University cmslewis@stanford.edu Abstract In this project, I explore the effectiveness of the Naive Bayes Classifier

More information

Modeling the Effect of Meter in Rhythmic Categorization: Preliminary Results

Modeling the Effect of Meter in Rhythmic Categorization: Preliminary Results Modeling the Effect of Meter in Rhythmic Categorization: Preliminary Results Peter Desain and Henkjan Honing,2 Music, Mind, Machine Group NICI, University of Nijmegen P.O. Box 904, 6500 HE Nijmegen The

More information

A Framework for Segmentation of Interview Videos

A Framework for Segmentation of Interview Videos A Framework for Segmentation of Interview Videos Omar Javed, Sohaib Khan, Zeeshan Rasheed, Mubarak Shah Computer Vision Lab School of Electrical Engineering and Computer Science University of Central Florida

More information

Computational Modelling of Harmony

Computational Modelling of Harmony Computational Modelling of Harmony Simon Dixon Centre for Digital Music, Queen Mary University of London, Mile End Rd, London E1 4NS, UK simon.dixon@elec.qmul.ac.uk http://www.elec.qmul.ac.uk/people/simond

More information

Differences in Metrical Structure Confound Tempo Judgments Justin London, August 2009

Differences in Metrical Structure Confound Tempo Judgments Justin London, August 2009 Presented at the Society for Music Perception and Cognition biannual meeting August 2009. Abstract Musical tempo is usually regarded as simply the rate of the tactus or beat, yet most rhythms involve multiple,

More information

A PRELIMINARY COMPUTATIONAL MODEL OF IMMANENT ACCENT SALIENCE IN TONAL MUSIC

A PRELIMINARY COMPUTATIONAL MODEL OF IMMANENT ACCENT SALIENCE IN TONAL MUSIC A PRELIMINARY COMPUTATIONAL MODEL OF IMMANENT ACCENT SALIENCE IN TONAL MUSIC Richard Parncutt Centre for Systematic Musicology University of Graz, Austria parncutt@uni-graz.at Erica Bisesi Centre for Systematic

More information

AP MUSIC THEORY 2016 SCORING GUIDELINES

AP MUSIC THEORY 2016 SCORING GUIDELINES AP MUSIC THEORY 2016 SCORING GUIDELINES Question 1 0---9 points Always begin with the regular scoring guide. Try an alternate scoring guide only if necessary. (See I.D.) I. Regular Scoring Guide A. Award

More information

2005 BY THE REGENTS OF THE UNIVERSITY OF CALIFORNIA. The Influence of Pitch Interval on the Perception of Polyrhythms

2005 BY THE REGENTS OF THE UNIVERSITY OF CALIFORNIA. The Influence of Pitch Interval on the Perception of Polyrhythms Music Perception Spring 2005, Vol. 22, No. 3, 425 440 2005 BY THE REGENTS OF THE UNIVERSITY OF CALIFORNIA ALL RIGHTS RESERVED. The Influence of Pitch Interval on the Perception of Polyrhythms DIRK MOELANTS

More information

Transcription An Historical Overview

Transcription An Historical Overview Transcription An Historical Overview By Daniel McEnnis 1/20 Overview of the Overview In the Beginning: early transcription systems Piszczalski, Moorer Note Detection Piszczalski, Foster, Chafe, Katayose,

More information

The Human, the Mechanical, and the Spaces in between: Explorations in Human-Robotic Musical Improvisation

The Human, the Mechanical, and the Spaces in between: Explorations in Human-Robotic Musical Improvisation Musical Metacreation: Papers from the 2013 AIIDE Workshop (WS-13-22) The Human, the Mechanical, and the Spaces in between: Explorations in Human-Robotic Musical Improvisation Scott Barton Worcester Polytechnic

More information

Estimating the Time to Reach a Target Frequency in Singing

Estimating the Time to Reach a Target Frequency in Singing THE NEUROSCIENCES AND MUSIC III: DISORDERS AND PLASTICITY Estimating the Time to Reach a Target Frequency in Singing Sean Hutchins a and David Campbell b a Department of Psychology, McGill University,

More information

Analysis and Clustering of Musical Compositions using Melody-based Features

Analysis and Clustering of Musical Compositions using Melody-based Features Analysis and Clustering of Musical Compositions using Melody-based Features Isaac Caswell Erika Ji December 13, 2013 Abstract This paper demonstrates that melodic structure fundamentally differentiates

More information

Creating a Feature Vector to Identify Similarity between MIDI Files

Creating a Feature Vector to Identify Similarity between MIDI Files Creating a Feature Vector to Identify Similarity between MIDI Files Joseph Stroud 2017 Honors Thesis Advised by Sergio Alvarez Computer Science Department, Boston College 1 Abstract Today there are many

More information

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION ULAŞ BAĞCI AND ENGIN ERZIN arxiv:0907.3220v1 [cs.sd] 18 Jul 2009 ABSTRACT. Music genre classification is an essential tool for

More information

Speaking in Minor and Major Keys

Speaking in Minor and Major Keys Chapter 5 Speaking in Minor and Major Keys 5.1. Introduction 28 The prosodic phenomena discussed in the foregoing chapters were all instances of linguistic prosody. Prosody, however, also involves extra-linguistic

More information

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm Georgia State University ScholarWorks @ Georgia State University Music Faculty Publications School of Music 2013 Chords not required: Incorporating horizontal and vertical aspects independently in a computer

More information

Rhythm related MIR tasks

Rhythm related MIR tasks Rhythm related MIR tasks Ajay Srinivasamurthy 1, André Holzapfel 1 1 MTG, Universitat Pompeu Fabra, Barcelona, Spain 10 July, 2012 Srinivasamurthy et al. (UPF) MIR tasks 10 July, 2012 1 / 23 1 Rhythm 2

More information

Measurement of overtone frequencies of a toy piano and perception of its pitch

Measurement of overtone frequencies of a toy piano and perception of its pitch Measurement of overtone frequencies of a toy piano and perception of its pitch PACS: 43.75.Mn ABSTRACT Akira Nishimura Department of Media and Cultural Studies, Tokyo University of Information Sciences,

More information

The Role of Accent Salience and Joint Accent Structure in Meter Perception

The Role of Accent Salience and Joint Accent Structure in Meter Perception Journal of Experimental Psychology: Human Perception and Performance 2009, Vol. 35, No. 1, 264 280 2009 American Psychological Association 0096-1523/09/$12.00 DOI: 10.1037/a0013482 The Role of Accent Salience

More information

Detecting Musical Key with Supervised Learning

Detecting Musical Key with Supervised Learning Detecting Musical Key with Supervised Learning Robert Mahieu Department of Electrical Engineering Stanford University rmahieu@stanford.edu Abstract This paper proposes and tests performance of two different

More information

Effects of Auditory and Motor Mental Practice in Memorized Piano Performance

Effects of Auditory and Motor Mental Practice in Memorized Piano Performance Bulletin of the Council for Research in Music Education Spring, 2003, No. 156 Effects of Auditory and Motor Mental Practice in Memorized Piano Performance Zebulon Highben Ohio State University Caroline

More information

Chord Classification of an Audio Signal using Artificial Neural Network

Chord Classification of an Audio Signal using Artificial Neural Network Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------

More information

An Integrated Music Chromaticism Model

An Integrated Music Chromaticism Model An Integrated Music Chromaticism Model DIONYSIOS POLITIS and DIMITRIOS MARGOUNAKIS Dept. of Informatics, School of Sciences Aristotle University of Thessaloniki University Campus, Thessaloniki, GR-541

More information

Week 14 Music Understanding and Classification

Week 14 Music Understanding and Classification Week 14 Music Understanding and Classification Roger B. Dannenberg Professor of Computer Science, Music & Art Overview n Music Style Classification n What s a classifier? n Naïve Bayesian Classifiers n

More information

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION Halfdan Rump, Shigeki Miyabe, Emiru Tsunoo, Nobukata Ono, Shigeki Sagama The University of Tokyo, Graduate

More information

The Tone Height of Multiharmonic Sounds. Introduction

The Tone Height of Multiharmonic Sounds. Introduction Music-Perception Winter 1990, Vol. 8, No. 2, 203-214 I990 BY THE REGENTS OF THE UNIVERSITY OF CALIFORNIA The Tone Height of Multiharmonic Sounds ROY D. PATTERSON MRC Applied Psychology Unit, Cambridge,

More information

Tonal Cognition INTRODUCTION

Tonal Cognition INTRODUCTION Tonal Cognition CAROL L. KRUMHANSL AND PETRI TOIVIAINEN Department of Psychology, Cornell University, Ithaca, New York 14853, USA Department of Music, University of Jyväskylä, Jyväskylä, Finland ABSTRACT:

More information

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis Automatic characterization of ornamentation from bassoon recordings for expressive synthesis Montserrat Puiggròs, Emilia Gómez, Rafael Ramírez, Xavier Serra Music technology Group Universitat Pompeu Fabra

More information

THE SOUND OF SADNESS: THE EFFECT OF PERFORMERS EMOTIONS ON AUDIENCE RATINGS

THE SOUND OF SADNESS: THE EFFECT OF PERFORMERS EMOTIONS ON AUDIENCE RATINGS THE SOUND OF SADNESS: THE EFFECT OF PERFORMERS EMOTIONS ON AUDIENCE RATINGS Anemone G. W. Van Zijl, Geoff Luck Department of Music, University of Jyväskylä, Finland Anemone.vanzijl@jyu.fi Abstract Very

More information

Auditory Illusions. Diana Deutsch. The sounds we perceive do not always correspond to those that are

Auditory Illusions. Diana Deutsch. The sounds we perceive do not always correspond to those that are In: E. Bruce Goldstein (Ed) Encyclopedia of Perception, Volume 1, Sage, 2009, pp 160-164. Auditory Illusions Diana Deutsch The sounds we perceive do not always correspond to those that are presented. When

More information

CSC475 Music Information Retrieval

CSC475 Music Information Retrieval CSC475 Music Information Retrieval Monophonic pitch extraction George Tzanetakis University of Victoria 2014 G. Tzanetakis 1 / 32 Table of Contents I 1 Motivation and Terminology 2 Psychacoustics 3 F0

More information

Melody Retrieval On The Web

Melody Retrieval On The Web Melody Retrieval On The Web Thesis proposal for the degree of Master of Science at the Massachusetts Institute of Technology M.I.T Media Laboratory Fall 2000 Thesis supervisor: Barry Vercoe Professor,

More information

Statistical learning and probabilistic prediction in music cognition: mechanisms of stylistic enculturation

Statistical learning and probabilistic prediction in music cognition: mechanisms of stylistic enculturation Ann. N.Y. Acad. Sci. ISSN 0077-8923 ANNALS OF THE NEW YORK ACADEMY OF SCIENCES Special Issue: The Neurosciences and Music VI ORIGINAL ARTICLE Statistical learning and probabilistic prediction in music

More information

AP Music Theory. Sample Student Responses and Scoring Commentary. Inside: Free Response Question 1. Scoring Guideline.

AP Music Theory. Sample Student Responses and Scoring Commentary. Inside: Free Response Question 1. Scoring Guideline. 2017 AP Music Theory Sample Student Responses and Scoring Commentary Inside: Free Response Question 1 Scoring Guideline Student Samples Scoring Commentary 2017 The College Board. College Board, Advanced

More information

SHORT TERM PITCH MEMORY IN WESTERN vs. OTHER EQUAL TEMPERAMENT TUNING SYSTEMS

SHORT TERM PITCH MEMORY IN WESTERN vs. OTHER EQUAL TEMPERAMENT TUNING SYSTEMS SHORT TERM PITCH MEMORY IN WESTERN vs. OTHER EQUAL TEMPERAMENT TUNING SYSTEMS Areti Andreopoulou Music and Audio Research Laboratory New York University, New York, USA aa1510@nyu.edu Morwaread Farbood

More information

PERFORMING ARTS Curriculum Framework K - 12

PERFORMING ARTS Curriculum Framework K - 12 PERFORMING ARTS Curriculum Framework K - 12 Litchfield School District Approved 4/2016 1 Philosophy of Performing Arts Education The Litchfield School District performing arts program seeks to provide

More information

Query By Humming: Finding Songs in a Polyphonic Database

Query By Humming: Finding Songs in a Polyphonic Database Query By Humming: Finding Songs in a Polyphonic Database John Duchi Computer Science Department Stanford University jduchi@stanford.edu Benjamin Phipps Computer Science Department Stanford University bphipps@stanford.edu

More information

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES Vishweshwara Rao and Preeti Rao Digital Audio Processing Lab, Electrical Engineering Department, IIT-Bombay, Powai,

More information

LESSON 1 PITCH NOTATION AND INTERVALS

LESSON 1 PITCH NOTATION AND INTERVALS FUNDAMENTALS I 1 Fundamentals I UNIT-I LESSON 1 PITCH NOTATION AND INTERVALS Sounds that we perceive as being musical have four basic elements; pitch, loudness, timbre, and duration. Pitch is the relative

More information

Automatic Laughter Detection

Automatic Laughter Detection Automatic Laughter Detection Mary Knox Final Project (EECS 94) knoxm@eecs.berkeley.edu December 1, 006 1 Introduction Laughter is a powerful cue in communication. It communicates to listeners the emotional

More information

Evaluating Melodic Encodings for Use in Cover Song Identification

Evaluating Melodic Encodings for Use in Cover Song Identification Evaluating Melodic Encodings for Use in Cover Song Identification David D. Wickland wickland@uoguelph.ca David A. Calvert dcalvert@uoguelph.ca James Harley jharley@uoguelph.ca ABSTRACT Cover song identification

More information