A hierarchical self-organizing map model for sequence recognition

Size: px
Start display at page:

Download "A hierarchical self-organizing map model for sequence recognition"

Transcription

1 A hierarchical self-organizing map model for sequence recognition Otávio Augusto S. Carpinteiro Instituto de Engenharia Elétrica Escola Federal de Engenharia de Itajubá Av. BPS 1303, Itajubá, MG, , Brazil h-page: phone: fax: March 1999 Keywords: artificial intelligence, artificial neural networks, pattern recognition, self-organizing map.

2 Abstract The paper presents the analysis of an original hierarchical neural model on a complex sequence the complete sixteenth four-part fugue in G minor of the Well-Tempered Clavier (vol. I) of J. S. Bach. The model makes an effective use of context information, through its hierarchical topology and its embedded time integrators, and that enables it to keep a very good account of past events. The model performs sequence classification and discrimination efficiently. It has application in domains which require pattern recognition, or particularly, which demand recognizing either a set of sequences of vectors in time or sub-sequences into a unique and large sequence of vectors in time.

3 1. Introduction Many researchers have developed artificial neural models to classify sequences in time. Several models based on Kohonen s self-organizing feature map [1], nonetheless, have faced some wellknown flaws. In windowed data models, as in Kangas model [2], the input vectors are concatenated in a fixed-size window which slides in time. The size of the memory to the past inputs is limited to the size of the window. The most serious deficiency of such models is that they become computationally expensive as wider windows are required. In time integral models, as in Chappell and Taylor s model [3], the activation of a unit is a combination of its current input and its former outputs decayed in time. Such models suffer from loss of context. Sequences that slightly differ in their initial elements (e.g., abccc, baccc, and bbccc) would probably have the same classification. In non-orthodox models, neither windows nor time integrators are employed. In James and Miikkulainen s model [4], for instance, when a vector in the sequence is input, the output unit in the map which wins the competition for that vector is disabled for further competition, and its activation decays in time to indicate which vector in the sequence it is representing. Each winning output unit represents just a single vector in the input sequence, and so, the representation for the whole input sequence is given by a sequence of winning output units in the map. This distributed representation is more complicated because it has to take into consideration not only the winning units but also their activations. For example, input sequences with the same elements but in different order (e.g., abcd and dcba) will have the same winning units, and so, one has to look at their activations to verify which input sequence the map is representing. Another disadvantage is that the model is unable to recognize sub-sequences inside a large and unique input sequence. The reason is that, after winning the competition in the map, the winning units are disabled for further competitions, and consequently, identical or similar sub-sequences inserted into different points of the large sequence will never hold identical or similar representations in the map. The model presented here is also based on Kohonen s map. However, thanks to its hierarchical topology and its embedded time integrators, it makes an efficient use of context information which prevents it from suffering from the flaws mentioned above. Artificial neural models have been widely employed in the musical domain. Among them, we may cite models for pitch perception [5, 6], chord perception [7], tonal perception [8, 9, 10, 11, 12], musical time perception [13], perception of musical sequences [14], musical pattern categorizations [15, 16], and musical composition [17, 18, 19, 20, 21]. Gjerdingen s work [15, 16] is the closest to the area of musical sequence recognition. His neural model was required to classify groups of notes produced by segmentation into categories. Unfortunately though, nothing was mentioned about existence or not of similar sequences of patterns in the training set, so that one could not assess performance of the net in terms of classifications and misclassifications of such sequences.

4 This paper reports our final research in neural models applied to the musical domain. The initial results were published in [22]. The research reported here, however, is much more relevant, for the musical sequences employed in the studies were much more complex than those employed in [22], and the results obtained and conclusions are also much more significant. In [22], we applied the model to a single part (or line) of a fugue of J. S. Bach. Here, the model is trained and evaluated on a more complex sequence in time a complete four-part fugue of J. S. Bach. In the following sections, we shall describe the representation, model, and the experiment on the musical sequences. 2. Representation for the sequences The concept behind the representation is the division of a musical piece into equal size time intervals. The small figure in a musical sequence is set to be the time interval (TI). Thus, all other figures in the sequence are multiples of TI. For example, if TI is an eighth note, a quarter note lasts two TIs, a half note lasts four TIs, and so on. A time interval counter (TIC) may also be defined. One TIC lasts one TI. TIC is the unit in which the musical sequence is measured. Therefore, at each TIC, either there is a rest, or a note onset, or a note sustained. The input data in the experiment consists of a sequence of musical intervals, which corresponds to the Bach s fugue. Data is input one TIC at a time. Fifteen neural units are used in the representation. Each unit represents one musical interval ranging from an octave down to an octave up. We assume here, therefore, that listeners are able to separate out voices if intervals between notes in the voices are greater than a fifteenth. When there is a rest, none of the input units receives activation. Otherwise, when a note is onset or sustained, the unit corresponding to the interval receives activation. The representation for musical sequences takes into consideration all musical voices, and therefore, is complex because the voices interact. The representation assumes three facts. First, any note onset occurring in a TIC makes up an interval with all notes onset or sustained which occurred in the TIC immediately before. Second, the representation does represent the intervals which occur in a TIC, but does not represent multiple instances occurring through the voices in the TIC. Third, at any given TIC, an interval corresponding to a note onset masks any occurrence of the same interval corresponding to a note sustained. Let us consider the multivoiced musical sequence in figure 1, and a time integrator with decay rate of 0.5 applied to input units. The sequence lasts 8 TICs, and TI is a sixteenth note. Units corresponding to intervals made up by a note onset or note sustained receive an activation of 1.0 and 0.5 respectively. Table 1 displays then the integrated value of activations of the units when inputting, TIC by TIC, the sequence in figure 1.

5 4 G 8 < 4 I : !)! -!)!)!)! -!(! -?? Figure 1: A multivoiced musical sequence TIC Table 1: Representation for the musical sequence in figure 1 Activations of Units Representing Intervals -8-7 ::: ::: ::: ::: 2 ::: 1.0 ::: ::: ::: ::: ::: ::: ::: ::: ::: ::: ::: ::: ::: The model The model is made up of two self-organizing maps (SOMs), as shown in figure 2. Its features, performance, and potential are better evaluated in [22, 23]. The problem of loss of context which occurs in other models, the analysis of our model, the analysis of Kohonen s SOM model, and the comparison of our model with Kohonen s in a simple temporal pattern recognition problem are reported in [23]. In this paper, we made use of contrived sequences to explain the behaviour of the two models in the presence of context information. We explained the effect of the time integrators on the results produced as well. In our model, the bottom SOM is responsible for making up representations which take into consideration not only the information given by the input vectors, but also the information given by the context in which these vectors are inserted. Therefore, the information passed up to the top SOM is much more acurate, and consequently, so are the classifications. We finally show that our model has much better performance than that of Kohonen s.

6 Top SOM Map Time Integrator Λ Bottom SOM Map Time Integrator V(t ) Figure 2: The model The input to the model is a sequence in time of m-dimensional vectors, S 1 = V(1), V(2),..., V(t),..., V(z), where the components of each vector are non-negative real values. The sequence is presented to the input layer of the bottom SOM, one vector at a time. The input layer has m units, one for each component of the input vector V(t), and a time integrator. The activation X(t) of the units in the input layer is given by X( t) = V( t) + δ 1X( t 1) (1) where δ 1 (0,1) is the decay rate. For each input vector X(t), the winning unit i * (t) in the map 1 is the unit which has the smallest distance Ψ(i,t). For each output unit i, Ψ(i,t) is given by the Euclidean distance between the input vector X(t) and the unit s weight vector W i. Each output unit i in the neighbourhood N * (t) of the winning unit i * (t) has its weight W i updated by W ( t + 1) = W ( t) + α ϒ ( i)[ X( t) Wi( t)] (2) i i where α (0,1) is the learning rate. ϒ(i) is the neighbourhood interaction function [24], a Gaussian type function, and is given by ϒ () i = κ + κ e 1 2 * 2 κ 3[ Φ( ii, ( t))] 2 σ 2 (3) where κ 1, κ 2, and κ 3 are constants, σ is the radius of the neighbourhood N * (t), and Φ(i,i * (t)) is the distance in the map between the unit i and the winning unit i * (t). The distance Φ(i,i ) between any two units i and i in the map is calculated according to the maximum norm, { } ' '' ' '' ' '' Φ(i, i ) = max l l, c c (4) where (l,c ) and (l,c ) are the coordinates of the units i and i respectively in the map.

7 The input to the top SOM is determined by the distances Φ(i,i * (t)) of the n units in the map of the bottom SOM. The input is thus a sequence in time of n-dimensional vectors, S 2 = Λ(Φ(i,i * (1))), Λ(Φ(i,i * (2))),..., Λ(Φ(i,i * (t))),..., Λ(Φ(i,i * (z))), where Λ is a n-dimensional transfer function on a n-dimensional space domain. Λ is defined as * * 1 κφ( ii, ( t)) if i N ( t) * ΛΦ ( ( ii, ( t))) = 0 otherwise (5) where κ is a constant, and N * (t) is a neighbourhood of the winning unit. The sequence S 2 is then presented to the input layer of the top SOM, one vector at a time. The input layer has n units, one for each component of the input vector Λ(Φ(i,i * (t))), and a time integrator. The activation X(t) of the units in the input layer is thus given by where δ 2 (0,1) is the decay rate. * X() t = ΛΦ ( ( i, i ())) t + δ2x( t 1) (6) The dynamics of the top SOM is identical to that of the bottom SOM. 4. The experiment We wanted the model to be tested on musical sequences, because the musical domain set three strong conditions on the model. First, that the model were able to recognize both a set of input sequences and a set of sub-sequences within a large and unique input sequence. The model was required to recognize a set of input sequences when the whole sequence was segmented. Otherwise, the entire piece consisted of a unique input sequence, and the model was thus required to recognize sub-sequences of that sequence. Second, that the model classified sequences (or sub-sequences) properly in the presence of noise. The reason followed from the fact that any two sequences which differed slightly had to achieve similar classifications. Third, that the model recognized sequences (or sub-sequences) in a very precise form. The reason for the latter is that any two sequences which shared either some intervals, or even all intervals, but in an alternative order or rhythm, were musically different, and as a consequence, had to be recognized as distinct. The experiment was on recognition of the instances of a theme occurring in a complex sequence the complete sixteenth four-part fugue in G minor of the first volume of The Well-Tempered Clavier of Bach. The fugue had 544 TICs, and TI was a sixteenth note. The theme of the fugue is shown in figure 3.

8 I 22 S?!! [ [! 4!! theme?!!!! PNN! -!... Figure 3: Theme of the sixteenth fugue in G minor The fugue in G minor was chosen for several reasons. First, as many of fugues of Bach, it has four voices. Second, it possesses many perfect and modified instances of theme. Third, it includes two cases of stretto 2. One case occurs between its seventeenth and eighteenth bars, in which two instances of theme overlap, and the other case occurs between its twenty eighth and thirtieth bars, in which three instances of theme overlap. Fourth, the thematic material of the theme is extensively developed throughout the fugue. Such developments 3, although quite similar to the theme, are not instances of theme. Fifth, very common intervals, as seconds up and down, occur extensively in the theme as well as in many passages in the fugue. All facts above are usually present in real situations, for example, in those in which humans are asked to perform thematic recognition in musical domains. Apart from providing a typical situation in a real domain, such facts also increase very much the level of difficulty of the domain to which the artificial neural model is applied. The input data consisted of two sets, hereafter referred to as input set I and input set II. Input set I consisted of a large and unique sequence of musical intervals, which corresponded to the fugue. Input set II contained many sequences, which were formed by segmenting the fugue whenever there were rests. The experiment pursued two aims. Firstly, to determine whether the model recognizes all instances of the theme in the fugue. Secondly, to determine whether any other sequence (or sub-sequence), which was not an instance, was not misclassified as theme. The training of the two SOMs of the model took place in two phases coarse-mapping and fine-tuning. The initial learning rate was set to 0.5, and the initial size of the neighbourhood was set to the size of the map. In the coarse-mapping phase, the learning rate and the radius of the neighbourhood were reduced linearly whereas in the fine-tuning phase, they were kept constant in 0.01 and 1 respectively. The coarse-mapping phase took 20%, and the fine-tuning phase took 80% of the total number of epochs. The initial weights were given randomly, in the range between 0 and 0.1, to both SOMs. Different values for decay rate were tested. In the bottom SOM of model II, it varied from 0.1 to 0.7, and in the top SOM, from 0.5 to 0.9. We present here, however, only the results using decay rates of 0.3 and 0.7 for the bottom and top SOM respectively. Such results, reached with those decay rates, were the best for all studies performed. The bottom SOM of model II was tested with map size of 15 15, and was trained in 700 epochs. The top SOM was tested with map size of 18 18, and trained in 850 epochs. Two

9 transfer functions Λ were tested. The first was given by equation 5, with neighbourhood N * (t) = {i Φ(i,i * (t)) < 2}, and κ = 0.5. The second was also given by equation 5, but with neighbourhood N * (t) = {i Φ(i,i * (t)) < 4}, and κ = We report here studies using the second transfer function only, for it produced the better results. The input layer of the bottom SOM of model II held fifteen units, one for each musical interval ranging from an octave down to an octave up. The representation employed in these units is fully described in section 2. The experiment comprised five studies. In the last four, in order to study the effect of noise on the classifications, reinforcement in activation was given to input units when representing instances of theme. For example, in the second experiment, note onset and note sustained received activations of 0.1 and 0.07 respectively. When corresponding to instances of the theme, they received instead, activations of 0.5 and 0.35 respectively. Table 2 shows the activation values of notes onset and sustained, whether reinforced or not, as well as the input set employed in each study 4. Table 2: Parameter values of the studies Study Input Reinforcement Note Note N. Onset N. Sustained Set Value Onset Sustained (Reinforced) (Reinforced) I I II I III I IV I V II A sequence (or sub-sequence) S a is said to have the same classification as that of the * * theme S t if the distance Φ(ia(), z it ()) z < 2, where i * a () z and i * t ( z) are the last winning units of Sa and S t. In case of S a be also an instance of the theme, the mean error of the instance S a, Ê(S a ), is then given by '' ta * * ( ia( z), it ( z)) Ê( S a ) = Φ '' ' ' ( t t + 1) z= ta a a (7) ' '' where t a and t a are the initial and final TICs of the instance Sa 5. The total mean error is given by the sum of the mean errors of all instances. We have selected some instances of theme to present here. Those instances hold some peculiarities, whether in terms of similarity with the theme, or convergence to the theme classification, or context in which they were inserted, which make them representatives among the whole set of instances.

10 Figures 4 to 10 plot, for each selected instance of theme, the distances between each winning unit of the instance and its corresponding winning unit in the theme for each study carried out on the model. We used equation 4 to measure those distances in the map of the top SOM. Table 3 displays the main characteristics of the instances of theme. Table 4 shows, for each study, the total error, and the mean errors of each instance of theme. Table 5 displays the classifications and misclassifications of the studies. Figure 11 plots the total mean error of classifications in accordance with reinforcement provided in the first four studies study I study II study III study IV study V Distance Figure 4: Classifications of the first instance of theme (TICs 33 50) relative to theme. The instance occurs in the fourth voice the highest one concurrently with another voice. The instance differs from the theme in its first two TICs. The classifications of the instance yielded by the model only converge to that of the theme in the fourth and fifth studies. TIC

11 study I study II study III study IV study V Distance Figure 5: Classifications of the seventh instance of theme (TICs ) relative to theme. The instance occurs in the first voice, concurrently with three other voices. The instance is a perfect copy of the theme. The classifications of the instance yielded by the model do not converge to that of the theme in any of the studies. TIC

12 study I study II study III study IV study V Distance Figure 6: Classifications of the eighth instance of theme (TICs ) relative to theme. The instance occurs in the third voice, and is a perfect copy of the theme. At its beginning, the eighth instance occurs concurrently with three other voices. From TIC 283 onwards, it occurs concurrently with just one voice. The classifications of the instance produced by the model only converge to that of the theme in the second, third, fourth, and fifth studies. TIC

13 study I study II study III study IV study V Distance Figure 7: Classifications of the eleventh instance of theme (TICs ) relative to theme. The instance occurs in the third voice, concurrently with two other voices. The instance differs from the theme in its first two TICs, and in two TICs in its middle. The classifications of the instance yielded by the model only converge to that of the theme in the fifth study. TIC

14 study I study II study III study IV study V Distance Figure 8: Classifications of the twelfth instance of theme (TICs ) relative to theme. The instance occurs in the fourth voice, and is a perfect copy of the theme. At its beginning, the instance occurs unaccompanied. From TIC 443 onwards, it occurs concurrently with two other voices, and from TIC 451 onwards, it occurs concurrently with three other voices. The classifications of the instance produced by the model do not converge to that of the theme in any of the studies. TIC

15 study I study II study III study IV study V Distance Figure 9: Classifications of the thirteenth instance of theme (TICs ) relative to theme. The instance occurs in the second voice, and is a perfect copy of the theme. At its beginning, the instance occurs concurrently with two other voices. From TIC 451 onwards, it occurs concurrently with three other voices, and from TIC 459 onwards, it occurs concurrently with just one voice. The classifications of the instance yielded by the model only converge to that of the theme in the second, third, fourth, and fifth studies. TIC

16 study I study II study III study IV study V Distance Figure 10: Classifications of the fourteenth instance of theme (TICs ) relative to theme. The instance occurs in the first voice, concurrently with three other voices. The instance is a perfect copy of the theme. The classifications of the instance produced by the model do not converge to that of the theme in any of the studies. TIC

17 Table 3: Main characteristics of the instances of theme Voice: voice in which the instance occurs (1: lowest; 4: highest); No. Voices: number of voices sounding concurrently (including the one in which the instance occurs); Dif. TICs: difference in number of TICs between the instance and the theme; Conv.: studies in which the instance converges to the classification of the theme; TIC: TICs in which the instance occurs; Voice No. V oices Dif.TICs Conv. TIC 1 3 4/5 73{ /5 209{ /5 313{ { { /4/2 2/3/4/5 449{ /5 97{ /5 521{ /4/5 185{ /2 2/3/4/5 273{ { /4/5 497{ /5 33{ /5 337{ /5 233{ /3/4 441{456

18 Table 4: Mean errors of the instances of theme TIC oftheerrors Instances Mean Study: I Study: II Study: III Study: IV Study: V 33{ { { { { { { { { { { { { { { { Total Table 5: Classifications and misclassifications of the model Study No. Hits Classications. No F ailures No. Miscl. Minor Misclassications No. Miscl. jor Ma I II III IV V

19 Total Mean Error Reinforcement Figure 11: Total mean error of classifications

20 Some conclusions may be drawn from the results. First, as displayed in table 5, the model held a high number of misclassifications in the third study. Such a high number was due to the fact that the model classified an intermediate part of theme as its final part, and consequently, kept on misclassifying intermediate parts of instances of theme as their final parts as well. Second, as shown in tables 4 and 5, the model exhibited a similar good performance in the fourth and fifth studies. In the fourth, the model exhibited lower mean errors, whereas in the fifth study, it presented better results for classifications and misclassifications. The model satisfied, therefore, the first of our required conditions as stated in section 4, i.e., it was able to perform efficiently both on a set of input sequences and on a set of sub-sequences within a large and unique input sequence. Third, by analysing the results displayed in the figures 4 to 11, and in the table 3, one may observe that the model was fault tolerant to errors, satisfying therefore, the second of our required conditions as stated in section 4. It classified properly several instances which differed slightly from the theme, whether in the pitch or in the duration of one single note. The model performed classification efficiently in the presence of noise as well. When instances of theme occurred concurrently with other polyphonic voices, the degree of noise was so high that it caused the model not to classify instances correctly. We may observe that in the results of studies I, II, and III. However, when thematic reinforcement was given to instances in a greater amount, the remaining polyphonic voices started playing roles of noisy backgrounds, and then, the model started classifying rightly instances of theme. That may be observed in the results of studies IV and V. Fourth, as it may be observed in figures 5, 8, 10, and in the table 3, the model failed, in all studies, in recognizing three instances of theme. It succeeded, however, as shown in figures 6 and 9, and in the table 3 in recognizing two other instances of theme in the last four studies. These instances, which occur between TICs 265 and 282, TICs 273 and 290, TICs 441 and 456, TICs 449 and 466, and TICs 457 and 460, overlapped through the voices, making up the two cases of stretto present in the fugue. One may conclude, therefore, that the recognition of strettos was not performed reasonably by the model. Fifth, the model presented a fairly high number of minor misclassifications in the fourth and fifth studies, according to the results in the table 5. Yet, we should not consider severely those results, because they are much more a case of `inertia' than a proper error. We considered a case of minor misclassification when the model kept on classifying as instance of theme the next few TICs which followed the instance. In respect to the major errors yet, the model performed perfectly, not producing anyone. The results thus reveal that the third of our required conditions, as stated in the section 4, was finely satisfied. Sixth, by comparing studies IV and V in tables 4 and 5, one may verify that there is not a significant difference between their results. The model exhibited lower mean errors in the study IV, but otherwise, it presented better results in terms of classifications and misclassifications in the study V. The straightforward conclusion which may be drawn is thus,

21 that the segmentation of the fugue on rests, i.e., the resetting of the model on inputs which corresponds to rests, does not yield any improvement in terms of classification. Seventh, by analysing the relation between number of voices sounding and the mean errors in the tables 3 and 4, one may observe that the recognition of an instance of theme becomes more difficult when the number of voices sounding simultaneously increases. The only exception is in the instance of theme occurring along TICs 449 to 466. The mean errors in the fourth and fifth studies are 6.56 and 8.28 respectively. They are much higher, for example, than those in the same studies occurring in the instances along TICs 97 to 114, and 521 to 538. Nevertheless, one must see that the instance of theme along TICs 449 to 466 is the one which, together with the instances along TICs 441 to 456, and 457 to 460, makes up the second stretto present in the fugue. Hence, the entrance of the first instance from TICs 441 to 456, and of the third instance from TICs 457 to 460, probably contributed to those high mean errors achieved. Finally, there seems not to be any relation between the voice in which the instance of theme occurs and the ease in recognizing it. Let us present three examples taken from the tables 3 and 4. The first is given by the instances I a and I b from TICs 185 to 201, and from 33 to 50 respectively. I a occurs in the third voice, and I b in the fourth one. Both instances occur together with another voice. The mean errors of I a are lower than those of I b. The second example is produced by the instances I c and I d, from TICs 313 to 330, and 337 to 353 respectively. I c occurs in the first voice, I d in the fourth, and there are two other voices sounding concurrently with those instances. The mean errors of I c are higher than those of I d, with the exception of that one in the study II. The third example is given by the instances I e from TICs 521 to 538, and I f from TICs 233 to 250. I e occurs in the second voice, and I f in the fourth one. There are three other voices sounding concurrently with the instances. The mean errors of I e are lower than those of I f in the studies IV and V. Thus, the recognition of instances of theme was not facilitated when the instances occurred in any specific voice. 5. Conclusion An original representation for musical sequences, and an original artificial neural model for sequence classification is presented. The model has a topology made up of two self-organizing map networks, one on top of the other. It encodes and manipulates context information effectively, and that enables it to perform sequence classification and discrimination efficiently. The model has application in domains which demand classifying either a set of sequences of vectors or sub-sequences into a unique and large sequence of vectors in time. The results obtained have shown that the artificial neural model was able to perform efficiently sequence classification and discrimination. The model was able to classify properly most of the instances of theme occurring in the musical piece. It is worth noticing that the model performed classification even in the presence of noise, i.e., even when instances occurred modified, in different past contexts, and amidst different polyphonic voices in the musical piece.

22 The model could also discriminate instances of theme from sequences that shared some similarity with the theme. Very many of these pseudo-instance sequences occurred in the fugue. Rightly yet, the model did not classify them as instances of theme. The parameters of the model and the model itself are not, in any way, tied up to any specific application, whether it be a musical piece or any other domain. Indeed, we are currently doing further research on the model, applying it to the electrical engineering domain, to predict the demand for electrical load. The first results of this work are to be published in [25]. Footnotes 1 Also known as array, grid, or output layer. 2 Stretto is a musical passage where two or more instances of theme overlap. 3 In the fugal domain, the developments of thematic material take place in parts called episodes. 4 Reinforcement was provided from the seventh common TIC between the theme and any of its instances. 5 The initial TIC is the seventh TIC of the instance. References [1] T. Kohonen. Self-Organization and Associative Memory. Springer-Verlag, Berlin, third edition, [2] J. Kangas. On the Analysis of Pattern Sequences by Self-Organizing Maps. PhD thesis, Laboratory of Computer and Information Science, Helsinki University of Technology, Rakentajanaukio 2 C, SF-02150, Finland, [3] G. J. Chappell and J. G. Taylor. The temporal Kohonen map. Neural Networks, 6: , [4] D. L. James and R. Miikkulainen. SARDNET: a self-organizing feature map for sequences. In G. Tesauro, D. S. Touretzky, and T. K. Leen, editors, Proceedings of the Advances in Neural Information Processing Systems, volume 7. Morgan Kaufmann, [5] H. Sano and B. Jenkins. A neural network model for pitch perception. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [6] I. Taylor and M. Greenhough. Modelling pitch perception with adaptive resonance theory artificial neural networks. Connection Science, 6(2&3): , 1994.

23 [7] B. Laden and D. Keefe. The representation of pitch in a neural net model of chord classification. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [8] D. Scarborough, B. Miller, and J. Jones. Connectionist models for tonal analysis. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [9] M. Leman. The ontogenesis of tonal semantics: results of a computer study. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [10] J. Bharucha. Music cognition and perceptual facilitation: a connectionist framework. Music Perception, 5(1):1 30, [11] J. Bharucha. Pitch, harmony, and neural nets: a psychological perspective. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [12] J. Bharucha and P. Todd. Modeling the perception of tonal structure with neural nets. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [13] P. Desain and H. Honing. The quantization of musical time: a connectionist approach. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [14] M. Page. Modelling the perception of musical sequences with self-organizing neural networks. Connection Science, 6(2&3): , [15] R. Gjerdingen. Categorization of musical patterns by self-organizing neuronlike networks. Music Perception, 7(4): , [16] R. Gjerdingen. Using connectionist models to explore complex musical patterns. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [17] P. Todd. A connectionist approach to algorithmic composition. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [18] J. Lewis. Algorithms for music composition by neural nets: improved CBR paradigms. Proceedings of the International Computer Music Conference, Computer Music Association, pages , [19] J. Lewis. Creation by refinement and the problem of algorithmic music composition. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, 1991.

24 [20] M. Mozer. Connectionist music composition based on melodic, stylistic, and psychophysical constraints. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [21] M. Mozer and T. Soukup. Connectionist music composition based on melodic and stylistic constraints. In R. Lippmann, J. Moody, and D. Touretzky, editors, Proceedings of the Advances in Neural Information Processing Systems, Morgan Kaufmann, volume 3, pages , [22] O. A. S. Carpinteiro. A hierarchical self-organizing map model for sequence recognition. Neural Processing Letters, 9:1 12, [23] O. A. S. Carpinteiro. A hierarchical self-organizing map model for pattern recognition. In L. Caloba and J. Barreto, editors, Proceedings of the Brazilian Congress on Artificial Neural Networks 97 (CBRN 97), pages , UFSC, Florianópolis, SC, Brazil, [24] Z. Lo and B. Bavarian. Improved rate of convergence in Kohonen neural network. In Proceedings of the International Joint Conference on Neural Networks, volume 2, pages , July [25] O. A. S. Carpinteiro. A hierarchical self-organizing map model in short-term load forecasting. To appear in the Proceedings of the Fifth International Conference on Engineering Applications of Neural Networks (EANN), Warsaw, Poland, September

Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies

Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies Judy Franklin Computer Science Department Smith College Northampton, MA 01063 Abstract Recurrent (neural) networks have

More information

MELONET I: Neural Nets for Inventing Baroque-Style Chorale Variations

MELONET I: Neural Nets for Inventing Baroque-Style Chorale Variations MELONET I: Neural Nets for Inventing Baroque-Style Chorale Variations Dominik Hornel dominik@ira.uka.de Institut fur Logik, Komplexitat und Deduktionssysteme Universitat Fridericiana Karlsruhe (TH) Am

More information

Recurrent Neural Networks and Pitch Representations for Music Tasks

Recurrent Neural Networks and Pitch Representations for Music Tasks Recurrent Neural Networks and Pitch Representations for Music Tasks Judy A. Franklin Smith College Department of Computer Science Northampton, MA 01063 jfranklin@cs.smith.edu Abstract We present results

More information

NetNeg: A Connectionist-Agent Integrated System for Representing Musical Knowledge

NetNeg: A Connectionist-Agent Integrated System for Representing Musical Knowledge From: AAAI Technical Report SS-99-05. Compilation copyright 1999, AAAI (www.aaai.org). All rights reserved. NetNeg: A Connectionist-Agent Integrated System for Representing Musical Knowledge Dan Gang and

More information

Notes on David Temperley s What s Key for Key? The Krumhansl-Schmuckler Key-Finding Algorithm Reconsidered By Carley Tanoue

Notes on David Temperley s What s Key for Key? The Krumhansl-Schmuckler Key-Finding Algorithm Reconsidered By Carley Tanoue Notes on David Temperley s What s Key for Key? The Krumhansl-Schmuckler Key-Finding Algorithm Reconsidered By Carley Tanoue I. Intro A. Key is an essential aspect of Western music. 1. Key provides the

More information

Bach in a Box - Real-Time Harmony

Bach in a Box - Real-Time Harmony Bach in a Box - Real-Time Harmony Randall R. Spangler and Rodney M. Goodman* Computation and Neural Systems California Institute of Technology, 136-93 Pasadena, CA 91125 Jim Hawkinst 88B Milton Grove Stoke

More information

Algorithmic Music Composition

Algorithmic Music Composition Algorithmic Music Composition MUS-15 Jan Dreier July 6, 2015 1 Introduction The goal of algorithmic music composition is to automate the process of creating music. One wants to create pleasant music without

More information

Robert Alexandru Dobre, Cristian Negrescu

Robert Alexandru Dobre, Cristian Negrescu ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Automatic Music Transcription Software Based on Constant Q

More information

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t MPEG-7 FOR CONTENT-BASED MUSIC PROCESSING Λ Emilia GÓMEZ, Fabien GOUYON, Perfecto HERRERA and Xavier AMATRIAIN Music Technology Group, Universitat Pompeu Fabra, Barcelona, SPAIN http://www.iua.upf.es/mtg

More information

Distortion Analysis Of Tamil Language Characters Recognition

Distortion Analysis Of Tamil Language Characters Recognition www.ijcsi.org 390 Distortion Analysis Of Tamil Language Characters Recognition Gowri.N 1, R. Bhaskaran 2, 1. T.B.A.K. College for Women, Kilakarai, 2. School Of Mathematics, Madurai Kamaraj University,

More information

Computer Coordination With Popular Music: A New Research Agenda 1

Computer Coordination With Popular Music: A New Research Agenda 1 Computer Coordination With Popular Music: A New Research Agenda 1 Roger B. Dannenberg roger.dannenberg@cs.cmu.edu http://www.cs.cmu.edu/~rbd School of Computer Science Carnegie Mellon University Pittsburgh,

More information

A probabilistic approach to determining bass voice leading in melodic harmonisation

A probabilistic approach to determining bass voice leading in melodic harmonisation A probabilistic approach to determining bass voice leading in melodic harmonisation Dimos Makris a, Maximos Kaliakatsos-Papakostas b, and Emilios Cambouropoulos b a Department of Informatics, Ionian University,

More information

Hidden Markov Model based dance recognition

Hidden Markov Model based dance recognition Hidden Markov Model based dance recognition Dragutin Hrenek, Nenad Mikša, Robert Perica, Pavle Prentašić and Boris Trubić University of Zagreb, Faculty of Electrical Engineering and Computing Unska 3,

More information

Extracting Significant Patterns from Musical Strings: Some Interesting Problems.

Extracting Significant Patterns from Musical Strings: Some Interesting Problems. Extracting Significant Patterns from Musical Strings: Some Interesting Problems. Emilios Cambouropoulos Austrian Research Institute for Artificial Intelligence Vienna, Austria emilios@ai.univie.ac.at Abstract

More information

Early Applications of Information Theory to Music

Early Applications of Information Theory to Music Early Applications of Information Theory to Music Marcus T. Pearce Centre for Cognition, Computation and Culture, Goldsmiths College, University of London, New Cross, London SE14 6NW m.pearce@gold.ac.uk

More information

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS Andrew N. Robertson, Mark D. Plumbley Centre for Digital Music

More information

Experiments on musical instrument separation using multiplecause

Experiments on musical instrument separation using multiplecause Experiments on musical instrument separation using multiplecause models J Klingseisen and M D Plumbley* Department of Electronic Engineering King's College London * - Corresponding Author - mark.plumbley@kcl.ac.uk

More information

10 Visualization of Tonal Content in the Symbolic and Audio Domains

10 Visualization of Tonal Content in the Symbolic and Audio Domains 10 Visualization of Tonal Content in the Symbolic and Audio Domains Petri Toiviainen Department of Music PO Box 35 (M) 40014 University of Jyväskylä Finland ptoiviai@campus.jyu.fi Abstract Various computational

More information

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu

More information

Tonal Cognition INTRODUCTION

Tonal Cognition INTRODUCTION Tonal Cognition CAROL L. KRUMHANSL AND PETRI TOIVIAINEN Department of Psychology, Cornell University, Ithaca, New York 14853, USA Department of Music, University of Jyväskylä, Jyväskylä, Finland ABSTRACT:

More information

Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors *

Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors * Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors * David Ortega-Pacheco and Hiram Calvo Centro de Investigación en Computación, Instituto Politécnico Nacional, Av. Juan

More information

arxiv: v1 [cs.sd] 8 Jun 2016

arxiv: v1 [cs.sd] 8 Jun 2016 Symbolic Music Data Version 1. arxiv:1.5v1 [cs.sd] 8 Jun 1 Christian Walder CSIRO Data1 7 London Circuit, Canberra,, Australia. christian.walder@data1.csiro.au June 9, 1 Abstract In this document, we introduce

More information

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Róisín Loughran roisin.loughran@ul.ie Jacqueline Walker jacqueline.walker@ul.ie Michael O Neill University

More information

Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals

Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals Eita Nakamura and Shinji Takaki National Institute of Informatics, Tokyo 101-8430, Japan eita.nakamura@gmail.com, takaki@nii.ac.jp

More information

Similarity and Categorisation in Boulez Parenthèse from the Third Piano Sonata: A Formal Analysis.

Similarity and Categorisation in Boulez Parenthèse from the Third Piano Sonata: A Formal Analysis. Similarity and Categorisation in Boulez Parenthèse from the Third Piano Sonata: A Formal Analysis. Christina Anagnostopoulou? and Alan Smaill y y? Faculty of Music, University of Edinburgh Division of

More information

Pitch Spelling Algorithms

Pitch Spelling Algorithms Pitch Spelling Algorithms David Meredith Centre for Computational Creativity Department of Computing City University, London dave@titanmusic.com www.titanmusic.com MaMuX Seminar IRCAM, Centre G. Pompidou,

More information

APPENDIX A: ERRATA TO SCORES OF THE PLAYER PIANO STUDIES

APPENDIX A: ERRATA TO SCORES OF THE PLAYER PIANO STUDIES APPENDIX A: ERRATA TO SCORES OF THE PLAYER PIANO STUDIES Conlon Nancarrow s hand-written scores, while generally quite precise, contain numerous errors. Most commonly these are errors of omission (e.g.,

More information

MUSI-6201 Computational Music Analysis

MUSI-6201 Computational Music Analysis MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)

More information

On time: the influence of tempo, structure and style on the timing of grace notes in skilled musical performance

On time: the influence of tempo, structure and style on the timing of grace notes in skilled musical performance RHYTHM IN MUSIC PERFORMANCE AND PERCEIVED STRUCTURE 1 On time: the influence of tempo, structure and style on the timing of grace notes in skilled musical performance W. Luke Windsor, Rinus Aarts, Peter

More information

SYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS

SYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS Published by Institute of Electrical Engineers (IEE). 1998 IEE, Paul Masri, Nishan Canagarajah Colloquium on "Audio and Music Technology"; November 1998, London. Digest No. 98/470 SYNTHESIS FROM MUSICAL

More information

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You Chris Lewis Stanford University cmslewis@stanford.edu Abstract In this project, I explore the effectiveness of the Naive Bayes Classifier

More information

Various Artificial Intelligence Techniques For Automated Melody Generation

Various Artificial Intelligence Techniques For Automated Melody Generation Various Artificial Intelligence Techniques For Automated Melody Generation Nikahat Kazi Computer Engineering Department, Thadomal Shahani Engineering College, Mumbai, India Shalini Bhatia Assistant Professor,

More information

Harmony and tonality The vertical dimension. HST 725 Lecture 11 Music Perception & Cognition

Harmony and tonality The vertical dimension. HST 725 Lecture 11 Music Perception & Cognition Harvard-MIT Division of Health Sciences and Technology HST.725: Music Perception and Cognition Prof. Peter Cariani Harmony and tonality The vertical dimension HST 725 Lecture 11 Music Perception & Cognition

More information

Laboratory Assignment 3. Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB

Laboratory Assignment 3. Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB Laboratory Assignment 3 Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB PURPOSE In this laboratory assignment, you will use MATLAB to synthesize the audio tones that make up a well-known

More information

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance Methodologies for Expressiveness Modeling of and for Music Performance by Giovanni De Poli Center of Computational Sonology, Department of Information Engineering, University of Padova, Padova, Italy About

More information

Improving Performance in Neural Networks Using a Boosting Algorithm

Improving Performance in Neural Networks Using a Boosting Algorithm - Improving Performance in Neural Networks Using a Boosting Algorithm Harris Drucker AT&T Bell Laboratories Holmdel, NJ 07733 Robert Schapire AT&T Bell Laboratories Murray Hill, NJ 07974 Patrice Simard

More information

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception LEARNING AUDIO SHEET MUSIC CORRESPONDENCES Matthias Dorfer Department of Computational Perception Short Introduction... I am a PhD Candidate in the Department of Computational Perception at Johannes Kepler

More information

Etna Builder - Interactively Building Advanced Graphical Tree Representations of Music

Etna Builder - Interactively Building Advanced Graphical Tree Representations of Music Etna Builder - Interactively Building Advanced Graphical Tree Representations of Music Wolfgang Chico-Töpfer SAS Institute GmbH In der Neckarhelle 162 D-69118 Heidelberg e-mail: woccnews@web.de Etna Builder

More information

Categorization of ICMR Using Feature Extraction Strategy And MIR With Ensemble Learning

Categorization of ICMR Using Feature Extraction Strategy And MIR With Ensemble Learning Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 57 (2015 ) 686 694 3rd International Conference on Recent Trends in Computing 2015 (ICRTC-2015) Categorization of ICMR

More information

Reducing False Positives in Video Shot Detection

Reducing False Positives in Video Shot Detection Reducing False Positives in Video Shot Detection Nithya Manickam Computer Science & Engineering Department Indian Institute of Technology, Bombay Powai, India - 400076 mnitya@cse.iitb.ac.in Sharat Chandran

More information

Expressive performance in music: Mapping acoustic cues onto facial expressions

Expressive performance in music: Mapping acoustic cues onto facial expressions International Symposium on Performance Science ISBN 978-94-90306-02-1 The Author 2011, Published by the AEC All rights reserved Expressive performance in music: Mapping acoustic cues onto facial expressions

More information

An Empirical Comparison of Tempo Trackers

An Empirical Comparison of Tempo Trackers An Empirical Comparison of Tempo Trackers Simon Dixon Austrian Research Institute for Artificial Intelligence Schottengasse 3, A-1010 Vienna, Austria simon@oefai.at An Empirical Comparison of Tempo Trackers

More information

A Beat Tracking System for Audio Signals

A Beat Tracking System for Audio Signals A Beat Tracking System for Audio Signals Simon Dixon Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria. simon@ai.univie.ac.at April 7, 2000 Abstract We present

More information

arxiv: v1 [cs.lg] 15 Jun 2016

arxiv: v1 [cs.lg] 15 Jun 2016 Deep Learning for Music arxiv:1606.04930v1 [cs.lg] 15 Jun 2016 Allen Huang Department of Management Science and Engineering Stanford University allenh@cs.stanford.edu Abstract Raymond Wu Department of

More information

y POWER USER MUSIC PRODUCTION and PERFORMANCE With the MOTIF ES Mastering the Sample SLICE function

y POWER USER MUSIC PRODUCTION and PERFORMANCE With the MOTIF ES Mastering the Sample SLICE function y POWER USER MUSIC PRODUCTION and PERFORMANCE With the MOTIF ES Mastering the Sample SLICE function Phil Clendeninn Senior Product Specialist Technology Products Yamaha Corporation of America Working with

More information

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular Music Mood Sheng Xu, Albert Peyton, Ryan Bhular What is Music Mood A psychological & musical topic Human emotions conveyed in music can be comprehended from two aspects: Lyrics Music Factors that affect

More information

Each copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printed page of such transmission.

Each copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printed page of such transmission. Modeling the Perception of Tonal Structure with Neural Nets Author(s): Jamshed J. Bharucha and Peter M. Todd Source: Computer Music Journal, Vol. 13, No. 4 (Winter, 1989), pp. 44-53 Published by: The MIT

More information

Music Composition with RNN

Music Composition with RNN Music Composition with RNN Jason Wang Department of Statistics Stanford University zwang01@stanford.edu Abstract Music composition is an interesting problem that tests the creativity capacities of artificial

More information

Rhythm together with melody is one of the basic elements in music. According to Longuet-Higgins

Rhythm together with melody is one of the basic elements in music. According to Longuet-Higgins 5 Quantisation Rhythm together with melody is one of the basic elements in music. According to Longuet-Higgins ([LH76]) human listeners are much more sensitive to the perception of rhythm than to the perception

More information

Deep Neural Networks Scanning for patterns (aka convolutional networks) Bhiksha Raj

Deep Neural Networks Scanning for patterns (aka convolutional networks) Bhiksha Raj Deep Neural Networks Scanning for patterns (aka convolutional networks) Bhiksha Raj 1 Story so far MLPs are universal function approximators Boolean functions, classifiers, and regressions MLPs can be

More information

The Human, the Mechanical, and the Spaces in between: Explorations in Human-Robotic Musical Improvisation

The Human, the Mechanical, and the Spaces in between: Explorations in Human-Robotic Musical Improvisation Musical Metacreation: Papers from the 2013 AIIDE Workshop (WS-13-22) The Human, the Mechanical, and the Spaces in between: Explorations in Human-Robotic Musical Improvisation Scott Barton Worcester Polytechnic

More information

Predicting the immediate future with Recurrent Neural Networks: Pre-training and Applications

Predicting the immediate future with Recurrent Neural Networks: Pre-training and Applications Predicting the immediate future with Recurrent Neural Networks: Pre-training and Applications Introduction Brandon Richardson December 16, 2011 Research preformed from the last 5 years has shown that the

More information

Exploring the Rules in Species Counterpoint

Exploring the Rules in Species Counterpoint Exploring the Rules in Species Counterpoint Iris Yuping Ren 1 University of Rochester yuping.ren.iris@gmail.com Abstract. In this short paper, we present a rule-based program for generating the upper part

More information

CS229 Project Report Polyphonic Piano Transcription

CS229 Project Report Polyphonic Piano Transcription CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project

More information

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene Beat Extraction from Expressive Musical Performances Simon Dixon, Werner Goebl and Emilios Cambouropoulos Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria.

More information

Evaluating Melodic Encodings for Use in Cover Song Identification

Evaluating Melodic Encodings for Use in Cover Song Identification Evaluating Melodic Encodings for Use in Cover Song Identification David D. Wickland wickland@uoguelph.ca David A. Calvert dcalvert@uoguelph.ca James Harley jharley@uoguelph.ca ABSTRACT Cover song identification

More information

A Case Based Approach to the Generation of Musical Expression

A Case Based Approach to the Generation of Musical Expression A Case Based Approach to the Generation of Musical Expression Taizan Suzuki Takenobu Tokunaga Hozumi Tanaka Department of Computer Science Tokyo Institute of Technology 2-12-1, Oookayama, Meguro, Tokyo

More information

Outline. Why do we classify? Audio Classification

Outline. Why do we classify? Audio Classification Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification Implementation Future Work Why do we classify

More information

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC Scientific Journal of Impact Factor (SJIF): 5.71 International Journal of Advance Engineering and Research Development Volume 5, Issue 04, April -2018 e-issn (O): 2348-4470 p-issn (P): 2348-6406 MUSICAL

More information

Structure and Interpretation of Rhythm and Timing 1

Structure and Interpretation of Rhythm and Timing 1 henkjan honing Structure and Interpretation of Rhythm and Timing Rhythm, as it is performed and perceived, is only sparingly addressed in music theory. Eisting theories of rhythmic structure are often

More information

Melodic Pattern Segmentation of Polyphonic Music as a Set Partitioning Problem

Melodic Pattern Segmentation of Polyphonic Music as a Set Partitioning Problem Melodic Pattern Segmentation of Polyphonic Music as a Set Partitioning Problem Tsubasa Tanaka and Koichi Fujii Abstract In polyphonic music, melodic patterns (motifs) are frequently imitated or repeated,

More information

Bach-Prop: Modeling Bach s Harmonization Style with a Back- Propagation Network

Bach-Prop: Modeling Bach s Harmonization Style with a Back- Propagation Network Indiana Undergraduate Journal of Cognitive Science 1 (2006) 3-14 Copyright 2006 IUJCS. All rights reserved Bach-Prop: Modeling Bach s Harmonization Style with a Back- Propagation Network Rob Meyerson Cognitive

More information

Each copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printed page of such transmission.

Each copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printed page of such transmission. A Connectionist Approach to Algorithmic Composition Author(s): Peter M. Todd Source: Computer Music Journal, Vol. 13, No. 4 (Winter, 1989), pp. 27-43 Published by: The MIT Press Stable URL: http://www.jstor.org/stable/3679551

More information

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the

More information

Pattern Discovery and Matching in Polyphonic Music and Other Multidimensional Datasets

Pattern Discovery and Matching in Polyphonic Music and Other Multidimensional Datasets Pattern Discovery and Matching in Polyphonic Music and Other Multidimensional Datasets David Meredith Department of Computing, City University, London. dave@titanmusic.com Geraint A. Wiggins Department

More information

Jazz Melody Generation and Recognition

Jazz Melody Generation and Recognition Jazz Melody Generation and Recognition Joseph Victor December 14, 2012 Introduction In this project, we attempt to use machine learning methods to study jazz solos. The reason we study jazz in particular

More information

University of California Press is collaborating with JSTOR to digitize, preserve and extend access to Music Perception: An Interdisciplinary Journal.

University of California Press is collaborating with JSTOR to digitize, preserve and extend access to Music Perception: An Interdisciplinary Journal. An Algorithm for Harmonic Analysis Author(s): David Temperley Source: Music Perception: An Interdisciplinary Journal, Vol. 15, No. 1 (Fall, 1997), pp. 31-68 Published by: University of California Press

More information

Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment

Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment Gus G. Xia Dartmouth College Neukom Institute Hanover, NH, USA gxia@dartmouth.edu Roger B. Dannenberg Carnegie

More information

a start time signature, an end time signature, a start divisions value, an end divisions value, a start beat, an end beat.

a start time signature, an end time signature, a start divisions value, an end divisions value, a start beat, an end beat. The KIAM System in the C@merata Task at MediaEval 2016 Marina Mytrova Keldysh Institute of Applied Mathematics Russian Academy of Sciences Moscow, Russia mytrova@keldysh.ru ABSTRACT The KIAM system is

More information

HST 725 Music Perception & Cognition Assignment #1 =================================================================

HST 725 Music Perception & Cognition Assignment #1 ================================================================= HST.725 Music Perception and Cognition, Spring 2009 Harvard-MIT Division of Health Sciences and Technology Course Director: Dr. Peter Cariani HST 725 Music Perception & Cognition Assignment #1 =================================================================

More information

2. Problem formulation

2. Problem formulation Artificial Neural Networks in the Automatic License Plate Recognition. Ascencio López José Ignacio, Ramírez Martínez José María Facultad de Ciencias Universidad Autónoma de Baja California Km. 103 Carretera

More information

Course Overview. Assessments What are the essential elements and. aptitude and aural acuity? meaning and expression in music?

Course Overview. Assessments What are the essential elements and. aptitude and aural acuity? meaning and expression in music? BEGINNING PIANO / KEYBOARD CLASS This class is open to all students in grades 9-12 who wish to acquire basic piano skills. It is appropriate for students in band, orchestra, and chorus as well as the non-performing

More information

A Bayesian Network for Real-Time Musical Accompaniment

A Bayesian Network for Real-Time Musical Accompaniment A Bayesian Network for Real-Time Musical Accompaniment Christopher Raphael Department of Mathematics and Statistics, University of Massachusetts at Amherst, Amherst, MA 01003-4515, raphael~math.umass.edu

More information

Lab P-6: Synthesis of Sinusoidal Signals A Music Illusion. A k cos.! k t C k / (1)

Lab P-6: Synthesis of Sinusoidal Signals A Music Illusion. A k cos.! k t C k / (1) DSP First, 2e Signal Processing First Lab P-6: Synthesis of Sinusoidal Signals A Music Illusion Pre-Lab: Read the Pre-Lab and do all the exercises in the Pre-Lab section prior to attending lab. Verification:

More information

Music Radar: A Web-based Query by Humming System

Music Radar: A Web-based Query by Humming System Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,

More information

PHYSICS OF MUSIC. 1.) Charles Taylor, Exploring Music (Music Library ML3805 T )

PHYSICS OF MUSIC. 1.) Charles Taylor, Exploring Music (Music Library ML3805 T ) REFERENCES: 1.) Charles Taylor, Exploring Music (Music Library ML3805 T225 1992) 2.) Juan Roederer, Physics and Psychophysics of Music (Music Library ML3805 R74 1995) 3.) Physics of Sound, writeup in this

More information

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng S. Zhu, P. Ji, W. Kuang and J. Yang Institute of Acoustics, CAS, O.21, Bei-Si-huan-Xi Road, 100190 Beijing,

More information

Predicting Variation of Folk Songs: A Corpus Analysis Study on the Memorability of Melodies Janssen, B.D.; Burgoyne, J.A.; Honing, H.J.

Predicting Variation of Folk Songs: A Corpus Analysis Study on the Memorability of Melodies Janssen, B.D.; Burgoyne, J.A.; Honing, H.J. UvA-DARE (Digital Academic Repository) Predicting Variation of Folk Songs: A Corpus Analysis Study on the Memorability of Melodies Janssen, B.D.; Burgoyne, J.A.; Honing, H.J. Published in: Frontiers in

More information

VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS. O. Javed, S. Khan, Z. Rasheed, M.Shah. {ojaved, khan, zrasheed,

VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS. O. Javed, S. Khan, Z. Rasheed, M.Shah. {ojaved, khan, zrasheed, VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS O. Javed, S. Khan, Z. Rasheed, M.Shah {ojaved, khan, zrasheed, shah}@cs.ucf.edu Computer Vision Lab School of Electrical Engineering and Computer

More information

CHAPTER ONE TWO-PART COUNTERPOINT IN FIRST SPECIES (1:1)

CHAPTER ONE TWO-PART COUNTERPOINT IN FIRST SPECIES (1:1) HANDBOOK OF TONAL COUNTERPOINT G. HEUSSENSTAMM Page 1 CHAPTER ONE TWO-PART COUNTERPOINT IN FIRST SPECIES (1:1) What is counterpoint? Counterpoint is the art of combining melodies; each part has its own

More information

Modeling the Effect of Meter in Rhythmic Categorization: Preliminary Results

Modeling the Effect of Meter in Rhythmic Categorization: Preliminary Results Modeling the Effect of Meter in Rhythmic Categorization: Preliminary Results Peter Desain and Henkjan Honing,2 Music, Mind, Machine Group NICI, University of Nijmegen P.O. Box 904, 6500 HE Nijmegen The

More information

Broken Wires Diagnosis Method Numerical Simulation Based on Smart Cable Structure

Broken Wires Diagnosis Method Numerical Simulation Based on Smart Cable Structure PHOTONIC SENSORS / Vol. 4, No. 4, 2014: 366 372 Broken Wires Diagnosis Method Numerical Simulation Based on Smart Cable Structure Sheng LI 1*, Min ZHOU 2, and Yan YANG 3 1 National Engineering Laboratory

More information

The Sparsity of Simple Recurrent Networks in Musical Structure Learning

The Sparsity of Simple Recurrent Networks in Musical Structure Learning The Sparsity of Simple Recurrent Networks in Musical Structure Learning Kat R. Agres (kra9@cornell.edu) Department of Psychology, Cornell University, 211 Uris Hall Ithaca, NY 14853 USA Jordan E. DeLong

More information

Polyphonic music transcription through dynamic networks and spectral pattern identification

Polyphonic music transcription through dynamic networks and spectral pattern identification Polyphonic music transcription through dynamic networks and spectral pattern identification Antonio Pertusa and José M. Iñesta Departamento de Lenguajes y Sistemas Informáticos Universidad de Alicante,

More information

Computing, Artificial Intelligence, and Music. A History and Exploration of Current Research. Josh Everist CS 427 5/12/05

Computing, Artificial Intelligence, and Music. A History and Exploration of Current Research. Josh Everist CS 427 5/12/05 Computing, Artificial Intelligence, and Music A History and Exploration of Current Research Josh Everist CS 427 5/12/05 Introduction. As an art, music is older than mathematics. Humans learned to manipulate

More information

Week 14 Music Understanding and Classification

Week 14 Music Understanding and Classification Week 14 Music Understanding and Classification Roger B. Dannenberg Professor of Computer Science, Music & Art Overview n Music Style Classification n What s a classifier? n Naïve Bayesian Classifiers n

More information

Calculating Dissonance in Chopin s Étude Op. 10 No. 1

Calculating Dissonance in Chopin s Étude Op. 10 No. 1 Calculating Dissonance in Chopin s Étude Op. 10 No. 1 Nikita Mamedov and Robert Peck Department of Music nmamed1@lsu.edu Abstract. The twenty-seven études of Frédéric Chopin are exemplary works that display

More information

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE Copyright SFA - InterNoise 2000 1 inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering 27-30 August 2000, Nice, FRANCE I-INCE Classification: 7.9 THE FUTURE OF SOUND

More information

Study Guide. Solutions to Selected Exercises. Foundations of Music and Musicianship with CD-ROM. 2nd Edition. David Damschroder

Study Guide. Solutions to Selected Exercises. Foundations of Music and Musicianship with CD-ROM. 2nd Edition. David Damschroder Study Guide Solutions to Selected Exercises Foundations of Music and Musicianship with CD-ROM 2nd Edition by David Damschroder Solutions to Selected Exercises 1 CHAPTER 1 P1-4 Do exercises a-c. Remember

More information

A Framework for Segmentation of Interview Videos

A Framework for Segmentation of Interview Videos A Framework for Segmentation of Interview Videos Omar Javed, Sohaib Khan, Zeeshan Rasheed, Mubarak Shah Computer Vision Lab School of Electrical Engineering and Computer Science University of Central Florida

More information

The Tone Height of Multiharmonic Sounds. Introduction

The Tone Height of Multiharmonic Sounds. Introduction Music-Perception Winter 1990, Vol. 8, No. 2, 203-214 I990 BY THE REGENTS OF THE UNIVERSITY OF CALIFORNIA The Tone Height of Multiharmonic Sounds ROY D. PATTERSON MRC Applied Psychology Unit, Cambridge,

More information

Some researchers in the computational sciences have considered music computation, including music reproduction

Some researchers in the computational sciences have considered music computation, including music reproduction INFORMS Journal on Computing Vol. 18, No. 3, Summer 2006, pp. 321 338 issn 1091-9856 eissn 1526-5528 06 1803 0321 informs doi 10.1287/ioc.1050.0131 2006 INFORMS Recurrent Neural Networks for Music Computation

More information

Ashton Allan MU 228 Tonality within Aaron Copland s Piano Variations

Ashton Allan MU 228 Tonality within Aaron Copland s Piano Variations Ashton Allan MU 228 Tonality within Aaron Copland s Piano Variations The closest Aaron Copland ever got to atonal music was his 1930 composition, Piano Variations. This work, constructed from twenty independently

More information

THE EFFECT OF EXPERTISE IN EVALUATING EMOTIONS IN MUSIC

THE EFFECT OF EXPERTISE IN EVALUATING EMOTIONS IN MUSIC THE EFFECT OF EXPERTISE IN EVALUATING EMOTIONS IN MUSIC Fabio Morreale, Raul Masu, Antonella De Angeli, Patrizio Fava Department of Information Engineering and Computer Science, University Of Trento, Italy

More information

Perceptual Evaluation of Automatically Extracted Musical Motives

Perceptual Evaluation of Automatically Extracted Musical Motives Perceptual Evaluation of Automatically Extracted Musical Motives Oriol Nieto 1, Morwaread M. Farbood 2 Dept. of Music and Performing Arts Professions, New York University, USA 1 oriol@nyu.edu, 2 mfarbood@nyu.edu

More information

Automatic meter extraction from MIDI files (Extraction automatique de mètres à partir de fichiers MIDI)

Automatic meter extraction from MIDI files (Extraction automatique de mètres à partir de fichiers MIDI) Journées d'informatique Musicale, 9 e édition, Marseille, 9-1 mai 00 Automatic meter extraction from MIDI files (Extraction automatique de mètres à partir de fichiers MIDI) Benoit Meudic Ircam - Centre

More information

Florida Performing Fine Arts Assessment Item Specifications for Benchmarks in Course: M/J Chorus 3

Florida Performing Fine Arts Assessment Item Specifications for Benchmarks in Course: M/J Chorus 3 Task A/B/C/D Item Type Florida Performing Fine Arts Assessment Course Title: M/J Chorus 3 Course Number: 1303020 Abbreviated Title: M/J CHORUS 3 Course Length: Year Course Level: 2 PERFORMING Benchmarks

More information

Doctor of Philosophy

Doctor of Philosophy University of Adelaide Elder Conservatorium of Music Faculty of Humanities and Social Sciences Declarative Computer Music Programming: using Prolog to generate rule-based musical counterpoints by Robert

More information

From Score to Performance: A Tutorial to Rubato Software Part I: Metro- and MeloRubette Part II: PerformanceRubette

From Score to Performance: A Tutorial to Rubato Software Part I: Metro- and MeloRubette Part II: PerformanceRubette From Score to Performance: A Tutorial to Rubato Software Part I: Metro- and MeloRubette Part II: PerformanceRubette May 6, 2016 Authors: Part I: Bill Heinze, Alison Lee, Lydia Michel, Sam Wong Part II:

More information

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm Georgia State University ScholarWorks @ Georgia State University Music Faculty Publications School of Music 2013 Chords not required: Incorporating horizontal and vertical aspects independently in a computer

More information