A hierarchical self-organizing map model for sequence recognition
|
|
- Emerald Shaw
- 6 years ago
- Views:
Transcription
1 A hierarchical self-organizing map model for sequence recognition Otávio Augusto S. Carpinteiro Instituto de Engenharia Elétrica Escola Federal de Engenharia de Itajubá Av. BPS 1303, Itajubá, MG, , Brazil h-page: phone: fax: March 1999 Keywords: artificial intelligence, artificial neural networks, pattern recognition, self-organizing map.
2 Abstract The paper presents the analysis of an original hierarchical neural model on a complex sequence the complete sixteenth four-part fugue in G minor of the Well-Tempered Clavier (vol. I) of J. S. Bach. The model makes an effective use of context information, through its hierarchical topology and its embedded time integrators, and that enables it to keep a very good account of past events. The model performs sequence classification and discrimination efficiently. It has application in domains which require pattern recognition, or particularly, which demand recognizing either a set of sequences of vectors in time or sub-sequences into a unique and large sequence of vectors in time.
3 1. Introduction Many researchers have developed artificial neural models to classify sequences in time. Several models based on Kohonen s self-organizing feature map [1], nonetheless, have faced some wellknown flaws. In windowed data models, as in Kangas model [2], the input vectors are concatenated in a fixed-size window which slides in time. The size of the memory to the past inputs is limited to the size of the window. The most serious deficiency of such models is that they become computationally expensive as wider windows are required. In time integral models, as in Chappell and Taylor s model [3], the activation of a unit is a combination of its current input and its former outputs decayed in time. Such models suffer from loss of context. Sequences that slightly differ in their initial elements (e.g., abccc, baccc, and bbccc) would probably have the same classification. In non-orthodox models, neither windows nor time integrators are employed. In James and Miikkulainen s model [4], for instance, when a vector in the sequence is input, the output unit in the map which wins the competition for that vector is disabled for further competition, and its activation decays in time to indicate which vector in the sequence it is representing. Each winning output unit represents just a single vector in the input sequence, and so, the representation for the whole input sequence is given by a sequence of winning output units in the map. This distributed representation is more complicated because it has to take into consideration not only the winning units but also their activations. For example, input sequences with the same elements but in different order (e.g., abcd and dcba) will have the same winning units, and so, one has to look at their activations to verify which input sequence the map is representing. Another disadvantage is that the model is unable to recognize sub-sequences inside a large and unique input sequence. The reason is that, after winning the competition in the map, the winning units are disabled for further competitions, and consequently, identical or similar sub-sequences inserted into different points of the large sequence will never hold identical or similar representations in the map. The model presented here is also based on Kohonen s map. However, thanks to its hierarchical topology and its embedded time integrators, it makes an efficient use of context information which prevents it from suffering from the flaws mentioned above. Artificial neural models have been widely employed in the musical domain. Among them, we may cite models for pitch perception [5, 6], chord perception [7], tonal perception [8, 9, 10, 11, 12], musical time perception [13], perception of musical sequences [14], musical pattern categorizations [15, 16], and musical composition [17, 18, 19, 20, 21]. Gjerdingen s work [15, 16] is the closest to the area of musical sequence recognition. His neural model was required to classify groups of notes produced by segmentation into categories. Unfortunately though, nothing was mentioned about existence or not of similar sequences of patterns in the training set, so that one could not assess performance of the net in terms of classifications and misclassifications of such sequences.
4 This paper reports our final research in neural models applied to the musical domain. The initial results were published in [22]. The research reported here, however, is much more relevant, for the musical sequences employed in the studies were much more complex than those employed in [22], and the results obtained and conclusions are also much more significant. In [22], we applied the model to a single part (or line) of a fugue of J. S. Bach. Here, the model is trained and evaluated on a more complex sequence in time a complete four-part fugue of J. S. Bach. In the following sections, we shall describe the representation, model, and the experiment on the musical sequences. 2. Representation for the sequences The concept behind the representation is the division of a musical piece into equal size time intervals. The small figure in a musical sequence is set to be the time interval (TI). Thus, all other figures in the sequence are multiples of TI. For example, if TI is an eighth note, a quarter note lasts two TIs, a half note lasts four TIs, and so on. A time interval counter (TIC) may also be defined. One TIC lasts one TI. TIC is the unit in which the musical sequence is measured. Therefore, at each TIC, either there is a rest, or a note onset, or a note sustained. The input data in the experiment consists of a sequence of musical intervals, which corresponds to the Bach s fugue. Data is input one TIC at a time. Fifteen neural units are used in the representation. Each unit represents one musical interval ranging from an octave down to an octave up. We assume here, therefore, that listeners are able to separate out voices if intervals between notes in the voices are greater than a fifteenth. When there is a rest, none of the input units receives activation. Otherwise, when a note is onset or sustained, the unit corresponding to the interval receives activation. The representation for musical sequences takes into consideration all musical voices, and therefore, is complex because the voices interact. The representation assumes three facts. First, any note onset occurring in a TIC makes up an interval with all notes onset or sustained which occurred in the TIC immediately before. Second, the representation does represent the intervals which occur in a TIC, but does not represent multiple instances occurring through the voices in the TIC. Third, at any given TIC, an interval corresponding to a note onset masks any occurrence of the same interval corresponding to a note sustained. Let us consider the multivoiced musical sequence in figure 1, and a time integrator with decay rate of 0.5 applied to input units. The sequence lasts 8 TICs, and TI is a sixteenth note. Units corresponding to intervals made up by a note onset or note sustained receive an activation of 1.0 and 0.5 respectively. Table 1 displays then the integrated value of activations of the units when inputting, TIC by TIC, the sequence in figure 1.
5 4 G 8 < 4 I : !)! -!)!)!)! -!(! -?? Figure 1: A multivoiced musical sequence TIC Table 1: Representation for the musical sequence in figure 1 Activations of Units Representing Intervals -8-7 ::: ::: ::: ::: 2 ::: 1.0 ::: ::: ::: ::: ::: ::: ::: ::: ::: ::: ::: ::: ::: The model The model is made up of two self-organizing maps (SOMs), as shown in figure 2. Its features, performance, and potential are better evaluated in [22, 23]. The problem of loss of context which occurs in other models, the analysis of our model, the analysis of Kohonen s SOM model, and the comparison of our model with Kohonen s in a simple temporal pattern recognition problem are reported in [23]. In this paper, we made use of contrived sequences to explain the behaviour of the two models in the presence of context information. We explained the effect of the time integrators on the results produced as well. In our model, the bottom SOM is responsible for making up representations which take into consideration not only the information given by the input vectors, but also the information given by the context in which these vectors are inserted. Therefore, the information passed up to the top SOM is much more acurate, and consequently, so are the classifications. We finally show that our model has much better performance than that of Kohonen s.
6 Top SOM Map Time Integrator Λ Bottom SOM Map Time Integrator V(t ) Figure 2: The model The input to the model is a sequence in time of m-dimensional vectors, S 1 = V(1), V(2),..., V(t),..., V(z), where the components of each vector are non-negative real values. The sequence is presented to the input layer of the bottom SOM, one vector at a time. The input layer has m units, one for each component of the input vector V(t), and a time integrator. The activation X(t) of the units in the input layer is given by X( t) = V( t) + δ 1X( t 1) (1) where δ 1 (0,1) is the decay rate. For each input vector X(t), the winning unit i * (t) in the map 1 is the unit which has the smallest distance Ψ(i,t). For each output unit i, Ψ(i,t) is given by the Euclidean distance between the input vector X(t) and the unit s weight vector W i. Each output unit i in the neighbourhood N * (t) of the winning unit i * (t) has its weight W i updated by W ( t + 1) = W ( t) + α ϒ ( i)[ X( t) Wi( t)] (2) i i where α (0,1) is the learning rate. ϒ(i) is the neighbourhood interaction function [24], a Gaussian type function, and is given by ϒ () i = κ + κ e 1 2 * 2 κ 3[ Φ( ii, ( t))] 2 σ 2 (3) where κ 1, κ 2, and κ 3 are constants, σ is the radius of the neighbourhood N * (t), and Φ(i,i * (t)) is the distance in the map between the unit i and the winning unit i * (t). The distance Φ(i,i ) between any two units i and i in the map is calculated according to the maximum norm, { } ' '' ' '' ' '' Φ(i, i ) = max l l, c c (4) where (l,c ) and (l,c ) are the coordinates of the units i and i respectively in the map.
7 The input to the top SOM is determined by the distances Φ(i,i * (t)) of the n units in the map of the bottom SOM. The input is thus a sequence in time of n-dimensional vectors, S 2 = Λ(Φ(i,i * (1))), Λ(Φ(i,i * (2))),..., Λ(Φ(i,i * (t))),..., Λ(Φ(i,i * (z))), where Λ is a n-dimensional transfer function on a n-dimensional space domain. Λ is defined as * * 1 κφ( ii, ( t)) if i N ( t) * ΛΦ ( ( ii, ( t))) = 0 otherwise (5) where κ is a constant, and N * (t) is a neighbourhood of the winning unit. The sequence S 2 is then presented to the input layer of the top SOM, one vector at a time. The input layer has n units, one for each component of the input vector Λ(Φ(i,i * (t))), and a time integrator. The activation X(t) of the units in the input layer is thus given by where δ 2 (0,1) is the decay rate. * X() t = ΛΦ ( ( i, i ())) t + δ2x( t 1) (6) The dynamics of the top SOM is identical to that of the bottom SOM. 4. The experiment We wanted the model to be tested on musical sequences, because the musical domain set three strong conditions on the model. First, that the model were able to recognize both a set of input sequences and a set of sub-sequences within a large and unique input sequence. The model was required to recognize a set of input sequences when the whole sequence was segmented. Otherwise, the entire piece consisted of a unique input sequence, and the model was thus required to recognize sub-sequences of that sequence. Second, that the model classified sequences (or sub-sequences) properly in the presence of noise. The reason followed from the fact that any two sequences which differed slightly had to achieve similar classifications. Third, that the model recognized sequences (or sub-sequences) in a very precise form. The reason for the latter is that any two sequences which shared either some intervals, or even all intervals, but in an alternative order or rhythm, were musically different, and as a consequence, had to be recognized as distinct. The experiment was on recognition of the instances of a theme occurring in a complex sequence the complete sixteenth four-part fugue in G minor of the first volume of The Well-Tempered Clavier of Bach. The fugue had 544 TICs, and TI was a sixteenth note. The theme of the fugue is shown in figure 3.
8 I 22 S?!! [ [! 4!! theme?!!!! PNN! -!... Figure 3: Theme of the sixteenth fugue in G minor The fugue in G minor was chosen for several reasons. First, as many of fugues of Bach, it has four voices. Second, it possesses many perfect and modified instances of theme. Third, it includes two cases of stretto 2. One case occurs between its seventeenth and eighteenth bars, in which two instances of theme overlap, and the other case occurs between its twenty eighth and thirtieth bars, in which three instances of theme overlap. Fourth, the thematic material of the theme is extensively developed throughout the fugue. Such developments 3, although quite similar to the theme, are not instances of theme. Fifth, very common intervals, as seconds up and down, occur extensively in the theme as well as in many passages in the fugue. All facts above are usually present in real situations, for example, in those in which humans are asked to perform thematic recognition in musical domains. Apart from providing a typical situation in a real domain, such facts also increase very much the level of difficulty of the domain to which the artificial neural model is applied. The input data consisted of two sets, hereafter referred to as input set I and input set II. Input set I consisted of a large and unique sequence of musical intervals, which corresponded to the fugue. Input set II contained many sequences, which were formed by segmenting the fugue whenever there were rests. The experiment pursued two aims. Firstly, to determine whether the model recognizes all instances of the theme in the fugue. Secondly, to determine whether any other sequence (or sub-sequence), which was not an instance, was not misclassified as theme. The training of the two SOMs of the model took place in two phases coarse-mapping and fine-tuning. The initial learning rate was set to 0.5, and the initial size of the neighbourhood was set to the size of the map. In the coarse-mapping phase, the learning rate and the radius of the neighbourhood were reduced linearly whereas in the fine-tuning phase, they were kept constant in 0.01 and 1 respectively. The coarse-mapping phase took 20%, and the fine-tuning phase took 80% of the total number of epochs. The initial weights were given randomly, in the range between 0 and 0.1, to both SOMs. Different values for decay rate were tested. In the bottom SOM of model II, it varied from 0.1 to 0.7, and in the top SOM, from 0.5 to 0.9. We present here, however, only the results using decay rates of 0.3 and 0.7 for the bottom and top SOM respectively. Such results, reached with those decay rates, were the best for all studies performed. The bottom SOM of model II was tested with map size of 15 15, and was trained in 700 epochs. The top SOM was tested with map size of 18 18, and trained in 850 epochs. Two
9 transfer functions Λ were tested. The first was given by equation 5, with neighbourhood N * (t) = {i Φ(i,i * (t)) < 2}, and κ = 0.5. The second was also given by equation 5, but with neighbourhood N * (t) = {i Φ(i,i * (t)) < 4}, and κ = We report here studies using the second transfer function only, for it produced the better results. The input layer of the bottom SOM of model II held fifteen units, one for each musical interval ranging from an octave down to an octave up. The representation employed in these units is fully described in section 2. The experiment comprised five studies. In the last four, in order to study the effect of noise on the classifications, reinforcement in activation was given to input units when representing instances of theme. For example, in the second experiment, note onset and note sustained received activations of 0.1 and 0.07 respectively. When corresponding to instances of the theme, they received instead, activations of 0.5 and 0.35 respectively. Table 2 shows the activation values of notes onset and sustained, whether reinforced or not, as well as the input set employed in each study 4. Table 2: Parameter values of the studies Study Input Reinforcement Note Note N. Onset N. Sustained Set Value Onset Sustained (Reinforced) (Reinforced) I I II I III I IV I V II A sequence (or sub-sequence) S a is said to have the same classification as that of the * * theme S t if the distance Φ(ia(), z it ()) z < 2, where i * a () z and i * t ( z) are the last winning units of Sa and S t. In case of S a be also an instance of the theme, the mean error of the instance S a, Ê(S a ), is then given by '' ta * * ( ia( z), it ( z)) Ê( S a ) = Φ '' ' ' ( t t + 1) z= ta a a (7) ' '' where t a and t a are the initial and final TICs of the instance Sa 5. The total mean error is given by the sum of the mean errors of all instances. We have selected some instances of theme to present here. Those instances hold some peculiarities, whether in terms of similarity with the theme, or convergence to the theme classification, or context in which they were inserted, which make them representatives among the whole set of instances.
10 Figures 4 to 10 plot, for each selected instance of theme, the distances between each winning unit of the instance and its corresponding winning unit in the theme for each study carried out on the model. We used equation 4 to measure those distances in the map of the top SOM. Table 3 displays the main characteristics of the instances of theme. Table 4 shows, for each study, the total error, and the mean errors of each instance of theme. Table 5 displays the classifications and misclassifications of the studies. Figure 11 plots the total mean error of classifications in accordance with reinforcement provided in the first four studies study I study II study III study IV study V Distance Figure 4: Classifications of the first instance of theme (TICs 33 50) relative to theme. The instance occurs in the fourth voice the highest one concurrently with another voice. The instance differs from the theme in its first two TICs. The classifications of the instance yielded by the model only converge to that of the theme in the fourth and fifth studies. TIC
11 study I study II study III study IV study V Distance Figure 5: Classifications of the seventh instance of theme (TICs ) relative to theme. The instance occurs in the first voice, concurrently with three other voices. The instance is a perfect copy of the theme. The classifications of the instance yielded by the model do not converge to that of the theme in any of the studies. TIC
12 study I study II study III study IV study V Distance Figure 6: Classifications of the eighth instance of theme (TICs ) relative to theme. The instance occurs in the third voice, and is a perfect copy of the theme. At its beginning, the eighth instance occurs concurrently with three other voices. From TIC 283 onwards, it occurs concurrently with just one voice. The classifications of the instance produced by the model only converge to that of the theme in the second, third, fourth, and fifth studies. TIC
13 study I study II study III study IV study V Distance Figure 7: Classifications of the eleventh instance of theme (TICs ) relative to theme. The instance occurs in the third voice, concurrently with two other voices. The instance differs from the theme in its first two TICs, and in two TICs in its middle. The classifications of the instance yielded by the model only converge to that of the theme in the fifth study. TIC
14 study I study II study III study IV study V Distance Figure 8: Classifications of the twelfth instance of theme (TICs ) relative to theme. The instance occurs in the fourth voice, and is a perfect copy of the theme. At its beginning, the instance occurs unaccompanied. From TIC 443 onwards, it occurs concurrently with two other voices, and from TIC 451 onwards, it occurs concurrently with three other voices. The classifications of the instance produced by the model do not converge to that of the theme in any of the studies. TIC
15 study I study II study III study IV study V Distance Figure 9: Classifications of the thirteenth instance of theme (TICs ) relative to theme. The instance occurs in the second voice, and is a perfect copy of the theme. At its beginning, the instance occurs concurrently with two other voices. From TIC 451 onwards, it occurs concurrently with three other voices, and from TIC 459 onwards, it occurs concurrently with just one voice. The classifications of the instance yielded by the model only converge to that of the theme in the second, third, fourth, and fifth studies. TIC
16 study I study II study III study IV study V Distance Figure 10: Classifications of the fourteenth instance of theme (TICs ) relative to theme. The instance occurs in the first voice, concurrently with three other voices. The instance is a perfect copy of the theme. The classifications of the instance produced by the model do not converge to that of the theme in any of the studies. TIC
17 Table 3: Main characteristics of the instances of theme Voice: voice in which the instance occurs (1: lowest; 4: highest); No. Voices: number of voices sounding concurrently (including the one in which the instance occurs); Dif. TICs: difference in number of TICs between the instance and the theme; Conv.: studies in which the instance converges to the classification of the theme; TIC: TICs in which the instance occurs; Voice No. V oices Dif.TICs Conv. TIC 1 3 4/5 73{ /5 209{ /5 313{ { { /4/2 2/3/4/5 449{ /5 97{ /5 521{ /4/5 185{ /2 2/3/4/5 273{ { /4/5 497{ /5 33{ /5 337{ /5 233{ /3/4 441{456
18 Table 4: Mean errors of the instances of theme TIC oftheerrors Instances Mean Study: I Study: II Study: III Study: IV Study: V 33{ { { { { { { { { { { { { { { { Total Table 5: Classifications and misclassifications of the model Study No. Hits Classications. No F ailures No. Miscl. Minor Misclassications No. Miscl. jor Ma I II III IV V
19 Total Mean Error Reinforcement Figure 11: Total mean error of classifications
20 Some conclusions may be drawn from the results. First, as displayed in table 5, the model held a high number of misclassifications in the third study. Such a high number was due to the fact that the model classified an intermediate part of theme as its final part, and consequently, kept on misclassifying intermediate parts of instances of theme as their final parts as well. Second, as shown in tables 4 and 5, the model exhibited a similar good performance in the fourth and fifth studies. In the fourth, the model exhibited lower mean errors, whereas in the fifth study, it presented better results for classifications and misclassifications. The model satisfied, therefore, the first of our required conditions as stated in section 4, i.e., it was able to perform efficiently both on a set of input sequences and on a set of sub-sequences within a large and unique input sequence. Third, by analysing the results displayed in the figures 4 to 11, and in the table 3, one may observe that the model was fault tolerant to errors, satisfying therefore, the second of our required conditions as stated in section 4. It classified properly several instances which differed slightly from the theme, whether in the pitch or in the duration of one single note. The model performed classification efficiently in the presence of noise as well. When instances of theme occurred concurrently with other polyphonic voices, the degree of noise was so high that it caused the model not to classify instances correctly. We may observe that in the results of studies I, II, and III. However, when thematic reinforcement was given to instances in a greater amount, the remaining polyphonic voices started playing roles of noisy backgrounds, and then, the model started classifying rightly instances of theme. That may be observed in the results of studies IV and V. Fourth, as it may be observed in figures 5, 8, 10, and in the table 3, the model failed, in all studies, in recognizing three instances of theme. It succeeded, however, as shown in figures 6 and 9, and in the table 3 in recognizing two other instances of theme in the last four studies. These instances, which occur between TICs 265 and 282, TICs 273 and 290, TICs 441 and 456, TICs 449 and 466, and TICs 457 and 460, overlapped through the voices, making up the two cases of stretto present in the fugue. One may conclude, therefore, that the recognition of strettos was not performed reasonably by the model. Fifth, the model presented a fairly high number of minor misclassifications in the fourth and fifth studies, according to the results in the table 5. Yet, we should not consider severely those results, because they are much more a case of `inertia' than a proper error. We considered a case of minor misclassification when the model kept on classifying as instance of theme the next few TICs which followed the instance. In respect to the major errors yet, the model performed perfectly, not producing anyone. The results thus reveal that the third of our required conditions, as stated in the section 4, was finely satisfied. Sixth, by comparing studies IV and V in tables 4 and 5, one may verify that there is not a significant difference between their results. The model exhibited lower mean errors in the study IV, but otherwise, it presented better results in terms of classifications and misclassifications in the study V. The straightforward conclusion which may be drawn is thus,
21 that the segmentation of the fugue on rests, i.e., the resetting of the model on inputs which corresponds to rests, does not yield any improvement in terms of classification. Seventh, by analysing the relation between number of voices sounding and the mean errors in the tables 3 and 4, one may observe that the recognition of an instance of theme becomes more difficult when the number of voices sounding simultaneously increases. The only exception is in the instance of theme occurring along TICs 449 to 466. The mean errors in the fourth and fifth studies are 6.56 and 8.28 respectively. They are much higher, for example, than those in the same studies occurring in the instances along TICs 97 to 114, and 521 to 538. Nevertheless, one must see that the instance of theme along TICs 449 to 466 is the one which, together with the instances along TICs 441 to 456, and 457 to 460, makes up the second stretto present in the fugue. Hence, the entrance of the first instance from TICs 441 to 456, and of the third instance from TICs 457 to 460, probably contributed to those high mean errors achieved. Finally, there seems not to be any relation between the voice in which the instance of theme occurs and the ease in recognizing it. Let us present three examples taken from the tables 3 and 4. The first is given by the instances I a and I b from TICs 185 to 201, and from 33 to 50 respectively. I a occurs in the third voice, and I b in the fourth one. Both instances occur together with another voice. The mean errors of I a are lower than those of I b. The second example is produced by the instances I c and I d, from TICs 313 to 330, and 337 to 353 respectively. I c occurs in the first voice, I d in the fourth, and there are two other voices sounding concurrently with those instances. The mean errors of I c are higher than those of I d, with the exception of that one in the study II. The third example is given by the instances I e from TICs 521 to 538, and I f from TICs 233 to 250. I e occurs in the second voice, and I f in the fourth one. There are three other voices sounding concurrently with the instances. The mean errors of I e are lower than those of I f in the studies IV and V. Thus, the recognition of instances of theme was not facilitated when the instances occurred in any specific voice. 5. Conclusion An original representation for musical sequences, and an original artificial neural model for sequence classification is presented. The model has a topology made up of two self-organizing map networks, one on top of the other. It encodes and manipulates context information effectively, and that enables it to perform sequence classification and discrimination efficiently. The model has application in domains which demand classifying either a set of sequences of vectors or sub-sequences into a unique and large sequence of vectors in time. The results obtained have shown that the artificial neural model was able to perform efficiently sequence classification and discrimination. The model was able to classify properly most of the instances of theme occurring in the musical piece. It is worth noticing that the model performed classification even in the presence of noise, i.e., even when instances occurred modified, in different past contexts, and amidst different polyphonic voices in the musical piece.
22 The model could also discriminate instances of theme from sequences that shared some similarity with the theme. Very many of these pseudo-instance sequences occurred in the fugue. Rightly yet, the model did not classify them as instances of theme. The parameters of the model and the model itself are not, in any way, tied up to any specific application, whether it be a musical piece or any other domain. Indeed, we are currently doing further research on the model, applying it to the electrical engineering domain, to predict the demand for electrical load. The first results of this work are to be published in [25]. Footnotes 1 Also known as array, grid, or output layer. 2 Stretto is a musical passage where two or more instances of theme overlap. 3 In the fugal domain, the developments of thematic material take place in parts called episodes. 4 Reinforcement was provided from the seventh common TIC between the theme and any of its instances. 5 The initial TIC is the seventh TIC of the instance. References [1] T. Kohonen. Self-Organization and Associative Memory. Springer-Verlag, Berlin, third edition, [2] J. Kangas. On the Analysis of Pattern Sequences by Self-Organizing Maps. PhD thesis, Laboratory of Computer and Information Science, Helsinki University of Technology, Rakentajanaukio 2 C, SF-02150, Finland, [3] G. J. Chappell and J. G. Taylor. The temporal Kohonen map. Neural Networks, 6: , [4] D. L. James and R. Miikkulainen. SARDNET: a self-organizing feature map for sequences. In G. Tesauro, D. S. Touretzky, and T. K. Leen, editors, Proceedings of the Advances in Neural Information Processing Systems, volume 7. Morgan Kaufmann, [5] H. Sano and B. Jenkins. A neural network model for pitch perception. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [6] I. Taylor and M. Greenhough. Modelling pitch perception with adaptive resonance theory artificial neural networks. Connection Science, 6(2&3): , 1994.
23 [7] B. Laden and D. Keefe. The representation of pitch in a neural net model of chord classification. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [8] D. Scarborough, B. Miller, and J. Jones. Connectionist models for tonal analysis. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [9] M. Leman. The ontogenesis of tonal semantics: results of a computer study. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [10] J. Bharucha. Music cognition and perceptual facilitation: a connectionist framework. Music Perception, 5(1):1 30, [11] J. Bharucha. Pitch, harmony, and neural nets: a psychological perspective. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [12] J. Bharucha and P. Todd. Modeling the perception of tonal structure with neural nets. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [13] P. Desain and H. Honing. The quantization of musical time: a connectionist approach. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [14] M. Page. Modelling the perception of musical sequences with self-organizing neural networks. Connection Science, 6(2&3): , [15] R. Gjerdingen. Categorization of musical patterns by self-organizing neuronlike networks. Music Perception, 7(4): , [16] R. Gjerdingen. Using connectionist models to explore complex musical patterns. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [17] P. Todd. A connectionist approach to algorithmic composition. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [18] J. Lewis. Algorithms for music composition by neural nets: improved CBR paradigms. Proceedings of the International Computer Music Conference, Computer Music Association, pages , [19] J. Lewis. Creation by refinement and the problem of algorithmic music composition. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, 1991.
24 [20] M. Mozer. Connectionist music composition based on melodic, stylistic, and psychophysical constraints. In P. Todd and D. Loy, editors, Music and Connectionism, pages The MIT Press, Cambridge, MA, [21] M. Mozer and T. Soukup. Connectionist music composition based on melodic and stylistic constraints. In R. Lippmann, J. Moody, and D. Touretzky, editors, Proceedings of the Advances in Neural Information Processing Systems, Morgan Kaufmann, volume 3, pages , [22] O. A. S. Carpinteiro. A hierarchical self-organizing map model for sequence recognition. Neural Processing Letters, 9:1 12, [23] O. A. S. Carpinteiro. A hierarchical self-organizing map model for pattern recognition. In L. Caloba and J. Barreto, editors, Proceedings of the Brazilian Congress on Artificial Neural Networks 97 (CBRN 97), pages , UFSC, Florianópolis, SC, Brazil, [24] Z. Lo and B. Bavarian. Improved rate of convergence in Kohonen neural network. In Proceedings of the International Joint Conference on Neural Networks, volume 2, pages , July [25] O. A. S. Carpinteiro. A hierarchical self-organizing map model in short-term load forecasting. To appear in the Proceedings of the Fifth International Conference on Engineering Applications of Neural Networks (EANN), Warsaw, Poland, September
Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies
Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies Judy Franklin Computer Science Department Smith College Northampton, MA 01063 Abstract Recurrent (neural) networks have
More informationMELONET I: Neural Nets for Inventing Baroque-Style Chorale Variations
MELONET I: Neural Nets for Inventing Baroque-Style Chorale Variations Dominik Hornel dominik@ira.uka.de Institut fur Logik, Komplexitat und Deduktionssysteme Universitat Fridericiana Karlsruhe (TH) Am
More informationRecurrent Neural Networks and Pitch Representations for Music Tasks
Recurrent Neural Networks and Pitch Representations for Music Tasks Judy A. Franklin Smith College Department of Computer Science Northampton, MA 01063 jfranklin@cs.smith.edu Abstract We present results
More informationNetNeg: A Connectionist-Agent Integrated System for Representing Musical Knowledge
From: AAAI Technical Report SS-99-05. Compilation copyright 1999, AAAI (www.aaai.org). All rights reserved. NetNeg: A Connectionist-Agent Integrated System for Representing Musical Knowledge Dan Gang and
More informationNotes on David Temperley s What s Key for Key? The Krumhansl-Schmuckler Key-Finding Algorithm Reconsidered By Carley Tanoue
Notes on David Temperley s What s Key for Key? The Krumhansl-Schmuckler Key-Finding Algorithm Reconsidered By Carley Tanoue I. Intro A. Key is an essential aspect of Western music. 1. Key provides the
More informationBach in a Box - Real-Time Harmony
Bach in a Box - Real-Time Harmony Randall R. Spangler and Rodney M. Goodman* Computation and Neural Systems California Institute of Technology, 136-93 Pasadena, CA 91125 Jim Hawkinst 88B Milton Grove Stoke
More informationAlgorithmic Music Composition
Algorithmic Music Composition MUS-15 Jan Dreier July 6, 2015 1 Introduction The goal of algorithmic music composition is to automate the process of creating music. One wants to create pleasant music without
More informationRobert Alexandru Dobre, Cristian Negrescu
ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Automatic Music Transcription Software Based on Constant Q
More information2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t
MPEG-7 FOR CONTENT-BASED MUSIC PROCESSING Λ Emilia GÓMEZ, Fabien GOUYON, Perfecto HERRERA and Xavier AMATRIAIN Music Technology Group, Universitat Pompeu Fabra, Barcelona, SPAIN http://www.iua.upf.es/mtg
More informationDistortion Analysis Of Tamil Language Characters Recognition
www.ijcsi.org 390 Distortion Analysis Of Tamil Language Characters Recognition Gowri.N 1, R. Bhaskaran 2, 1. T.B.A.K. College for Women, Kilakarai, 2. School Of Mathematics, Madurai Kamaraj University,
More informationComputer Coordination With Popular Music: A New Research Agenda 1
Computer Coordination With Popular Music: A New Research Agenda 1 Roger B. Dannenberg roger.dannenberg@cs.cmu.edu http://www.cs.cmu.edu/~rbd School of Computer Science Carnegie Mellon University Pittsburgh,
More informationA probabilistic approach to determining bass voice leading in melodic harmonisation
A probabilistic approach to determining bass voice leading in melodic harmonisation Dimos Makris a, Maximos Kaliakatsos-Papakostas b, and Emilios Cambouropoulos b a Department of Informatics, Ionian University,
More informationHidden Markov Model based dance recognition
Hidden Markov Model based dance recognition Dragutin Hrenek, Nenad Mikša, Robert Perica, Pavle Prentašić and Boris Trubić University of Zagreb, Faculty of Electrical Engineering and Computing Unska 3,
More informationExtracting Significant Patterns from Musical Strings: Some Interesting Problems.
Extracting Significant Patterns from Musical Strings: Some Interesting Problems. Emilios Cambouropoulos Austrian Research Institute for Artificial Intelligence Vienna, Austria emilios@ai.univie.ac.at Abstract
More informationEarly Applications of Information Theory to Music
Early Applications of Information Theory to Music Marcus T. Pearce Centre for Cognition, Computation and Culture, Goldsmiths College, University of London, New Cross, London SE14 6NW m.pearce@gold.ac.uk
More informationPOST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS
POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS Andrew N. Robertson, Mark D. Plumbley Centre for Digital Music
More informationExperiments on musical instrument separation using multiplecause
Experiments on musical instrument separation using multiplecause models J Klingseisen and M D Plumbley* Department of Electronic Engineering King's College London * - Corresponding Author - mark.plumbley@kcl.ac.uk
More information10 Visualization of Tonal Content in the Symbolic and Audio Domains
10 Visualization of Tonal Content in the Symbolic and Audio Domains Petri Toiviainen Department of Music PO Box 35 (M) 40014 University of Jyväskylä Finland ptoiviai@campus.jyu.fi Abstract Various computational
More informationInstrument Recognition in Polyphonic Mixtures Using Spectral Envelopes
Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu
More informationTonal Cognition INTRODUCTION
Tonal Cognition CAROL L. KRUMHANSL AND PETRI TOIVIAINEN Department of Psychology, Cornell University, Ithaca, New York 14853, USA Department of Music, University of Jyväskylä, Jyväskylä, Finland ABSTRACT:
More informationAutomatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors *
Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors * David Ortega-Pacheco and Hiram Calvo Centro de Investigación en Computación, Instituto Politécnico Nacional, Av. Juan
More informationarxiv: v1 [cs.sd] 8 Jun 2016
Symbolic Music Data Version 1. arxiv:1.5v1 [cs.sd] 8 Jun 1 Christian Walder CSIRO Data1 7 London Circuit, Canberra,, Australia. christian.walder@data1.csiro.au June 9, 1 Abstract In this document, we introduce
More informationMusical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons
Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Róisín Loughran roisin.loughran@ul.ie Jacqueline Walker jacqueline.walker@ul.ie Michael O Neill University
More informationCharacteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals
Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals Eita Nakamura and Shinji Takaki National Institute of Informatics, Tokyo 101-8430, Japan eita.nakamura@gmail.com, takaki@nii.ac.jp
More informationSimilarity and Categorisation in Boulez Parenthèse from the Third Piano Sonata: A Formal Analysis.
Similarity and Categorisation in Boulez Parenthèse from the Third Piano Sonata: A Formal Analysis. Christina Anagnostopoulou? and Alan Smaill y y? Faculty of Music, University of Edinburgh Division of
More informationPitch Spelling Algorithms
Pitch Spelling Algorithms David Meredith Centre for Computational Creativity Department of Computing City University, London dave@titanmusic.com www.titanmusic.com MaMuX Seminar IRCAM, Centre G. Pompidou,
More informationAPPENDIX A: ERRATA TO SCORES OF THE PLAYER PIANO STUDIES
APPENDIX A: ERRATA TO SCORES OF THE PLAYER PIANO STUDIES Conlon Nancarrow s hand-written scores, while generally quite precise, contain numerous errors. Most commonly these are errors of omission (e.g.,
More informationMUSI-6201 Computational Music Analysis
MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)
More informationOn time: the influence of tempo, structure and style on the timing of grace notes in skilled musical performance
RHYTHM IN MUSIC PERFORMANCE AND PERCEIVED STRUCTURE 1 On time: the influence of tempo, structure and style on the timing of grace notes in skilled musical performance W. Luke Windsor, Rinus Aarts, Peter
More informationSYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS
Published by Institute of Electrical Engineers (IEE). 1998 IEE, Paul Masri, Nishan Canagarajah Colloquium on "Audio and Music Technology"; November 1998, London. Digest No. 98/470 SYNTHESIS FROM MUSICAL
More informationTake a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University
Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You Chris Lewis Stanford University cmslewis@stanford.edu Abstract In this project, I explore the effectiveness of the Naive Bayes Classifier
More informationVarious Artificial Intelligence Techniques For Automated Melody Generation
Various Artificial Intelligence Techniques For Automated Melody Generation Nikahat Kazi Computer Engineering Department, Thadomal Shahani Engineering College, Mumbai, India Shalini Bhatia Assistant Professor,
More informationHarmony and tonality The vertical dimension. HST 725 Lecture 11 Music Perception & Cognition
Harvard-MIT Division of Health Sciences and Technology HST.725: Music Perception and Cognition Prof. Peter Cariani Harmony and tonality The vertical dimension HST 725 Lecture 11 Music Perception & Cognition
More informationLaboratory Assignment 3. Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB
Laboratory Assignment 3 Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB PURPOSE In this laboratory assignment, you will use MATLAB to synthesize the audio tones that make up a well-known
More informationAbout Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance
Methodologies for Expressiveness Modeling of and for Music Performance by Giovanni De Poli Center of Computational Sonology, Department of Information Engineering, University of Padova, Padova, Italy About
More informationImproving Performance in Neural Networks Using a Boosting Algorithm
- Improving Performance in Neural Networks Using a Boosting Algorithm Harris Drucker AT&T Bell Laboratories Holmdel, NJ 07733 Robert Schapire AT&T Bell Laboratories Murray Hill, NJ 07974 Patrice Simard
More informationLEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception
LEARNING AUDIO SHEET MUSIC CORRESPONDENCES Matthias Dorfer Department of Computational Perception Short Introduction... I am a PhD Candidate in the Department of Computational Perception at Johannes Kepler
More informationEtna Builder - Interactively Building Advanced Graphical Tree Representations of Music
Etna Builder - Interactively Building Advanced Graphical Tree Representations of Music Wolfgang Chico-Töpfer SAS Institute GmbH In der Neckarhelle 162 D-69118 Heidelberg e-mail: woccnews@web.de Etna Builder
More informationCategorization of ICMR Using Feature Extraction Strategy And MIR With Ensemble Learning
Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 57 (2015 ) 686 694 3rd International Conference on Recent Trends in Computing 2015 (ICRTC-2015) Categorization of ICMR
More informationReducing False Positives in Video Shot Detection
Reducing False Positives in Video Shot Detection Nithya Manickam Computer Science & Engineering Department Indian Institute of Technology, Bombay Powai, India - 400076 mnitya@cse.iitb.ac.in Sharat Chandran
More informationExpressive performance in music: Mapping acoustic cues onto facial expressions
International Symposium on Performance Science ISBN 978-94-90306-02-1 The Author 2011, Published by the AEC All rights reserved Expressive performance in music: Mapping acoustic cues onto facial expressions
More informationAn Empirical Comparison of Tempo Trackers
An Empirical Comparison of Tempo Trackers Simon Dixon Austrian Research Institute for Artificial Intelligence Schottengasse 3, A-1010 Vienna, Austria simon@oefai.at An Empirical Comparison of Tempo Trackers
More informationA Beat Tracking System for Audio Signals
A Beat Tracking System for Audio Signals Simon Dixon Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria. simon@ai.univie.ac.at April 7, 2000 Abstract We present
More informationarxiv: v1 [cs.lg] 15 Jun 2016
Deep Learning for Music arxiv:1606.04930v1 [cs.lg] 15 Jun 2016 Allen Huang Department of Management Science and Engineering Stanford University allenh@cs.stanford.edu Abstract Raymond Wu Department of
More informationy POWER USER MUSIC PRODUCTION and PERFORMANCE With the MOTIF ES Mastering the Sample SLICE function
y POWER USER MUSIC PRODUCTION and PERFORMANCE With the MOTIF ES Mastering the Sample SLICE function Phil Clendeninn Senior Product Specialist Technology Products Yamaha Corporation of America Working with
More informationMusic Mood. Sheng Xu, Albert Peyton, Ryan Bhular
Music Mood Sheng Xu, Albert Peyton, Ryan Bhular What is Music Mood A psychological & musical topic Human emotions conveyed in music can be comprehended from two aspects: Lyrics Music Factors that affect
More informationEach copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printed page of such transmission.
Modeling the Perception of Tonal Structure with Neural Nets Author(s): Jamshed J. Bharucha and Peter M. Todd Source: Computer Music Journal, Vol. 13, No. 4 (Winter, 1989), pp. 44-53 Published by: The MIT
More informationMusic Composition with RNN
Music Composition with RNN Jason Wang Department of Statistics Stanford University zwang01@stanford.edu Abstract Music composition is an interesting problem that tests the creativity capacities of artificial
More informationRhythm together with melody is one of the basic elements in music. According to Longuet-Higgins
5 Quantisation Rhythm together with melody is one of the basic elements in music. According to Longuet-Higgins ([LH76]) human listeners are much more sensitive to the perception of rhythm than to the perception
More informationDeep Neural Networks Scanning for patterns (aka convolutional networks) Bhiksha Raj
Deep Neural Networks Scanning for patterns (aka convolutional networks) Bhiksha Raj 1 Story so far MLPs are universal function approximators Boolean functions, classifiers, and regressions MLPs can be
More informationThe Human, the Mechanical, and the Spaces in between: Explorations in Human-Robotic Musical Improvisation
Musical Metacreation: Papers from the 2013 AIIDE Workshop (WS-13-22) The Human, the Mechanical, and the Spaces in between: Explorations in Human-Robotic Musical Improvisation Scott Barton Worcester Polytechnic
More informationPredicting the immediate future with Recurrent Neural Networks: Pre-training and Applications
Predicting the immediate future with Recurrent Neural Networks: Pre-training and Applications Introduction Brandon Richardson December 16, 2011 Research preformed from the last 5 years has shown that the
More informationExploring the Rules in Species Counterpoint
Exploring the Rules in Species Counterpoint Iris Yuping Ren 1 University of Rochester yuping.ren.iris@gmail.com Abstract. In this short paper, we present a rule-based program for generating the upper part
More informationCS229 Project Report Polyphonic Piano Transcription
CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project
More informationHowever, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene
Beat Extraction from Expressive Musical Performances Simon Dixon, Werner Goebl and Emilios Cambouropoulos Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria.
More informationEvaluating Melodic Encodings for Use in Cover Song Identification
Evaluating Melodic Encodings for Use in Cover Song Identification David D. Wickland wickland@uoguelph.ca David A. Calvert dcalvert@uoguelph.ca James Harley jharley@uoguelph.ca ABSTRACT Cover song identification
More informationA Case Based Approach to the Generation of Musical Expression
A Case Based Approach to the Generation of Musical Expression Taizan Suzuki Takenobu Tokunaga Hozumi Tanaka Department of Computer Science Tokyo Institute of Technology 2-12-1, Oookayama, Meguro, Tokyo
More informationOutline. Why do we classify? Audio Classification
Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification Implementation Future Work Why do we classify
More informationInternational Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC
Scientific Journal of Impact Factor (SJIF): 5.71 International Journal of Advance Engineering and Research Development Volume 5, Issue 04, April -2018 e-issn (O): 2348-4470 p-issn (P): 2348-6406 MUSICAL
More informationStructure and Interpretation of Rhythm and Timing 1
henkjan honing Structure and Interpretation of Rhythm and Timing Rhythm, as it is performed and perceived, is only sparingly addressed in music theory. Eisting theories of rhythmic structure are often
More informationMelodic Pattern Segmentation of Polyphonic Music as a Set Partitioning Problem
Melodic Pattern Segmentation of Polyphonic Music as a Set Partitioning Problem Tsubasa Tanaka and Koichi Fujii Abstract In polyphonic music, melodic patterns (motifs) are frequently imitated or repeated,
More informationBach-Prop: Modeling Bach s Harmonization Style with a Back- Propagation Network
Indiana Undergraduate Journal of Cognitive Science 1 (2006) 3-14 Copyright 2006 IUJCS. All rights reserved Bach-Prop: Modeling Bach s Harmonization Style with a Back- Propagation Network Rob Meyerson Cognitive
More informationEach copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printed page of such transmission.
A Connectionist Approach to Algorithmic Composition Author(s): Peter M. Todd Source: Computer Music Journal, Vol. 13, No. 4 (Winter, 1989), pp. 27-43 Published by: The MIT Press Stable URL: http://www.jstor.org/stable/3679551
More informationMelody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng
Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the
More informationPattern Discovery and Matching in Polyphonic Music and Other Multidimensional Datasets
Pattern Discovery and Matching in Polyphonic Music and Other Multidimensional Datasets David Meredith Department of Computing, City University, London. dave@titanmusic.com Geraint A. Wiggins Department
More informationJazz Melody Generation and Recognition
Jazz Melody Generation and Recognition Joseph Victor December 14, 2012 Introduction In this project, we attempt to use machine learning methods to study jazz solos. The reason we study jazz in particular
More informationUniversity of California Press is collaborating with JSTOR to digitize, preserve and extend access to Music Perception: An Interdisciplinary Journal.
An Algorithm for Harmonic Analysis Author(s): David Temperley Source: Music Perception: An Interdisciplinary Journal, Vol. 15, No. 1 (Fall, 1997), pp. 31-68 Published by: University of California Press
More informationImprovised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment
Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment Gus G. Xia Dartmouth College Neukom Institute Hanover, NH, USA gxia@dartmouth.edu Roger B. Dannenberg Carnegie
More informationa start time signature, an end time signature, a start divisions value, an end divisions value, a start beat, an end beat.
The KIAM System in the C@merata Task at MediaEval 2016 Marina Mytrova Keldysh Institute of Applied Mathematics Russian Academy of Sciences Moscow, Russia mytrova@keldysh.ru ABSTRACT The KIAM system is
More informationHST 725 Music Perception & Cognition Assignment #1 =================================================================
HST.725 Music Perception and Cognition, Spring 2009 Harvard-MIT Division of Health Sciences and Technology Course Director: Dr. Peter Cariani HST 725 Music Perception & Cognition Assignment #1 =================================================================
More information2. Problem formulation
Artificial Neural Networks in the Automatic License Plate Recognition. Ascencio López José Ignacio, Ramírez Martínez José María Facultad de Ciencias Universidad Autónoma de Baja California Km. 103 Carretera
More informationCourse Overview. Assessments What are the essential elements and. aptitude and aural acuity? meaning and expression in music?
BEGINNING PIANO / KEYBOARD CLASS This class is open to all students in grades 9-12 who wish to acquire basic piano skills. It is appropriate for students in band, orchestra, and chorus as well as the non-performing
More informationA Bayesian Network for Real-Time Musical Accompaniment
A Bayesian Network for Real-Time Musical Accompaniment Christopher Raphael Department of Mathematics and Statistics, University of Massachusetts at Amherst, Amherst, MA 01003-4515, raphael~math.umass.edu
More informationLab P-6: Synthesis of Sinusoidal Signals A Music Illusion. A k cos.! k t C k / (1)
DSP First, 2e Signal Processing First Lab P-6: Synthesis of Sinusoidal Signals A Music Illusion Pre-Lab: Read the Pre-Lab and do all the exercises in the Pre-Lab section prior to attending lab. Verification:
More informationMusic Radar: A Web-based Query by Humming System
Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,
More informationPHYSICS OF MUSIC. 1.) Charles Taylor, Exploring Music (Music Library ML3805 T )
REFERENCES: 1.) Charles Taylor, Exploring Music (Music Library ML3805 T225 1992) 2.) Juan Roederer, Physics and Psychophysics of Music (Music Library ML3805 R74 1995) 3.) Physics of Sound, writeup in this
More informationThe Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng
The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng S. Zhu, P. Ji, W. Kuang and J. Yang Institute of Acoustics, CAS, O.21, Bei-Si-huan-Xi Road, 100190 Beijing,
More informationPredicting Variation of Folk Songs: A Corpus Analysis Study on the Memorability of Melodies Janssen, B.D.; Burgoyne, J.A.; Honing, H.J.
UvA-DARE (Digital Academic Repository) Predicting Variation of Folk Songs: A Corpus Analysis Study on the Memorability of Melodies Janssen, B.D.; Burgoyne, J.A.; Honing, H.J. Published in: Frontiers in
More informationVISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS. O. Javed, S. Khan, Z. Rasheed, M.Shah. {ojaved, khan, zrasheed,
VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS O. Javed, S. Khan, Z. Rasheed, M.Shah {ojaved, khan, zrasheed, shah}@cs.ucf.edu Computer Vision Lab School of Electrical Engineering and Computer
More informationCHAPTER ONE TWO-PART COUNTERPOINT IN FIRST SPECIES (1:1)
HANDBOOK OF TONAL COUNTERPOINT G. HEUSSENSTAMM Page 1 CHAPTER ONE TWO-PART COUNTERPOINT IN FIRST SPECIES (1:1) What is counterpoint? Counterpoint is the art of combining melodies; each part has its own
More informationModeling the Effect of Meter in Rhythmic Categorization: Preliminary Results
Modeling the Effect of Meter in Rhythmic Categorization: Preliminary Results Peter Desain and Henkjan Honing,2 Music, Mind, Machine Group NICI, University of Nijmegen P.O. Box 904, 6500 HE Nijmegen The
More informationBroken Wires Diagnosis Method Numerical Simulation Based on Smart Cable Structure
PHOTONIC SENSORS / Vol. 4, No. 4, 2014: 366 372 Broken Wires Diagnosis Method Numerical Simulation Based on Smart Cable Structure Sheng LI 1*, Min ZHOU 2, and Yan YANG 3 1 National Engineering Laboratory
More informationThe Sparsity of Simple Recurrent Networks in Musical Structure Learning
The Sparsity of Simple Recurrent Networks in Musical Structure Learning Kat R. Agres (kra9@cornell.edu) Department of Psychology, Cornell University, 211 Uris Hall Ithaca, NY 14853 USA Jordan E. DeLong
More informationPolyphonic music transcription through dynamic networks and spectral pattern identification
Polyphonic music transcription through dynamic networks and spectral pattern identification Antonio Pertusa and José M. Iñesta Departamento de Lenguajes y Sistemas Informáticos Universidad de Alicante,
More informationComputing, Artificial Intelligence, and Music. A History and Exploration of Current Research. Josh Everist CS 427 5/12/05
Computing, Artificial Intelligence, and Music A History and Exploration of Current Research Josh Everist CS 427 5/12/05 Introduction. As an art, music is older than mathematics. Humans learned to manipulate
More informationWeek 14 Music Understanding and Classification
Week 14 Music Understanding and Classification Roger B. Dannenberg Professor of Computer Science, Music & Art Overview n Music Style Classification n What s a classifier? n Naïve Bayesian Classifiers n
More informationCalculating Dissonance in Chopin s Étude Op. 10 No. 1
Calculating Dissonance in Chopin s Étude Op. 10 No. 1 Nikita Mamedov and Robert Peck Department of Music nmamed1@lsu.edu Abstract. The twenty-seven études of Frédéric Chopin are exemplary works that display
More informationinter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE
Copyright SFA - InterNoise 2000 1 inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering 27-30 August 2000, Nice, FRANCE I-INCE Classification: 7.9 THE FUTURE OF SOUND
More informationStudy Guide. Solutions to Selected Exercises. Foundations of Music and Musicianship with CD-ROM. 2nd Edition. David Damschroder
Study Guide Solutions to Selected Exercises Foundations of Music and Musicianship with CD-ROM 2nd Edition by David Damschroder Solutions to Selected Exercises 1 CHAPTER 1 P1-4 Do exercises a-c. Remember
More informationA Framework for Segmentation of Interview Videos
A Framework for Segmentation of Interview Videos Omar Javed, Sohaib Khan, Zeeshan Rasheed, Mubarak Shah Computer Vision Lab School of Electrical Engineering and Computer Science University of Central Florida
More informationThe Tone Height of Multiharmonic Sounds. Introduction
Music-Perception Winter 1990, Vol. 8, No. 2, 203-214 I990 BY THE REGENTS OF THE UNIVERSITY OF CALIFORNIA The Tone Height of Multiharmonic Sounds ROY D. PATTERSON MRC Applied Psychology Unit, Cambridge,
More informationSome researchers in the computational sciences have considered music computation, including music reproduction
INFORMS Journal on Computing Vol. 18, No. 3, Summer 2006, pp. 321 338 issn 1091-9856 eissn 1526-5528 06 1803 0321 informs doi 10.1287/ioc.1050.0131 2006 INFORMS Recurrent Neural Networks for Music Computation
More informationAshton Allan MU 228 Tonality within Aaron Copland s Piano Variations
Ashton Allan MU 228 Tonality within Aaron Copland s Piano Variations The closest Aaron Copland ever got to atonal music was his 1930 composition, Piano Variations. This work, constructed from twenty independently
More informationTHE EFFECT OF EXPERTISE IN EVALUATING EMOTIONS IN MUSIC
THE EFFECT OF EXPERTISE IN EVALUATING EMOTIONS IN MUSIC Fabio Morreale, Raul Masu, Antonella De Angeli, Patrizio Fava Department of Information Engineering and Computer Science, University Of Trento, Italy
More informationPerceptual Evaluation of Automatically Extracted Musical Motives
Perceptual Evaluation of Automatically Extracted Musical Motives Oriol Nieto 1, Morwaread M. Farbood 2 Dept. of Music and Performing Arts Professions, New York University, USA 1 oriol@nyu.edu, 2 mfarbood@nyu.edu
More informationAutomatic meter extraction from MIDI files (Extraction automatique de mètres à partir de fichiers MIDI)
Journées d'informatique Musicale, 9 e édition, Marseille, 9-1 mai 00 Automatic meter extraction from MIDI files (Extraction automatique de mètres à partir de fichiers MIDI) Benoit Meudic Ircam - Centre
More informationFlorida Performing Fine Arts Assessment Item Specifications for Benchmarks in Course: M/J Chorus 3
Task A/B/C/D Item Type Florida Performing Fine Arts Assessment Course Title: M/J Chorus 3 Course Number: 1303020 Abbreviated Title: M/J CHORUS 3 Course Length: Year Course Level: 2 PERFORMING Benchmarks
More informationDoctor of Philosophy
University of Adelaide Elder Conservatorium of Music Faculty of Humanities and Social Sciences Declarative Computer Music Programming: using Prolog to generate rule-based musical counterpoints by Robert
More informationFrom Score to Performance: A Tutorial to Rubato Software Part I: Metro- and MeloRubette Part II: PerformanceRubette
From Score to Performance: A Tutorial to Rubato Software Part I: Metro- and MeloRubette Part II: PerformanceRubette May 6, 2016 Authors: Part I: Bill Heinze, Alison Lee, Lydia Michel, Sam Wong Part II:
More informationChords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm
Georgia State University ScholarWorks @ Georgia State University Music Faculty Publications School of Music 2013 Chords not required: Incorporating horizontal and vertical aspects independently in a computer
More information