CREATING all forms of art [1], [2], [3], [4], including

Size: px
Start display at page:

Download "CREATING all forms of art [1], [2], [3], [4], including"

Transcription

1 Grammar Argumented LSTM Neural Networks with Note-Level Encoding for Music Composition Zheng Sun, Jiaqi Liu, Zewang Zhang, Jingwen Chen, Zhao Huo, Ching Hua Lee, and Xiao Zhang 1 arxiv: v1 [cs.lg] 16 Nov 2016 Abstract Creating any aesthetically pleasing piece of art, like music, has been a long time dream for artificial intelligence research. Based on recent success of long-short term memory (LSTM) on sequence learning, we put forward a novel system to reflect the thinking pattern of a musician. For data representation, we propose a note-level encoding method, which enables our model to simulate how human composes and polishes music phrases. To avoid failure against music theory, we invent a novel method, grammar argumented (GA) method. It can teach machine basic composing principles. In this method, we propose three rules as argumented grammars and three metrics for evaluation of machine-made music. Results show that comparing to basic LSTM, grammar argumented model s compositions have higher contents of diatonic scale notes, short pitch intervals, and chords. Index Terms Music composition, note-level encoding, LSTM neural networks, grammar argumented method 1 INTRODUCTION CREATING all forms of art [1], [2], [3], [4], including music, has been a long time pursue for artificial intelligence (AI) research. Even AI follows the principle as stated by the musicologist Leonard B. Meyer, that styles of music are in effect complex systems of probability relationships. In the early years, symbolic AI methods were popular and specific grammars describing a set of rules drive the composition [5], [6]. This method was latter improved by evolutionary algorithms in different ways [7], including the famous EMI project [8]. Latter, statistic model such as Markov chains and Hidden Markov model(hmm) became popular in algorithmic composition [9]. In the meanwhile, neural network(nn) has made remarkable progress in recognition and other field [10], including music composition using Recurrent neural networks(rnn) [11], [12], [13], [14] and Long-Short Term Memory(LSTM) [15], [16]. RNN and LSTM perform well in modeling sequential data, however, when they are applied on music composition, the outcomes will be drab and dull, and even fall into a group of harsh notes. Moreover, the generated music sometimes go against general music theory. We hope to teach machine basic composing principles, but merely neural networks or older grammar methods are incapable of it. In this work, we improvise LSTM with an original method named grammar argumented method (GA), which combine neural networks and grammars. We begin by training a LSTM neural network with a dataset full of musicianmade music. In the training process, machine learns sequencial information as much as possible. Then we feed a short Z. Sun, J. Liu, Z. Zhang J. Chen and X. Zhang are with the Department of Physics, Sun Yat-sen University, Guangzhou, P. R. China. {sunzh6, liujq33, zhangzw3, chenjw93, zhangxiao}@{mail2, mail2, mail2, mail2, mail}.sysu.edu.cn Z. Huo is with China University of Political Science and Law. huozhao@cupl.edu.cn C. H. Lee is with Institute of High Performance Computing. calvin-lee@ihpc.a-star.edu.sg Manuscript received xx/xx/xxxx; revised xx/xx/xxxx. music phrase to trigger the first phase of generation. Instead of adding the generated notes directly to music score, we evaluate each predicted result with some composing rules. The notes go against general music theory will be abandoned and new notes confirmed to rules can be accessed by repredicting. All amended results and their corresponding inputs will be added to training set, we retrain our model with updated training set and use the original generating method to do the second phase of (real) generation. We also adopt a new representation of notes by concatenating each note s duration and pitch as a single input vector. With this note-level encodig method, we enable machine to process real notes and think like human. Results show that our system is capable of composing pleasing melody, and it performs better than non-ga system in the percentages of notes in diatonic scale, pitch intervals in one octave, and chords. In other words, our specific system can learn basic composing principles and bring out common and melodious music. 2 METHODS 2.1 Note-Level Encoding Although machine learning method makes a significant progress in music composition, none of the related works really simulates how human creates music. During music composition, composers often focus on several typical music phrases. In the process of polishing music phrases, they need to deliberate each music note. Music note is a particular combination of pitch and duration, like a quarter note of E6. However, the related works either represent music as quantized time series [11], [13], [15], [17], [18], [19], or treat pitches and durations in separate neural networks [20], [21]. In this work, we encode music with note-level method. Because it is troublesome to simulate the process of polishing and generating notes with raw MIDI files, we need to convert each file to a sequence of music notes. First, we choose MIDI files of 106 Soft Piano music whose composers include Joe Hisaishi, Yiruma, Yoko Kanno and Shi Jin, et

2 2 Fig. 1. We extract music score from the MIDI files. Each note symbol contains the information of duration and pitch, both of which are then converted into one-hot vector and concatenated into one specific binary vector. (a) al. These compositions time signature are all 4/4. We only pick music with time signature 4/4 because music with different time structures is intractable for modeling. Then, we delete all the accompaniments and score annotations like Control Change Events and grace notes. Besides, we ignore the change of intensity and just keep the highest pitch note when more than two notes appear at the same time. This treatment can discretize melodies into a sequence of Note On Events and Note Off Events. We then encode each pair of Note On Event and Note Off Event into one 2-D vector. Each vector includes two dimensions: duration ranging from semiquaver to breve and pitch ranging from A0 to C8. For reducing dimentions of model s input layer, we transpose all melodies into C major/a minor, which eventually include 30 kinds of durations and 59 kinds of pitches. Then, we convert each note s duration and pitch into corresponding one-hot vector, one has 30 bits, another has 59 bits, as in Fig.1. Concatenate these two binary vectors, we can get specific representation for each note. Unlike the common used encoding methods, piano-roll representation, this method can represent both pitch and duration in one single vector. What is more significant, a sequence of real music notes enable our model to simulate how human generates and polishes music phrases. 2.2 Long-Short Term Memory Neural Networks Recurrent neural networks (RNN) perform well in sequence learning. Each hidden layer receives both input from the previous layer and input from itself one time step in the past. It enables the networks remember information of previous time steps. However, simple RNN does not perform well in long-term dependency because of vanishing gradients [22]. Long-short term memory (LSTM) neural network is a kind of advanced RNN and solve this problem. It use memory cell to store necessary information and various gates to control other information. The graphic structure of LSTM in Fig.2 shows how the data flows through the LSTM module. In the following we describe how the hidden state u (t) of LSTM is computed. To obtain u (t) for each time step, we input data v (t) and u (t 1) from last time step into the LSTM. b i, b o, b f, b c denote corresponding vectors of biases, whereas the subscript i, o, f, c means input, output, forget and cell gate, sequentially. W i, W o, W f, W c denote the matrices of weights between input vector and corresponding gate units and (b) Fig. 2. The graphic structure of LSTM. With v (t) and u (t 1) as inputs, LSTM layer outputs a hidden state u (t) that can convey temporal information. denotes element-wise multiplication and + denotes elementwise addition. The schematic diagram (a) shows what to input and output and (b) indicates the details that how the LSTM layer computes the current hidden state u (t). U i, U o, U f, U c denote the matrices of weights connecting hidden state from last time step to the corresponding gate units. First, with the previous temporal information u (t 1), we construct input gate i (t), output gate o (t) and forget gate f (t) to decide which parts of information from memory cell to pass through the gate. σ(x) = (1 + e ( x) ) 1 is an elementwise sigmoid function whose output value is between (0, 1) interval. i (t) = σ(w i v (t) + U i u (t 1) + b i ) (1) o (t) = σ(w o v (t) + U o u (t 1) + b o ) (2) f (t) = σ(w f v (t) + U f u (t 1) + b f ) (3) In the same way, we compute the new state C (t) of memory cell depending on the value of input v (t) and hidden state u (t 1) from last time step. The difference is that the activation here is tanh(x) = (1 e 2x )/(1 + e 2x ). C (t) = tanh(w c v (t) + U c u (t 1) + b c ) (4) Then we can define the update rule for the current state C (t) of memory cell of LSTM, a unit that stores and accumulates information. It is the new state C (t) of memory cell multiplied by the output of input gate i (t) plus the old state C (t 1) of memory cell from last time step multiplied by the output of forget gate f (t) : C (t) = i (t) C (t) + f (t) C (t 1) (5)

3 3 Finally, the current hidden state u (t) of LSTM is computed by the activated current state of memory cell under the control of output gate. u (t) = o (t) tanh(c (t) ) (6) 2.3 Grammar Argumented Method In an original method, after hundreds of epoches, the loss of model stops decreasing, we will consider machine has learned as much information as possible from the dataset. Once a seed input is given, the model will be able to predict notes continuously. However, it does not perform as well as expected because the model often generates results not conformed to basic composing principles, for example, too many overtones, acute change on pitch and unpleasing melodies. These results can ruin the composition. Our GA method can teach machine to learn general music knowledge and compose more harmonious notes without any manual intervention on the second phase of (real) generation. Before the details of GA go, we need to determine what kinds of rules to use in GA method. According to music theory and unmelodious pieces in results, we put forward three kinds of rules. The first rule is diatonic scale (Dia). In music theory, notes in a diatonic scale are the most pleasing. The C major scale is one of the diatonic scales and it is made up of seven distinct notes (C, D, E, F, G, A, and B). Because all data is converted to C major in this work, we choose C major scale rule in our experiments. Conventionally, machine without GA method often produces overtones (C#, D#, F#, G#, and A#) and ruins a nice piece. Although overtones sometimes have positive effects on music pieces, it is too difficult for machine to compose melodious music with all twelve tones in one octave. So we hope there are less overtones and more notes in C major scale in machine s compositions. The second rule is short pitch interval (SPI). The pitch interval of two successive notes usually does not span over one octave. Acute change on pitch often makes listener feel uncomfortable. We think only experienced composers can use pitch intervals spanning over one octave well, so short pitch intervals are more preferred in our results. The last rule is triads (Tri). In addition to notes in a scale, chords are also the key of composing tuneful music. Triads are the simplest chords and each kind of triads represents a pair of pitch intervals. There are four kinds of triads in total and each of them has specific music emotion. Furthermore, triads are also the basis of all seventh chords, which result in variegated music. We believe that composition s level is closely related to the amount of chords. With GA method, in the first phase of generation, we aim at achieving data amended with composing rules stated above. Before a fresh note is added to music score, we check that whether or not this new note are conformed to rules. When a discordant note is predicted, we go back to the output layer of model and resample from the output distribution, as in Fig.3. Loop this operation until a note conformed to rule is generated. For example, we suppose that the fresh note is (eighth, B6) and the last note in music score is (eighth, A5), their pitch interval will be 14 semitones. Because one octave only includes 12 semitones and this new note is non-conformed to SPI. Although pitch B6 may has the highest probability in output layer, we abandon it and resample for other notes conformed to SPI. When a note whose pitch is conformed to SPI appears, we add it to music score. At the same time, we record this amended result and the current input phrase. After the generating process, we mix all recored data with original training set, like what is shown at the bottom of Fig.3. In other words, we add a handful of fabricated data which includes the information of basic composing rules. Lastly, we retrain the model with the updated training set. In the second phase of (real) generation, we adopt original method, let machine generate notes continuously without any extra mechanism. This GA method enables machine to learn basic composing principles from data amended with music theory and bring out melodious music. 3 EXPERIMENTS Our model consists of one LSTM layer and one fully connected layer. The LSTM layer includes 128 cells and its input dimension is 89, which equals to the length of note s binary representation. There are 89 nodes in fully connected layer and it is also the output layer. The size of our dataset is 30000, to speed up the training process, we divide it into mini-batches, and the batch size is 64. We use Adam [23] to perform gradient descent optimization, learning rate is set to We build our model on a high-level neural networks library Keras [24] and use T ensorf low [25] as its tensor manipulation library. We train this model with original dataset and label this set of weights with Orig. With GA method, we use Orig to generate 100k notes for each rule and get three sets of amended data. Then we mix each set of data with dataset, mix all three sets with dataset, and get four new training sets at last. We retrain our model with them and label these four sets of weights with Dia, SPI, Tri, and MIX, corresponding to their rules. For statistics analysis, we adopt a public random seed to generate 100k notes with all five sets of weights, including Orig. 4 RESULTS We take a representative piece from machine s long composition, which is generated in MIX mode. This piece s music score is shown in Fig.4. At first, it is a strong evidence that machine excludes overtones and prefers notes in C major scale. There is only one overtone in this piece. Secondly, it is worth noticing that machine has already learned to use repeated rhythmic structure in one piece as human always does, for example, bar 4-5 and bar Listen from the beginning, we notice that the whole piece sounds soft and lyrical, which is consistent with music in dataset. Bar 3, 4, 6, and 12 have faster paces sandwiched between slow moving melodies. This variety of paces is also seen in musicianmade music. Besides, the bottom half is so pleasing that it can be directly used as theme of a new song. 5 EVALUATIONS According to music theory described in section2.3, we put forward three kinds of metrics: percentages of notes in

4 4 Fig. 3. The grammar argumented method. With a music phrase or note sequence, a well-trained LSTM neural network is able to predict the next note. The output layer consists of 89 nodes, according to note-level encoding method, the left 30 nodes represent duration and the right 59 nodes are for pitch. We sample from these two parts because duration and pitch s representations are all one-hot codes. After the sampling result is converted to note, we check it with composing rule, only notes comformed to rule can be added to phrase. When a non-comforming note is predicted, we resample on the output distribution until the comforming one appears. Then, we add this amended note and its corresponding input phrase to the original training set. Fig. 4. An example of machine s composition. It is generated in MIX mode and includes about 100 notes. diatonic scale (pdia), percentage of pitch intervals in one octave (pspi) and percentages of triads (ptri). All of them are based on note-level encoding, but they are generally valid as evaluation criteria for other composing algorithms. 5.1 pdia Table 1 shows the percentages of seven notes in C major scale, we find that GA method works with C major scale rule. For E6, its ratio increases by one percentage points. Except for Dia and MIX, Tri mode also produces high pdia. This phenomenon can be explained by pitch interval. Because pitch intervals in triads are included in major scale, adding data amended with triads rule also improves pdia. 5.2 pspi We figure out each pitch interval in the composition and calculate the percentage of pitch intervals in one octave. TABLE 1 Percentages of Notes in Diatonic Scale DS Orig Dia SPI Tri MIX C D E F G A B High pspi contributes to a harmonious composition. As what is shown in Table 2, SPI and MIX mode can compose music with high pspi. We also notice that without GA method (Orig mode), model generate more unharmonious notes, which result in more poor compositions.

5 5 5.3 ptri TABLE 2 Percentage of Pitch Intervals in One Octave Modes DS Orig Dia SPI Tri MIX pspi(%) TABLE 3 Percentages of Triads DS Orig Dia SPI Tri MIX Major Minor Augmented Diminished Table 3 shows percentages of triads. In Tri and MIX mode, machines composition includes more triads than other results. We notice that music composed in MIX mode performs well with all three metrics. It suggests that those three rules are not conflicting, they are coherent rules in music theory. 6 CONCLUSION Although it is difficult for machine to learn music theory rules with simple LSTM neural networks, we propose grammar argument method and enable our model to learn those rules by adding rule-amended data to dataset. GA method reduce unharmonious notes amount significantly. Our original note-level encoding method also contributes to this successful system. On note level, machine thinks like musicians and is able to learn advanced logic from the dataset. At last, three metrics provide a solution for evaluating machine-made music. Our GA method has the potential to include high-level music theory rules, such as repeated paragraphs, and it gives an approach towards music with global structure. ACKNOWLEDGMENTS The authors would like to thank Chuangjie Ren and Qingpei Liu for helpful discussions on neural networks. REFERENCES [1] K. Gregor, I. Danihelka, A. Graves, D. J. Rezende, and D. Wierstra, Draw: A recurrent neural network for image generation, Computer Science, [2] A. Graves, Generating sequences with recurrent neural networks, Computer Science, [3] L. A. Gatys, A. S. Ecker, and M. Bethge, A neural algorithm of artistic style, Computer Science, [4] A. van den Oord, N. Kalchbrenner, and K. Kavukcuoglu, Pixel recurrent neural networks, arxiv preprint arxiv: , [5] G. M. Rader, A method for composing simple traditional music by computer, Communications of the ACM, vol. 17, no. 11, pp , [6] J. D. Fernánd Ndez and F. Vico, Ai methods in algorithmic composition: a comprehensive survey, Journal of Artificial Intelligence Research, vol. 48, no. 48, pp , [7] THYWISSEN and KURT, Genotator: an environment for exploring the application of evolutionary techniques in computerassisted composition, Organised Sound, vol. 4, no. 2, pp , [8] D. Cope, Computer modeling of musical intelligence in emi, Computer Music Journal, vol. 16, no. 16, pp , [9] M. Allan, Harmonising chorales in the style of johann sebastian bach, Master s Thesis, School of Informatics, University of Edinburgh, [10] D. Silver, A. Huang, C. J. Maddison, A. Guez, L. Sifre, d. D. G. Van, J. Schrittwieser, I. Antonoglou, V. Panneershelvam, and M. Lanctot, Mastering the game of go with deep neural networks and tree search. Nature, vol. 529, no. 7587, pp , [11] P. M. Todd, A connectionist approach to algorithmic composition, Computer Music Journal, vol. 13, no. 4, pp , [12] M. C. MOZER, Neural network music composition by prediction: Exploring the benefits of psychoacoustic constraints and multiscale processing, Connection Science, vol. 6, no. 2-3, pp , [13] N. Boulanger-Lewandowski, Y. Bengio, and P. Vincent, Modeling temporal dependencies in high-dimensional sequences: Application to polyphonic music generation and transcription, arxiv preprint arxiv: , [14] S. Wermter, C. Weber, W. Duch, T. Honkela, and P. Koprinkovahristova, Artificial neural networks and machine learning icann 2014, Lecture Notes in Computer Science, vol. 8681, [15] D. Eck and J. Lapalme, Learning musical structure directly from sequences of music, University of Montreal, Department of Computer Science, CP, vol. 6128, [16] J. A. Franklin, Recurrent neural networks for music computation, Informs Journal on Computing, vol. 18, no. 3, pp , [17] K. Goel, R. Vohra, and J. Sahoo, Polyphonic music generation by modeling temporal dependencies using a rnn-dbn, in Artificial Neural Networks and Machine Learning ICANN Springer, 2014, pp [18] D. Eck and J. Schmidhuber, Finding temporal structure in music: Blues improvisation with lstm recurrent networks, in Neural Networks for Signal Processing, Proceedings of the th IEEE Workshop on. IEEE, 2002, pp [19] Q. Lyu, Z. Wu, and J. Zhu, Polyphonic music modelling with lstm-rtrbm, in Proceedings of the 23rd Annual ACM Conference on Multimedia Conference. ACM, 2015, pp [20] M. C. Mozer, Neural network music composition by prediction: Exploring the benefits of psychoacoustic constraints and multiscale processing, Connection Science, vol. 6, no. 2-3, pp , [21] J. A. Franklin, Recurrent neural networks for music computation, INFORMS Journal on Computing, vol. 18, no. 3, pp , [22] S. Hochreiter and J. Schmidhuber, Long short-term memory, Neural computation, vol. 9, no. 8, pp , [23] D. Kingma and J. Ba, Adam: A method for stochastic optimization, Computer Science, [24] F. Chollet, Keras, [25] M. Abadi, A. Agarwal, P. Barham, E. Brevdo, Z. Chen, C. Citro, G. S. Corrado, A. Davis, J. Dean, M. Devin, S. Ghemawat, I. Goodfellow, A. Harp, G. Irving, M. Isard, Y. Jia, R. Jozefowicz, L. Kaiser, M. Kudlur, J. Levenberg, D. Mané, R. Monga, S. Moore, D. Murray, C. Olah, M. Schuster, J. Shlens, B. Steiner, I. Sutskever, K. Talwar, P. Tucker, V. Vanhoucke, V. Vasudevan, F. Viégas, O. Vinyals, P. Warden, M. Wattenberg, M. Wicke, Y. Yu, and X. Zheng, TensorFlow: Large-scale machine learning on heterogeneous systems, 2015, software available from tensorflow.org. [Online]. Available:

Music Composition with RNN

Music Composition with RNN Music Composition with RNN Jason Wang Department of Statistics Stanford University zwang01@stanford.edu Abstract Music composition is an interesting problem that tests the creativity capacities of artificial

More information

An AI Approach to Automatic Natural Music Transcription

An AI Approach to Automatic Natural Music Transcription An AI Approach to Automatic Natural Music Transcription Michael Bereket Stanford University Stanford, CA mbereket@stanford.edu Karey Shi Stanford Univeristy Stanford, CA kareyshi@stanford.edu Abstract

More information

Talking Drums: Generating drum grooves with neural networks

Talking Drums: Generating drum grooves with neural networks Talking Drums: Generating drum grooves with neural networks P. Hutchings 1 1 Monash University, Melbourne, Australia arxiv:1706.09558v1 [cs.sd] 29 Jun 2017 Presented is a method of generating a full drum

More information

arxiv: v1 [cs.lg] 15 Jun 2016

arxiv: v1 [cs.lg] 15 Jun 2016 Deep Learning for Music arxiv:1606.04930v1 [cs.lg] 15 Jun 2016 Allen Huang Department of Management Science and Engineering Stanford University allenh@cs.stanford.edu Abstract Raymond Wu Department of

More information

LSTM Neural Style Transfer in Music Using Computational Musicology

LSTM Neural Style Transfer in Music Using Computational Musicology LSTM Neural Style Transfer in Music Using Computational Musicology Jett Oristaglio Dartmouth College, June 4 2017 1. Introduction In the 2016 paper A Neural Algorithm of Artistic Style, Gatys et al. discovered

More information

Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies

Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies Judy Franklin Computer Science Department Smith College Northampton, MA 01063 Abstract Recurrent (neural) networks have

More information

A Unit Selection Methodology for Music Generation Using Deep Neural Networks

A Unit Selection Methodology for Music Generation Using Deep Neural Networks A Unit Selection Methodology for Music Generation Using Deep Neural Networks Mason Bretan Georgia Institute of Technology Atlanta, GA Gil Weinberg Georgia Institute of Technology Atlanta, GA Larry Heck

More information

RoboMozart: Generating music using LSTM networks trained per-tick on a MIDI collection with short music segments as input.

RoboMozart: Generating music using LSTM networks trained per-tick on a MIDI collection with short music segments as input. RoboMozart: Generating music using LSTM networks trained per-tick on a MIDI collection with short music segments as input. Joseph Weel 10321624 Bachelor thesis Credits: 18 EC Bachelor Opleiding Kunstmatige

More information

A STUDY ON LSTM NETWORKS FOR POLYPHONIC MUSIC SEQUENCE MODELLING

A STUDY ON LSTM NETWORKS FOR POLYPHONIC MUSIC SEQUENCE MODELLING A STUDY ON LSTM NETWORKS FOR POLYPHONIC MUSIC SEQUENCE MODELLING Adrien Ycart and Emmanouil Benetos Centre for Digital Music, Queen Mary University of London, UK {a.ycart, emmanouil.benetos}@qmul.ac.uk

More information

CHORD GENERATION FROM SYMBOLIC MELODY USING BLSTM NETWORKS

CHORD GENERATION FROM SYMBOLIC MELODY USING BLSTM NETWORKS CHORD GENERATION FROM SYMBOLIC MELODY USING BLSTM NETWORKS Hyungui Lim 1,2, Seungyeon Rhyu 1 and Kyogu Lee 1,2 3 Music and Audio Research Group, Graduate School of Convergence Science and Technology 4

More information

Modeling Temporal Tonal Relations in Polyphonic Music Through Deep Networks with a Novel Image-Based Representation

Modeling Temporal Tonal Relations in Polyphonic Music Through Deep Networks with a Novel Image-Based Representation INTRODUCTION Modeling Temporal Tonal Relations in Polyphonic Music Through Deep Networks with a Novel Image-Based Representation Ching-Hua Chuan 1, 2 1 University of North Florida 2 University of Miami

More information

Predicting Mozart s Next Note via Echo State Networks

Predicting Mozart s Next Note via Echo State Networks Predicting Mozart s Next Note via Echo State Networks Ąžuolas Krušna, Mantas Lukoševičius Faculty of Informatics Kaunas University of Technology Kaunas, Lithuania azukru@ktu.edu, mantas.lukosevicius@ktu.lt

More information

Learning Musical Structure Directly from Sequences of Music

Learning Musical Structure Directly from Sequences of Music Learning Musical Structure Directly from Sequences of Music Douglas Eck and Jasmin Lapalme Dept. IRO, Université de Montréal C.P. 6128, Montreal, Qc, H3C 3J7, Canada Technical Report 1300 Abstract This

More information

Audio: Generation & Extraction. Charu Jaiswal

Audio: Generation & Extraction. Charu Jaiswal Audio: Generation & Extraction Charu Jaiswal Music Composition which approach? Feed forward NN can t store information about past (or keep track of position in song) RNN as a single step predictor struggle

More information

CONDITIONING DEEP GENERATIVE RAW AUDIO MODELS FOR STRUCTURED AUTOMATIC MUSIC

CONDITIONING DEEP GENERATIVE RAW AUDIO MODELS FOR STRUCTURED AUTOMATIC MUSIC CONDITIONING DEEP GENERATIVE RAW AUDIO MODELS FOR STRUCTURED AUTOMATIC MUSIC Rachel Manzelli Vijay Thakkar Ali Siahkamari Brian Kulis Equal contributions ECE Department, Boston University {manzelli, thakkarv,

More information

Predicting Similar Songs Using Musical Structure Armin Namavari, Blake Howell, Gene Lewis

Predicting Similar Songs Using Musical Structure Armin Namavari, Blake Howell, Gene Lewis Predicting Similar Songs Using Musical Structure Armin Namavari, Blake Howell, Gene Lewis 1 Introduction In this work we propose a music genre classification method that directly analyzes the structure

More information

arxiv: v1 [cs.sd] 17 Dec 2018

arxiv: v1 [cs.sd] 17 Dec 2018 Learning to Generate Music with BachProp Florian Colombo School of Computer Science and School of Life Sciences École Polytechnique Fédérale de Lausanne, Lausanne, Switzerland florian.colombo@epfl.ch arxiv:1812.06669v1

More information

arxiv: v1 [cs.sd] 8 Jun 2016

arxiv: v1 [cs.sd] 8 Jun 2016 Symbolic Music Data Version 1. arxiv:1.5v1 [cs.sd] 8 Jun 1 Christian Walder CSIRO Data1 7 London Circuit, Canberra,, Australia. christian.walder@data1.csiro.au June 9, 1 Abstract In this document, we introduce

More information

Neural Aesthetic Image Reviewer

Neural Aesthetic Image Reviewer Neural Aesthetic Image Reviewer Wenshan Wang 1, Su Yang 1,3, Weishan Zhang 2, Jiulong Zhang 3 1 Shanghai Key Laboratory of Intelligent Information Processing School of Computer Science, Fudan University

More information

Deep Recurrent Music Writer: Memory-enhanced Variational Autoencoder-based Musical Score Composition and an Objective Measure

Deep Recurrent Music Writer: Memory-enhanced Variational Autoencoder-based Musical Score Composition and an Objective Measure Deep Recurrent Music Writer: Memory-enhanced Variational Autoencoder-based Musical Score Composition and an Objective Measure Romain Sabathé, Eduardo Coutinho, and Björn Schuller Department of Computing,

More information

A probabilistic approach to determining bass voice leading in melodic harmonisation

A probabilistic approach to determining bass voice leading in melodic harmonisation A probabilistic approach to determining bass voice leading in melodic harmonisation Dimos Makris a, Maximos Kaliakatsos-Papakostas b, and Emilios Cambouropoulos b a Department of Informatics, Ionian University,

More information

Blues Improviser. Greg Nelson Nam Nguyen

Blues Improviser. Greg Nelson Nam Nguyen Blues Improviser Greg Nelson (gregoryn@cs.utah.edu) Nam Nguyen (namphuon@cs.utah.edu) Department of Computer Science University of Utah Salt Lake City, UT 84112 Abstract Computer-generated music has long

More information

arxiv: v3 [cs.sd] 14 Jul 2017

arxiv: v3 [cs.sd] 14 Jul 2017 Music Generation with Variational Recurrent Autoencoder Supported by History Alexey Tikhonov 1 and Ivan P. Yamshchikov 2 1 Yandex, Berlin altsoph@gmail.com 2 Max Planck Institute for Mathematics in the

More information

arxiv: v1 [cs.sd] 12 Dec 2016

arxiv: v1 [cs.sd] 12 Dec 2016 A Unit Selection Methodology for Music Generation Using Deep Neural Networks Mason Bretan Georgia Tech Atlanta, GA Gil Weinberg Georgia Tech Atlanta, GA Larry Heck Google Research Mountain View, CA arxiv:1612.03789v1

More information

GENERATING NONTRIVIAL MELODIES FOR MUSIC AS A SERVICE

GENERATING NONTRIVIAL MELODIES FOR MUSIC AS A SERVICE GENERATING NONTRIVIAL MELODIES FOR MUSIC AS A SERVICE Yifei Teng U. of Illinois, Dept. of ECE teng9@illinois.edu Anny Zhao U. of Illinois, Dept. of ECE anzhao2@illinois.edu Camille Goudeseune U. of Illinois,

More information

On the mathematics of beauty: beautiful music

On the mathematics of beauty: beautiful music 1 On the mathematics of beauty: beautiful music A. M. Khalili Abstract The question of beauty has inspired philosophers and scientists for centuries, the study of aesthetics today is an active research

More information

arxiv: v1 [cs.sd] 18 Dec 2018

arxiv: v1 [cs.sd] 18 Dec 2018 BANDNET: A NEURAL NETWORK-BASED, MULTI-INSTRUMENT BEATLES-STYLE MIDI MUSIC COMPOSITION MACHINE Yichao Zhou,1,2 Wei Chu,1 Sam Young 1,3 Xin Chen 1 1 Snap Inc. 63 Market St, Venice, CA 90291, 2 Department

More information

Recurrent Neural Networks and Pitch Representations for Music Tasks

Recurrent Neural Networks and Pitch Representations for Music Tasks Recurrent Neural Networks and Pitch Representations for Music Tasks Judy A. Franklin Smith College Department of Computer Science Northampton, MA 01063 jfranklin@cs.smith.edu Abstract We present results

More information

Algorithmic Composition of Melodies with Deep Recurrent Neural Networks

Algorithmic Composition of Melodies with Deep Recurrent Neural Networks Algorithmic Composition of Melodies with Deep Recurrent Neural Networks Florian Colombo, Samuel P. Muscinelli, Alexander Seeholzer, Johanni Brea and Wulfram Gerstner Laboratory of Computational Neurosciences.

More information

Modeling Musical Context Using Word2vec

Modeling Musical Context Using Word2vec Modeling Musical Context Using Word2vec D. Herremans 1 and C.-H. Chuan 2 1 Queen Mary University of London, London, UK 2 University of North Florida, Jacksonville, USA We present a semantic vector space

More information

Sudhanshu Gautam *1, Sarita Soni 2. M-Tech Computer Science, BBAU Central University, Lucknow, Uttar Pradesh, India

Sudhanshu Gautam *1, Sarita Soni 2. M-Tech Computer Science, BBAU Central University, Lucknow, Uttar Pradesh, India International Journal of Scientific Research in Computer Science, Engineering and Information Technology 2018 IJSRCSEIT Volume 3 Issue 3 ISSN : 2456-3307 Artificial Intelligence Techniques for Music Composition

More information

Bach-Prop: Modeling Bach s Harmonization Style with a Back- Propagation Network

Bach-Prop: Modeling Bach s Harmonization Style with a Back- Propagation Network Indiana Undergraduate Journal of Cognitive Science 1 (2006) 3-14 Copyright 2006 IUJCS. All rights reserved Bach-Prop: Modeling Bach s Harmonization Style with a Back- Propagation Network Rob Meyerson Cognitive

More information

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception LEARNING AUDIO SHEET MUSIC CORRESPONDENCES Matthias Dorfer Department of Computational Perception Short Introduction... I am a PhD Candidate in the Department of Computational Perception at Johannes Kepler

More information

Finding Temporal Structure in Music: Blues Improvisation with LSTM Recurrent Networks

Finding Temporal Structure in Music: Blues Improvisation with LSTM Recurrent Networks Finding Temporal Structure in Music: Blues Improvisation with LSTM Recurrent Networks Douglas Eck and Jürgen Schmidhuber IDSIA Istituto Dalle Molle di Studi sull Intelligenza Artificiale Galleria 2, 6928

More information

Robert Alexandru Dobre, Cristian Negrescu

Robert Alexandru Dobre, Cristian Negrescu ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Automatic Music Transcription Software Based on Constant Q

More information

Musical Creativity. Jukka Toivanen Introduction to Computational Creativity Dept. of Computer Science University of Helsinki

Musical Creativity. Jukka Toivanen Introduction to Computational Creativity Dept. of Computer Science University of Helsinki Musical Creativity Jukka Toivanen Introduction to Computational Creativity Dept. of Computer Science University of Helsinki Basic Terminology Melody = linear succession of musical tones that the listener

More information

Generating Music with Recurrent Neural Networks

Generating Music with Recurrent Neural Networks Generating Music with Recurrent Neural Networks 27 October 2017 Ushini Attanayake Supervised by Christian Walder Co-supervised by Henry Gardner COMP3740 Project Work in Computing The Australian National

More information

Music genre classification using a hierarchical long short term memory (LSTM) model

Music genre classification using a hierarchical long short term memory (LSTM) model Chun Pui Tang, Ka Long Chui, Ying Kin Yu, Zhiliang Zeng, Kin Hong Wong, "Music Genre classification using a hierarchical Long Short Term Memory (LSTM) model", International Workshop on Pattern Recognition

More information

MELONET I: Neural Nets for Inventing Baroque-Style Chorale Variations

MELONET I: Neural Nets for Inventing Baroque-Style Chorale Variations MELONET I: Neural Nets for Inventing Baroque-Style Chorale Variations Dominik Hornel dominik@ira.uka.de Institut fur Logik, Komplexitat und Deduktionssysteme Universitat Fridericiana Karlsruhe (TH) Am

More information

Noise (Music) Composition Using Classification Algorithms Peter Wang (pwang01) December 15, 2017

Noise (Music) Composition Using Classification Algorithms Peter Wang (pwang01) December 15, 2017 Noise (Music) Composition Using Classification Algorithms Peter Wang (pwang01) December 15, 2017 Background Abstract I attempted a solution at using machine learning to compose music given a large corpus

More information

Chord Label Personalization through Deep Learning of Integrated Harmonic Interval-based Representations

Chord Label Personalization through Deep Learning of Integrated Harmonic Interval-based Representations Chord Label Personalization through Deep Learning of Integrated Harmonic Interval-based Representations Hendrik Vincent Koops 1, W. Bas de Haas 2, Jeroen Bransen 2, and Anja Volk 1 arxiv:1706.09552v1 [cs.sd]

More information

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the

More information

Jazz Melody Generation and Recognition

Jazz Melody Generation and Recognition Jazz Melody Generation and Recognition Joseph Victor December 14, 2012 Introduction In this project, we attempt to use machine learning methods to study jazz solos. The reason we study jazz in particular

More information

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You Chris Lewis Stanford University cmslewis@stanford.edu Abstract In this project, I explore the effectiveness of the Naive Bayes Classifier

More information

CS229 Project Report Polyphonic Piano Transcription

CS229 Project Report Polyphonic Piano Transcription CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project

More information

JazzGAN: Improvising with Generative Adversarial Networks

JazzGAN: Improvising with Generative Adversarial Networks JazzGAN: Improvising with Generative Adversarial Networks Nicholas Trieu and Robert M. Keller Harvey Mudd College Claremont, California, USA ntrieu@hmc.edu, keller@cs.hmc.edu Abstract For the purpose of

More information

Composing a melody with long-short term memory (LSTM) Recurrent Neural Networks. Konstantin Lackner

Composing a melody with long-short term memory (LSTM) Recurrent Neural Networks. Konstantin Lackner Composing a melody with long-short term memory (LSTM) Recurrent Neural Networks Konstantin Lackner Bachelor s thesis Composing a melody with long-short term memory (LSTM) Recurrent Neural Networks Konstantin

More information

arxiv: v2 [cs.sd] 15 Jun 2017

arxiv: v2 [cs.sd] 15 Jun 2017 Learning and Evaluating Musical Features with Deep Autoencoders Mason Bretan Georgia Tech Atlanta, GA Sageev Oore, Douglas Eck, Larry Heck Google Research Mountain View, CA arxiv:1706.04486v2 [cs.sd] 15

More information

Computational Modelling of Harmony

Computational Modelling of Harmony Computational Modelling of Harmony Simon Dixon Centre for Digital Music, Queen Mary University of London, Mile End Rd, London E1 4NS, UK simon.dixon@elec.qmul.ac.uk http://www.elec.qmul.ac.uk/people/simond

More information

Some researchers in the computational sciences have considered music computation, including music reproduction

Some researchers in the computational sciences have considered music computation, including music reproduction INFORMS Journal on Computing Vol. 18, No. 3, Summer 2006, pp. 321 338 issn 1091-9856 eissn 1526-5528 06 1803 0321 informs doi 10.1287/ioc.1050.0131 2006 INFORMS Recurrent Neural Networks for Music Computation

More information

Deep learning for music data processing

Deep learning for music data processing Deep learning for music data processing A personal (re)view of the state-of-the-art Jordi Pons www.jordipons.me Music Technology Group, DTIC, Universitat Pompeu Fabra, Barcelona. 31st January 2017 Jordi

More information

Algorithmic Music Composition using Recurrent Neural Networking

Algorithmic Music Composition using Recurrent Neural Networking Algorithmic Music Composition using Recurrent Neural Networking Kai-Chieh Huang kaichieh@stanford.edu Dept. of Electrical Engineering Quinlan Jung quinlanj@stanford.edu Dept. of Computer Science Jennifer

More information

Advances in Algorithmic Composition

Advances in Algorithmic Composition ISSN 1000-9825 CODEN RUXUEW E-mail: jos@iscasaccn Journal of Software Vol17 No2 February 2006 pp209 215 http://wwwjosorgcn DOI: 101360/jos170209 Tel/Fax: +86-10-62562563 2006 by Journal of Software All

More information

Various Artificial Intelligence Techniques For Automated Melody Generation

Various Artificial Intelligence Techniques For Automated Melody Generation Various Artificial Intelligence Techniques For Automated Melody Generation Nikahat Kazi Computer Engineering Department, Thadomal Shahani Engineering College, Mumbai, India Shalini Bhatia Assistant Professor,

More information

Building a Better Bach with Markov Chains

Building a Better Bach with Markov Chains Building a Better Bach with Markov Chains CS701 Implementation Project, Timothy Crocker December 18, 2015 1 Abstract For my implementation project, I explored the field of algorithmic music composition

More information

Chord Classification of an Audio Signal using Artificial Neural Network

Chord Classification of an Audio Signal using Artificial Neural Network Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------

More information

CSC475 Music Information Retrieval

CSC475 Music Information Retrieval CSC475 Music Information Retrieval Symbolic Music Representations George Tzanetakis University of Victoria 2014 G. Tzanetakis 1 / 30 Table of Contents I 1 Western Common Music Notation 2 Digital Formats

More information

Singing voice synthesis based on deep neural networks

Singing voice synthesis based on deep neural networks INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Singing voice synthesis based on deep neural networks Masanari Nishimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

More information

The Sparsity of Simple Recurrent Networks in Musical Structure Learning

The Sparsity of Simple Recurrent Networks in Musical Structure Learning The Sparsity of Simple Recurrent Networks in Musical Structure Learning Kat R. Agres (kra9@cornell.edu) Department of Psychology, Cornell University, 211 Uris Hall Ithaca, NY 14853 USA Jordan E. DeLong

More information

Deep Jammer: A Music Generation Model

Deep Jammer: A Music Generation Model Deep Jammer: A Music Generation Model Justin Svegliato and Sam Witty College of Information and Computer Sciences University of Massachusetts Amherst, MA 01003, USA {jsvegliato,switty}@cs.umass.edu Abstract

More information

Bach2Bach: Generating Music Using A Deep Reinforcement Learning Approach Nikhil Kotecha Columbia University

Bach2Bach: Generating Music Using A Deep Reinforcement Learning Approach Nikhil Kotecha Columbia University Bach2Bach: Generating Music Using A Deep Reinforcement Learning Approach Nikhil Kotecha Columbia University Abstract A model of music needs to have the ability to recall past details and have a clear,

More information

arxiv: v1 [cs.ai] 2 Mar 2017

arxiv: v1 [cs.ai] 2 Mar 2017 Sampling Variations of Lead Sheets arxiv:1703.00760v1 [cs.ai] 2 Mar 2017 Pierre Roy, Alexandre Papadopoulos, François Pachet Sony CSL, Paris roypie@gmail.com, pachetcsl@gmail.com, alexandre.papadopoulos@lip6.fr

More information

Automated sound generation based on image colour spectrum with using the recurrent neural network

Automated sound generation based on image colour spectrum with using the recurrent neural network Automated sound generation based on image colour spectrum with using the recurrent neural network N A Nikitin 1, V L Rozaliev 1, Yu A Orlova 1 and A V Alekseev 1 1 Volgograd State Technical University,

More information

PLANE TESSELATION WITH MUSICAL-SCALE TILES AND BIDIMENSIONAL AUTOMATIC COMPOSITION

PLANE TESSELATION WITH MUSICAL-SCALE TILES AND BIDIMENSIONAL AUTOMATIC COMPOSITION PLANE TESSELATION WITH MUSICAL-SCALE TILES AND BIDIMENSIONAL AUTOMATIC COMPOSITION ABSTRACT We present a method for arranging the notes of certain musical scales (pentatonic, heptatonic, Blues Minor and

More information

Singer Traits Identification using Deep Neural Network

Singer Traits Identification using Deep Neural Network Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic

More information

BachBot: Automatic composition in the style of Bach chorales

BachBot: Automatic composition in the style of Bach chorales BachBot: Automatic composition in the style of Bach chorales Developing, analyzing, and evaluating a deep LSTM model for musical style Feynman Liang Department of Engineering University of Cambridge M.Phil

More information

OPTICAL MUSIC RECOGNITION WITH CONVOLUTIONAL SEQUENCE-TO-SEQUENCE MODELS

OPTICAL MUSIC RECOGNITION WITH CONVOLUTIONAL SEQUENCE-TO-SEQUENCE MODELS OPTICAL MUSIC RECOGNITION WITH CONVOLUTIONAL SEQUENCE-TO-SEQUENCE MODELS First Author Affiliation1 author1@ismir.edu Second Author Retain these fake authors in submission to preserve the formatting Third

More information

Real-valued parametric conditioning of an RNN for interactive sound synthesis

Real-valued parametric conditioning of an RNN for interactive sound synthesis Real-valued parametric conditioning of an RNN for interactive sound synthesis Lonce Wyse Communications and New Media Department National University of Singapore Singapore lonce.acad@zwhome.org Abstract

More information

SentiMozart: Music Generation based on Emotions

SentiMozart: Music Generation based on Emotions SentiMozart: Music Generation based on Emotions Rishi Madhok 1,, Shivali Goel 2, and Shweta Garg 1, 1 Department of Computer Science and Engineering, Delhi Technological University, New Delhi, India 2

More information

A Transfer Learning Based Feature Extractor for Polyphonic Sound Event Detection Using Connectionist Temporal Classification

A Transfer Learning Based Feature Extractor for Polyphonic Sound Event Detection Using Connectionist Temporal Classification INTERSPEECH 17 August, 17, Stockholm, Sweden A Transfer Learning Based Feature Extractor for Polyphonic Sound Event Detection Using Connectionist Temporal Classification Yun Wang and Florian Metze Language

More information

Automatic Piano Music Transcription

Automatic Piano Music Transcription Automatic Piano Music Transcription Jianyu Fan Qiuhan Wang Xin Li Jianyu.Fan.Gr@dartmouth.edu Qiuhan.Wang.Gr@dartmouth.edu Xi.Li.Gr@dartmouth.edu 1. Introduction Writing down the score while listening

More information

arxiv: v1 [cs.sd] 20 Nov 2018

arxiv: v1 [cs.sd] 20 Nov 2018 COUPLED RECURRENT MODELS FOR POLYPHONIC MUSIC COMPOSITION John Thickstun 1, Zaid Harchaoui 2 & Dean P. Foster 3 & Sham M. Kakade 1,2 1 Allen School of Computer Science and Engineering, University of Washington,

More information

Towards End-to-End Raw Audio Music Synthesis

Towards End-to-End Raw Audio Music Synthesis To be published in: Proceedings of the 27th Conference on Artificial Neural Networks (ICANN), Rhodes, Greece, 2018. (Author s Preprint) Towards End-to-End Raw Audio Music Synthesis Manfred Eppe, Tayfun

More information

Music Segmentation Using Markov Chain Methods

Music Segmentation Using Markov Chain Methods Music Segmentation Using Markov Chain Methods Paul Finkelstein March 8, 2011 Abstract This paper will present just how far the use of Markov Chains has spread in the 21 st century. We will explain some

More information

Shimon the Robot Film Composer and DeepScore

Shimon the Robot Film Composer and DeepScore Shimon the Robot Film Composer and DeepScore Richard Savery and Gil Weinberg Georgia Institute of Technology {rsavery3, gilw} @gatech.edu Abstract. Composing for a film requires developing an understanding

More information

Learning to Create Jazz Melodies Using Deep Belief Nets

Learning to Create Jazz Melodies Using Deep Belief Nets Claremont Colleges Scholarship @ Claremont All HMC Faculty Publications and Research HMC Faculty Scholarship 1-1-2010 Learning to Create Jazz Melodies Using Deep Belief Nets Greg Bickerman '10 Harvey Mudd

More information

arxiv: v1 [cs.cv] 16 Jul 2017

arxiv: v1 [cs.cv] 16 Jul 2017 OPTICAL MUSIC RECOGNITION WITH CONVOLUTIONAL SEQUENCE-TO-SEQUENCE MODELS Eelco van der Wel University of Amsterdam eelcovdw@gmail.com Karen Ullrich University of Amsterdam karen.ullrich@uva.nl arxiv:1707.04877v1

More information

Hidden Markov Model based dance recognition

Hidden Markov Model based dance recognition Hidden Markov Model based dance recognition Dragutin Hrenek, Nenad Mikša, Robert Perica, Pavle Prentašić and Boris Trubić University of Zagreb, Faculty of Electrical Engineering and Computing Unska 3,

More information

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t MPEG-7 FOR CONTENT-BASED MUSIC PROCESSING Λ Emilia GÓMEZ, Fabien GOUYON, Perfecto HERRERA and Xavier AMATRIAIN Music Technology Group, Universitat Pompeu Fabra, Barcelona, SPAIN http://www.iua.upf.es/mtg

More information

Sequence generation and classification with VAEs and RNNs

Sequence generation and classification with VAEs and RNNs Jay Hennig 1 * Akash Umakantha 1 * Ryan Williamson 1 * 1. Introduction Variational autoencoders (VAEs) (Kingma & Welling, 2013) are a popular approach for performing unsupervised learning that can also

More information

Predicting Aesthetic Radar Map Using a Hierarchical Multi-task Network

Predicting Aesthetic Radar Map Using a Hierarchical Multi-task Network Predicting Aesthetic Radar Map Using a Hierarchical Multi-task Network Xin Jin 1,2,LeWu 1, Xinghui Zhou 1, Geng Zhao 1, Xiaokun Zhang 1, Xiaodong Li 1, and Shiming Ge 3(B) 1 Department of Cyber Security,

More information

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES PACS: 43.60.Lq Hacihabiboglu, Huseyin 1,2 ; Canagarajah C. Nishan 2 1 Sonic Arts Research Centre (SARC) School of Computer Science Queen s University

More information

Modelling Symbolic Music: Beyond the Piano Roll

Modelling Symbolic Music: Beyond the Piano Roll JMLR: Workshop and Conference Proceedings 63:174 189, 2016 ACML 2016 Modelling Symbolic Music: Beyond the Piano Roll Christian Walder Data61 at CSIRO, Australia. christian.walder@data61.csiro.au Editors:

More information

MUSIC scores are the main medium for transmitting music. In the past, the scores started being handwritten, later they

MUSIC scores are the main medium for transmitting music. In the past, the scores started being handwritten, later they MASTER THESIS DISSERTATION, MASTER IN COMPUTER VISION, SEPTEMBER 2017 1 Optical Music Recognition by Long Short-Term Memory Recurrent Neural Networks Arnau Baró-Mas Abstract Optical Music Recognition is

More information

Los Angeles Valley College MUS 200: INTRO TO MUSIC THEORY

Los Angeles Valley College MUS 200: INTRO TO MUSIC THEORY Los Angeles Valley College MUS 200: INTRO TO MUSIC THEORY FALL 2016 Tuesday/Thursday, 8:15am - 10:40am, M112 Timothy Herscovitch, professor E-mail and Phone: herscota@gmail.com / (818) 947-2346 (office)

More information

Music Generation from MIDI datasets

Music Generation from MIDI datasets Music Generation from MIDI datasets Moritz Hilscher, Novin Shahroudi 2 Institute of Computer Science, University of Tartu moritz.hilscher@student.hpi.de, 2 novin@ut.ee Abstract. Many approaches are being

More information

Music Representations

Music Representations Lecture Music Processing Music Representations Meinard Müller International Audio Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de Book: Fundamentals of Music Processing Meinard Müller Fundamentals

More information

Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment

Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment Gus G. Xia Dartmouth College Neukom Institute Hanover, NH, USA gxia@dartmouth.edu Roger B. Dannenberg Carnegie

More information

The Design of Efficient Viterbi Decoder and Realization by FPGA

The Design of Efficient Viterbi Decoder and Realization by FPGA Modern Applied Science; Vol. 6, No. 11; 212 ISSN 1913-1844 E-ISSN 1913-1852 Published by Canadian Center of Science and Education The Design of Efficient Viterbi Decoder and Realization by FPGA Liu Yanyan

More information

Generating Music from Text: Mapping Embeddings to a VAE s Latent Space

Generating Music from Text: Mapping Embeddings to a VAE s Latent Space MSc Artificial Intelligence Master Thesis Generating Music from Text: Mapping Embeddings to a VAE s Latent Space by Roderick van der Weerdt 10680195 August 15, 2018 36 EC January 2018 - August 2018 Supervisor:

More information

Neural Network for Music Instrument Identi cation

Neural Network for Music Instrument Identi cation Neural Network for Music Instrument Identi cation Zhiwen Zhang(MSE), Hanze Tu(CCRMA), Yuan Li(CCRMA) SUN ID: zhiwen, hanze, yuanli92 Abstract - In the context of music, instrument identi cation would contribute

More information

Non-chord Tone Identification

Non-chord Tone Identification Non-chord Tone Identification Yaolong Ju Centre for Interdisciplinary Research in Music Media and Technology (CIRMMT) Schulich School of Music McGill University SIMSSA XII Workshop 2017 Aug. 7 th, 2017

More information

A Discriminative Approach to Topic-based Citation Recommendation

A Discriminative Approach to Topic-based Citation Recommendation A Discriminative Approach to Topic-based Citation Recommendation Jie Tang and Jing Zhang Department of Computer Science and Technology, Tsinghua University, Beijing, 100084. China jietang@tsinghua.edu.cn,zhangjing@keg.cs.tsinghua.edu.cn

More information

arxiv: v1 [cs.sd] 29 Apr 2016

arxiv: v1 [cs.sd] 29 Apr 2016 Music transcription modelling and composition using deep learning Bob L. Sturm 1, João Felipe Santos 2, Oded Ben-Tal 3 and Iryna Korshunova 4 1 Centre for Digital Music, Queen Mary University of London

More information

Composer Style Attribution

Composer Style Attribution Composer Style Attribution Jacqueline Speiser, Vishesh Gupta Introduction Josquin des Prez (1450 1521) is one of the most famous composers of the Renaissance. Despite his fame, there exists a significant

More information

arxiv: v1 [cs.sd] 9 Dec 2017

arxiv: v1 [cs.sd] 9 Dec 2017 Music Generation by Deep Learning Challenges and Directions Jean-Pierre Briot François Pachet Sorbonne Universités, UPMC Univ Paris 06, CNRS, LIP6, Paris, France Jean-Pierre.Briot@lip6.fr Spotify Creator

More information

Artificial Intelligence Approaches to Music Composition

Artificial Intelligence Approaches to Music Composition Artificial Intelligence Approaches to Music Composition Richard Fox and Adil Khan Department of Computer Science Northern Kentucky University, Highland Heights, KY 41099 Abstract Artificial Intelligence

More information

Grade Six. MyMusicTheory.com. Composition Complete Course, Exercises & Answers PREVIEW. (ABRSM Syllabus) BY VICTORIA WILLIAMS BA MUSIC

Grade Six. MyMusicTheory.com. Composition Complete Course, Exercises & Answers PREVIEW. (ABRSM Syllabus) BY VICTORIA WILLIAMS BA MUSIC MyMusicTheory.com Grade Six Composition Complete Course, Exercises & Answers PREVIEW (ABRSM Syllabus) BY VICTORIA WILLIAMS BA MUSIC www.mymusictheory.com Published: 8th March 2015 1 This is a preview document

More information

Evaluating Melodic Encodings for Use in Cover Song Identification

Evaluating Melodic Encodings for Use in Cover Song Identification Evaluating Melodic Encodings for Use in Cover Song Identification David D. Wickland wickland@uoguelph.ca David A. Calvert dcalvert@uoguelph.ca James Harley jharley@uoguelph.ca ABSTRACT Cover song identification

More information

Appendix A Types of Recorded Chords

Appendix A Types of Recorded Chords Appendix A Types of Recorded Chords In this appendix, detailed lists of the types of recorded chords are presented. These lists include: The conventional name of the chord [13, 15]. The intervals between

More information