Specifying Features for Classical and Non-Classical Melody Evaluation

Similar documents
jsymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada

jsymbolic 2: New Developments and Research Opportunities

MUSIC THEORY CURRICULUM STANDARDS GRADES Students will sing, alone and with others, a varied repertoire of music.

CSC475 Music Information Retrieval

Outline. Why do we classify? Audio Classification

MHSIB.5 Composing and arranging music within specified guidelines a. Creates music incorporating expressive elements.

ILLINOIS LICENSURE TESTING SYSTEM

Exploring Melodic Features for the Classification and Retrieval of Traditional Music in the Context of Cultural Source

Analysis and Clustering of Musical Compositions using Melody-based Features

Computational Modelling of Harmony

LESSON 1 PITCH NOTATION AND INTERVALS

Music. Last Updated: May 28, 2015, 11:49 am NORTH CAROLINA ESSENTIAL STANDARDS

Sudhanshu Gautam *1, Sarita Soni 2. M-Tech Computer Science, BBAU Central University, Lucknow, Uttar Pradesh, India

A Basis for Characterizing Musical Genres

PLANE TESSELATION WITH MUSICAL-SCALE TILES AND BIDIMENSIONAL AUTOMATIC COMPOSITION

Woodlynne School District Curriculum Guide. General Music Grades 3-4

Composer Style Attribution

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

II. Prerequisites: Ability to play a band instrument, access to a working instrument

Music Genre Classification and Variance Comparison on Number of Genres

Grade Level 5-12 Subject Area: Vocal and Instrumental Music

Creating a Feature Vector to Identify Similarity between MIDI Files

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Topics in Computer Music Instrument Identification. Ioanna Karydi

The Human Features of Music.

Musical Harmonization with Constraints: A Survey. Overview. Computers and Music. Tonal Music

Curriculum Mapping Piano and Electronic Keyboard (L) Semester class (18 weeks)

PERFORMING ARTS. Head of Music: Cinzia Cursaro. Year 7 MUSIC Core Component 1 Term

A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES

Musical Creativity. Jukka Toivanen Introduction to Computational Creativity Dept. of Computer Science University of Helsinki

ILLINOIS LICENSURE TESTING SYSTEM

Articulation Clarity and distinct rendition in musical performance.

HST 725 Music Perception & Cognition Assignment #1 =================================================================

Music Theory. Fine Arts Curriculum Framework. Revised 2008

Curriculum Framework for Performing Arts

Grade 4 Music Curriculum Maps

TABLE OF CONTENTS CHAPTER 1 PREREQUISITES FOR WRITING AN ARRANGEMENT... 1

Power Standards and Benchmarks Orchestra 4-12

Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals

NCEA Level 2 Music (91275) 2012 page 1 of 6. Assessment Schedule 2012 Music: Demonstrate aural understanding through written representation (91275)

Doctor of Philosophy

CHOIR Grade 6. Benchmark 4: Students sing music written in two and three parts.

From quantitative empirï to musical performology: Experience in performance measurements and analyses

NEW YORK STATE TEACHER CERTIFICATION EXAMINATIONS

Lesson Week: August 17-19, 2016 Grade Level: 11 th & 12 th Subject: Advanced Placement Music Theory Prepared by: Aaron Williams Overview & Purpose:

Music (MUS) Courses. Music (MUS) 1

MUSI-6201 Computational Music Analysis

Automatic Music Clustering using Audio Attributes

Connecticut State Department of Education Music Standards Middle School Grades 6-8

EASTERN ARIZONA COLLEGE Elementary Theory

Murrieta Valley Unified School District High School Course Outline February 2006

Paper Reference. Paper Reference(s) 1426/03 Edexcel GCSE Music Paper 3 Listening and Appraising. Friday 18 May 2007 Afternoon Time: 1 hour 30 minutes

Copyright 2009 Pearson Education, Inc. or its affiliate(s). All rights reserved. NES, the NES logo, Pearson, the Pearson logo, and National

Music Curriculum Glossary

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University

Algorithmic Music Composition

Influence of timbre, presence/absence of tonal hierarchy and musical training on the perception of musical tension and relaxation schemas

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t

Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment

BRICK TOWNSHIP PUBLIC SCHOOLS (SUBJECT) CURRICULUM

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm

CS229 Project Report Polyphonic Piano Transcription

A Real-Time Genetic Algorithm in Human-Robot Musical Improvisation

Visual Arts, Music, Dance, and Theater Personal Curriculum

A Case Based Approach to the Generation of Musical Expression

Music, Grade 9, Open (AMU1O)

Topic 10. Multi-pitch Analysis

2014 Music Style and Composition GA 3: Aural and written examination

A CHROMA-BASED SALIENCE FUNCTION FOR MELODY AND BASS LINE ESTIMATION FROM MUSIC AUDIO SIGNALS

Automatic Piano Music Transcription

Feature-Based Analysis of Haydn String Quartets

K-12 Performing Arts - Music Standards Lincoln Community School Sources: ArtsEdge - National Standards for Arts Education

Semi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis

Robert Alexandru Dobre, Cristian Negrescu

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

CONCERT ORCHESTRA AND SYMPHONIC ORCHESTRA

A probabilistic approach to determining bass voice leading in melodic harmonisation

Exploring the Design Space of Symbolic Music Genre Classification Using Data Mining Techniques Ortiz-Arroyo, Daniel; Kofod, Christian

Detecting Musical Key with Supervised Learning

FINE ARTS Institutional (ILO), Program (PLO), and Course (SLO) Alignment

AP Music Theory Syllabus

Choir Scope and Sequence Grade 6-12

Perceptual Evaluation of Automatically Extracted Musical Motives

MUSIC (MUS) Music (MUS) 1

Curriculum Standard One: The student will listen to and analyze music critically, using the vocabulary and language of music.

2016 HSC Music 1 Aural Skills Marking Guidelines Written Examination

MUSIC (MU) Music (MU) 1

Eighth Grade Music Curriculum Guide Iredell-Statesville Schools

PHYSICS OF MUSIC. 1.) Charles Taylor, Exploring Music (Music Library ML3805 T )

Music 1. the aesthetic experience. Students are required to attend live concerts on and off-campus.

Music. Program Level Student Learning Outcomes

PERFORMING ARTS Curriculum Framework K - 12

Curriculum Standard One: The student will listen to and analyze music critically, using vocabulary and language of music.

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

AP Music Theory Syllabus

Grade One General Music

Automatic Composition of Music with Methods of Computational Intelligence

Curriculum Standard One: The student will listen to and analyze music critically, using the vocabulary and language of music.

Music Complexity Descriptors. Matt Stabile June 6 th, 2008

Transcription:

Specifying Features for Classical and Non-Classical Melody Evaluation Andrei D. Coronel Ateneo de Manila University acoronel@ateneo.edu Ariel A. Maguyon Ateneo de Manila University amaguyon@ateneo.edu Proceso L. Fernandez Ateneo de Manila University pfernandez@ateneo.edu ABSTRACT Algorithmic Composition for music is a progressive field of study. The success of an automated music-generating algorithm depends heavily on the fitness function that is used to score the generated music. This fitness function is generally based on music features that a given algorithm is programmed to measure. This study explores the features that are important for melody generation by investigating those features that can separate classical from nonclassical music in the context of melody, and that can help distinguish between 2 specific subgenres of both classical and non-classical music Chopin vs. Bach compositions, as well as jazz vs. contemporary compositions. The jsymbolic tool was used to collect 160 standard features from 400 music files. Using C4.5 to select significant features, and then running Naïve-Bayes and SVM classifiers, the study was able to determine melodic features that may later be used for formulating fitness function for automated music generation. Categories and Subject Descriptors H.5.5 [Computer Science and Music]: Sound and Music Computing methodologies and techniques. General Terms Algorithms, Measurement, Performance, Experimentation, Keywords Computer Music, Melodic Evaluation, Music Features, Melodic Features, Music Genre Classification 1. INTRODUCTION Computer generated music is a progressive area of interest. Interdisciplinary works have been developing various means to create musical compositions via algorithmic programming and software applications. Computer generated music can suggest either tool-based computer-aided compositions or algorithmic composition. The former requires users to employ nonalgorithmic interactive software to be able to create music from a composer s idea. The Algorithmic Composition (AC) can involve either an application of heuristic principles, automated learning techniques, or evolutionary programming. This study refers to AC methods rather than the former tool-based methods. The AC methods generally fall under one of two types: rule-based or evolutionary. The quality of the output of AC, especially in evolutionary methods such as Genetic Algorithms, relies heavily on a suitable fitness function. This function is used to score the computer-generated music by providing a quantifiable measure to what is originally a qualitative measure of aesthetic value. It is, thus, this function that determines whether or not the musical output is within an acceptable aesthetic quality. To arrive at a suitable fitness function for scoring computergenerated music, it is critical to identify the important music features to be measured for such a fitness function. One of the challenges, however, is that different music genres intuitively have different fitness functions. This study investigates various music features that can be used in the computer generation of classical music. It is shown that a set of 9 features is sufficient in accurately distinguishing between classical and non-classical music. These 9 features, which can be easily captured from MIDI files using the JSymbolic tool, can, thus, be explored in the fitness functions for AC methods that generate classical music. 2. REVIEW OF RELATED LITERATURE Many software applications and generative music systems available for Algorithmic Composition (AC), such as IMPROVISOR, Tune Smithy and Bloom can be found on the Internet. The main methods used by these applications generally fall under one of two categories: rule-based or evolutionary. Examples of rule-based AC include stochastic binary subdivision, auditals, key phrase animation, fractal interpolation, and expert novice pickers. These techniques involve the application of rhyhtmic generation to melodies, grammar-based melodic generation, computer graphics algorithms applied to music, and the use of knowledge-bases, respectively [6]. Evolutionary AC methods are those that basically run the generate-evaluate-repeat loop for music generation. Two of the most popular heuristics for evolutionary AC are Genetic Algorithm and Genetic Programming [10]. These methods rely on a fitness function for evaluating the generated music. Various fitness functions have been previously explored and applied to different AC methods. There is no recognized gold standard yet, as even modern studies in melodic extension may still partially employ human evaluators [4], or fully utilize them to score improvisations [5]. Included in the list of these scoring mechanisms are human critics, rule-based critics, learning-based critics, and global statistics. Since there are constraints when human critics are used, e.g., fatigue based on the repetitiveness of assessing the fitness of pieces of music, the search for more automatic evaluation techniques is a continuing study [11]. A prerequisite for the development of fitness function for automatic music evaluation is the extraction of features from a music piece. These feature values become the parameters to the fitness functions being developed. Research on feature extraction straight from audio or acoustic signals has already been done [9], and the fitness function that was developed for this took into

consideration music features such as spectral variation, count of sound, frequency strength, amplitude frequency, to name several important ones. The study focused on pairing each extracted music feature with appropriate weights, where the weight set is genre specific. This type of music analysis, however, does not isolate melody, since the features were extracted from audio signals from compressed music. Especially for evolutionary methods, melodic analysis is very relevant since melody is the factor being evolved in every turn of the iterative program. A study made by Towsey et al enumerated, categorized, and analyzed features based on melody by applying global statistics to a dataset of MIDI files. The study was able to identify 21 melodic features that are useful for melodic analysis. These features include pitch variety, dissonance of intervals, and contour direction, to name several. PCA analysis and subsequent clustering procedures were successful in identifying the strength of influence of the features to the potential fitness rating of melodies [11]. A more recent study by Alan Freitas et al enumerated and described important features for melodic evaluation, taking previous studies in consideration [2]. What these studies have not yet addressed, however, is the identification of what specific combination of features are optimal for melodic analysis involving particular genres, i.e., what features are more relevant for classical music analysis vs. non-classical or contemporary music. In this study, we extend the work of Towsey et al by verifying if the 21 melodic features that they have identified can optimally differentiate classical from non-classical genres, and between 2 specified subgenres for each of these. Identification of the important features is a step towards the development a better fitness function for the automated evaluation of music based on melodic features. The results of this study will be the foundation for the creation of fitness functions for evaluating melodies generated by evolutionary algorithms. In this context, this study is contributory to automated music composition since the latter requires a quantitative gauge of melodic scoring to be able to weed out melodies of low fitness and retain those of high fitness for subsequent iterative evolution (i.e. Genetic Algorithms). The construction of these fitness functions begin with determining the specific key features that define the quality of genre-specific melodic strings in the context of quantitative feature values. 3. METHODOLOGY The goal of the study is to specify which selected features or feature sets are optimal for effectively classifying music according to genre and subgenre in the context of melodic analysis. There have been numerous works on music evaluation where the music source is either a dataset of audio signals (compressed music files) or MIDI. This study will be using the latter, since melodic analysis is the focus of the study rather than the actual acoustic structure. Given that the focus of the study is melodic analysis in music represented by MIDI format files, the features extracted will be melodic features and not audio signal features. It is not the goal of this study to develop and implement a new algorithmic composition technique, however it is an essential step towards the development of a fitness function that may potentially evaluate melodies according to its quality in the context of genre matching. Collection of 400 MIDI Files: 100 Bach 100 Chopin 100 Jazz 100 Contemporary Data Set 0: 400 instances x 160 music features Data Set 1: 400 instances x 54 melody features C4.5 Feature Selection Data Set 1A: 400 x 13 Data Set 1B: 200 x 5 Data Set 1C: 200 x 3 Data Set 1D: 400 x 11 jsymbolic Feature Extraction Melodic Feature Analysis Comparative Analysis: Towsey Features Mapping Naïve Bayes and SVM Classification C4.5 features vs Towsey features on Data Set 2A: 400 x 13 Data Set 2B: 200 x 13 Data Set 2C: 200 x 13 Data Set 2D: 400 x 13 A. Classical vs Non-Classical B. Bach vs Chopin C. Contemporary vs Jazz D. Bach vs Chopin vs Contemporary vs Jazz Figure 1: Methodology in this research Data Set 2: 400 instances x 13 melody features Data Preparation The methodology for this study involved several steps (refer also to Fig. 1). 1. Acquire a music dataset (in MIDI format) with a balanced distribution of 4 subgenres: Bach, Chopin, Jazz (various artists), and Contemporary (Beatles). Bach and Chopin represent the classical group, while the Jazz and Contemporary melodies represent the non-classical group. The Beatles were chosen to represent contemporary music as majority of their melodies stay within whole major and minor non-augmented chords,

which is a harmonic trait of common contemporary music still applicable today. 2. Extract features from the music dataset (via jsymbolic). The extraction involves all 160 features that are extractable from each of the MIDI files. 3. Review the definition of the jsymbolic features in order to identify the features that are related to melody. Based on feature definitions, only 54 out of the 160 features may actually be applied to melodic analysis. Hence the extraneous 106 features were removed from the dataset. 4. Apply C4.5 decision tree algorithm to 4 datasets to determine which of the 54 melodic features are significant for specific classification. The following describes the 4 datasets: a. Dataset of 400 MIDI files, each labeled as either classical or non-classical. b. Dataset of 200 MIDI files, each labeled as either Bach or Chopin c. Dataset of 200 MIDI files, each labeled as either Beatles (contemporary) or Jazz. d. Dataset of 400 MIDI files each labeled as either Bach, Chopin, Beatles or Jazz. Note that the four dataset described above are subsets of the earlier dataset composed of 400 MIDI files composed of 160 features (please refer to Steps 1-2). 5. Create 4 new music datasets with reduced number of features, as recomm by the C4.5 results of the previous step. 6. Map the Towsey melodic features to the jsymbolicextractable features. 7. Create an additional 4 music datasets similar to Step 5 but using the Towsey mapped features of Step 6. 8. Run Naïve Bayes and SVM on the 8 datasets from Steps 4 & 6, and estimate the accuracy using 10-fold crossvalidation method. Accuracy here is measured by computing the number of correct classification divided by the total number of instances classified. 9. Determine the features involved in the best results for each specific classification challenge. identify instrument type, and subsequently removing them from the list of 160. What remains then would be melody-centric features. The table below shows which features are involved in melodic analysis and which are otherwise. Involved in Melodic Analysis Amount of Arpeggiation Average Melodic Interval Average Note Duration Average Note to Note Dynamics Change Average of Glissandos Average Time Between Attacks Changes of Meter Chromatic Motion Compound or Simple Meter Direction of Motion Distance Between Most Common Melodic Intervals Dominant Spread Duration of Melodic Arcs Glissando Prevalence Pitch Classes Pitches Maximum Note Duration Melodic Fifths Melodic Intervals in Lowest Line Not involved in Melodic Analysis Acoustic Guitar Fraction Average Number of independent Voices Average Time Between Attacks for Each Voice Average Variability of Time Between Attacks for each Voice Brass Fraction Combined Strength of Two Strongest Rhythmic Pulses Electric Guitar Fraction Electric Instrument Fraction Harmonicity of Two Strongest Rhythmic Pulses Importance of Bass Register Importance of High Register Importance of Loudest Voice Importance of Middle Register Initial Tempo Maximum Number of Independent Voices Number of Moderate Pulses Number of Pitched Instruments Number of Relatively Strong pulses Number of Strong Pullses 4. RESULTS AND ANALYSIS After subjecting a complete 400 MIDI dataset to the C4.5 decision tree algorithm, it was noted that the key features used in classifying music into separate genres included features that are not exclusive for melodic analysis, such as percussion-based features and an analysis of MIDI layers/voices as well as specific instrument fractions. These features are not useful in melodic analysis, especially since subsequent work after this study will involve the evaluation of evolved melodies, which does not include an analysis of MIDI instrument used, nor layering. Selecting features exclusive for melodic analysis was performed in view of this. Based on feature definitions, 54 of the available 160 music features were determined to be significant for melodic analysis. The 54 features were identified to be metrics for melody by determining features that focus on polyphony (inclusion of nonmelodic voices), non-pitch related features, and features that Melodic Octaves Melodic Thirds Melodic Tritones Minimum Note Duration Most Common Melodic Interval Most Common Melodic Interval Prevalence Most Common Pitch Class Most Common Pitch Class Prevalence Most Common Pitch Number of Unpitched instruments Orchestral Strings Fraction Percussion Prevalence Polyrhythms of Highest Line Relative of Highest Line Relative of Loudest Voice Saxophone Fraction Second Strongest Rhythm Pulse

Most Common Pitch Prevalence Number of Common Melodic Intervals Number of Common Pitches Overall Dynamic Pitch Class Variety Primary Register Quality Quintuple Meter Relative Strength of Most Common Interavals Relative Strength of Top Ptich Classes Relative Strenth of Top Pitches Rhythmic Variability Size of Melodc Arcs Staccato Incidence Stepwise Motion Stroing Tonal Centers Triple Meter Variability of Note Duration Variability of Time Between Attacks Variation Dynamics Vibrato Prevalence Strength of Second Strongest Rhythm Pulse Strength Ratio of Two Strongest Rhythmic Pulses String Ensemble Fraction String Keyboard Fraction Strongest Rhythmic Pulse Variability of Note Prevalence of Pitched Instruments Variability of Note Prevalence of Unpitched Instruments Variability of Number of Independent Voices Variation of Dynamics In Each Voice Violin Fraction Voice Equality - Dynamics Voice Equality - Melodic Leaps Voice Equality - Note Duration Voice Equality - Number of Notes Voice Equality - Voice Separation Woodwinds Fraction Table 1: Features involved with melody and features that are not based on melody Applying C4.5 decision tree algorithm on the datasets considering all 54 melodic features (first column of Table 1) resulted in the identification of what features are used in specific classification challenges. The results are as follows: Classification Challenge Classical vs. Non-Classical Relevant Features (Listed according to importance in C4.5 Decision Tree) Variation of Dynamics Amount of Arpeggiation Bach vs. Chopin Contemporary vs. Jazz Bach vs. Chopin vs. Contemporary vs. Jazz Minimum Note Duration Dominant Spread Primary Register Pitch Class variety Variation of Dynamics Chromatic Motion Triple Meter Most Common Pitch Class Chromatic Motion Size of Melodic Arcs Melodic Tritones Pitches Staccato Incidence Rhythmic Variability Number of Common Pitches Average Note Duration Rhythmic Variability Stepwise Motion Triple Meter Melodic Octaves Staccato Incidence Size of Melodic Arcs Pitch Classes Table 2: Features found to be significant based on C4.5 Decision Tree Algorithm for each classification challenge The features sets specific for each classification challenge form the potentially recomm features to be used whenever the goal is to measure the fitness of an evolved melody into either classical, non-classical, Contemporary or Jazz. The study by Towsey et al suggested the use 21 identified features. Based on careful analysis of features and feature definitions, it has been found that those suggested features correspond to 13 of of the 54 jsymbolic extractable music features. The following table shows the mapping between features extractable by jsymbolic and the actual features suggested by Towsey et al for melodic analysis: jsymbolic Extractable Feature Recomm Features for Towsey et al

Rhythmic Variability Direction of Motion Stepwise Motion Quality Melodic Octaves Melodic Thirds Melodic Tritones Melodic Fifths Rhythmic Rhythmic Variety, Rest Density, syncopaion, (Repeated Rhythmic values) Contour Direction, Contour Stability Climax Strength, (Repeated Pitch), Movement by step, Leap Returns Contour Direction, Contour Stability Key-centered, non-scal notes, dissonant intervals Table 3: Mapping between jsymbolic extractable features and the recommendations by Towsey, et al. After acquiring this information, a comparison of the classification accuracy in datasets involving the C4.5-suggested feature sets and the Toswey-based features was performed. Both Naïve-Bayes and SVM were used to check the performance of the classification with varying features sets. To estimate the actual performance of each of these classifiers to future data, 10-fold cross validation was performed. The tests showed the following accuracy results: Compari son Test 1. Classical vs. Nonclassical 2. Bach vs. Chopin 3. Contemp orary vs. Jazz 4. Bach vs. Chopin vs. Contemp orary vs. Naïve- Bayes Features by C4.5) Naïve- Bayes features mapped from Towsey at al SVM Features by C4.5 SVM Features mapped from Towsey et al) 96.5% 91.5% 97.5% 95.25% 75.5% 91.5% 77% 92.5% 97% 99% 98% 99% 94.5% 90.25% 95.5% 92% Jazz Table 4: Comparison among the classification accuracy results between the C4.5 recomm features and the features suggested by Towsey et al. The results presented in Table 4 indicate that the choice of the classification algorithm (Naïve Bayes and SVM) does not affect the preferred feature set for classification. That is, for each of the 4 tests performed, the features on which Naïve Bayes performed better were also those features where SVM registered a higher accuracy. The results have shown that, after a selection of features exclusive for melodic analysis, the features set constructed out of the results of the C4.5 Decision Tree algorithm (Table 2) produce a more accurate classification when compared to features that represent the suggested set by Towsey et al in classification challenges that involve more abstract genre categorization rather than more specific (subgenre) classification, in which the feature set that approximates Towsey et al s study proved to perform better (Table 4) However, the feature sets based on the C4.5 results was able to classify subgenres and outperform the Towset-based feature set when the MIDI files to be classified have a larger spectrum of differentiation (i.e. a dataset with 4 different subgenres, belong to both classical and non-classical groups). Overall, the results indicate the C4.5-based features are more appropriate than the Towsey features for classification involving a wide variety of elements (belonging to different music genres), while the Towsey features are still preferred for classifying within a genre. 5. CONCLUSION This study has mapped features suggested by Towsey et al for melodic analysis to the features that can be extracted by jsymbolic to produce a Towsey-based feature set for music genre classification. The classification accuracy of this Towsey-based feature set was compared with features sets constructed with features based on the results of C4.5 decision tree algorithm applied to a variety of classification challenges. All features involved in these tests were exclusive for melodic analysis (i.e. features that analyze instrument fraction were removed). Results have shown that C4.5-based feature sets vary for different classification challenges. This indicates that the features for fitness evaluation in future studies would have to consider various feature sets for computer generation of music from various genres. We have also shown that the Towsey-based melodic features may not always be the preferred features for classification, and later on for fitness evaluation. Specifically, the features that the C4.5 recommends appear to be better for classifying a wider variety of music samples. The Towsey-based feature set, however, still appears to be best used in a more specific subgenre classification of music files for datasets that contain elements from under the same major genre. These results prove to be useful as these feature sets will be used alongside distance measure to evaluate the quality of an evolved melody (i.e. output of an evolutionary algorithm) by analyzing the it against target melodies, in the continuing effort to develop an automated fitness function for melodic evaluation.

6. FURTHER STUDY It would be interesting to investigate the classification accuracy if the dataset size was increased significantly (e.g., several thousand records). The use of other classifiers, and the collection of the accuracy results from these will also allow better comparison between the C4.5 and Towsey features. Still, another possible direction can apply a different technique for feature selection in order to possibly identify different feature sets for different classification tasks. 7. REFERENCES [1] M. Alfonseca, M. Cebrian, and A. Ortega, A Fitness Function for Computer Generated Music using Genetic Algorithms, WSEAS Transactions on Information Science and Applications, Vol. 3:3, p.518-525, Mar. 2006. [2] A. Freitas,F. Guimaraes and R. Barbosa,"Ideas in Automatic Evaluation Methods for Melodies in Algorithmic Composition," Proceedings of the 9th Sound and Music Computing Conference, Copenhagen, Denmark, p.514-520, 2012 [3] J. Jensen, Evolutionary Music Composition, Institutt for datateknikk og informasjonsvitenskap NTNU University Library, 2011. [4] B. Johanson and R. Poli, GP-Music: An Interactive Genetic Programming System for Music Generation with Automated Fitness Raters, Proceedings of the Third Annual Conference, MIT Press, pp. 181-186, 1998. [5] A. Jordanous, A Fitness Function for Creativity in Jazz Improvisation and Beyond, Proceedings of the First International Conference on Computational Creativity (ICCCX), Lisbon, Portugal, 2010. [6] P. Langston, Six Techniques for Algorithmic Music Composition, 15th International Computer Music Conference (ICMC) in Columbus, Ohio, November 2-5, 1989. [7] M. Lo and S.M. Lucas, Evolving Musical Sequences with N- Gram Based Trainable Fitness Functions, IEEE International Conference on Evolutionary Computation, pp. 601-608, 2006. [8] B. Manaris et al, Developing Fitness Functions for Pleasant Music: Zipf's Law and Interactive Evolution Systems, Applications of Evolutionary Computing, EvoWorkkShops Proceedings,pp. 498-507, 2005. [9] T. Nozaki and K. Kameyama, Feature Selection for Useradaptive Content-Based Music Retrieval using Particle Swarm Optimization Proceedings of 2010 10th International Conference on Intelligent Systems Design and Applications (ISDA 10), pp. 941-946, 2010. [10] G. Papadopoulus and G. Wiggins, AI Methods for Algorithmic Composition: A Survey, a Critical View and Future, Proceedings from the AISB 99 Symposium on Musical Creativity, Edinburg, Scotland, pp. 110-117. [11] M. Towsey et al., Towards Melodic Extension Using Genetic Algorithms, Educational Technology and Society 4, vol. 2, 2001