Mood Tracking of Radio Station Broadcasts

Size: px
Start display at page:

Download "Mood Tracking of Radio Station Broadcasts"

Transcription

1 Mood Tracking of Radio Station Broadcasts Jacek Grekow Faculty of Computer Science, Bialystok University of Technology, Wiejska 45A, Bialystok , Poland Abstract. This paper presents an example of a system for the analysis of emotions contained within radio broadcasts. We prepared training data, did feature extraction, built classifiers for music/speech discrimination and for emotion detection in music. To study changes in emotions, we used recorded broadcasts from 4 selected European radio stations. The collected data allowed us to determine the dominant emotion in the radio broadcasts and construct maps visualizing the distribution of emotions in time. The obtained results provide a new interesting view of the emotional content of radio station broadcasts. Keywords: Emotion detection, Mood tracking, Audio feature extraction, Music information retrieval, Radio broadcasts. 1 Introduction The overwhelming number of media outlets is constantly growing. This also applies to radio stations available on the Internet, over satellite, and over the air. On the one hand, the number of opportunities to listen to various radio shows has grown, but on the other, choosing the right station has become more difficult. Music information retrieval helps those people who listen to the radio mainly for the music. This technology is able to make a general detection of the genre, artist, and even emotion. Listening to music is particularly emotional. People need a variety of emotions, and music is perfectly suited to provide them. Listening to the radio station throughout the day, whether we want to it not, we are affected by the transmitted emotional content. In this paper, we focus on emotional analysis of the music presented by radio stations. During the course of a radio broadcast, these emotions can take on a variety of shades, change several times with varying intensity. This paper presents a method of tracking changing emotions during the course of a radio broadcast. The collected data allowed to determine the dominant emotion in the radio broadcast and construct maps visualizing the distribution of emotions in time. Music emotion detection studies are mainly based on two popular approaches: categorical or dimensional. The categorical approach [1][2][3][4] describes emotions with a discrete number of classes - affective adjectives. In the dimensional approach [5][6], emotions are described as numerical values of valence

2 2 Jacek Grekow and arousal. In this way an emotion of a song is represented as a point on an emotion space. In this work, we use the categorical approach. There are several other studies on the issue of mood tracking [7][8][9]. Lu et al. [3], apart from detecting emotions, tracked them, divided the music into several independent segments, each of which contained a homogeneous emotional expression. The use of mood tracking for indexing and searching multimedia databases has been used in the work of Grekow and Ras [10]. One wonders how long it takes a person to recognize emotion in musical compositions to which he/she is listening. Bachorik et al. [11] concluded that the majority of music listeners need 8 seconds to identify the emotion of a piece. This time is closely related to the length of a segment of music during emotion detection. Xiao et al. [12] found that the segment length should be no shorter than 4 sec and no longer than 16 sec. In various studies on the detection of emotion in music the segments are varying lengths. In [4][1], segment length is 30 sec. A 25 second segment was used by Yang et al. [5]. Fifteen second clips are used as ground truth data by Schmidt et al. [13]. The length of 1 sec segment was used by Schmidt and Kim in [6]. In this work, we use 6-second segments. A comprehensive review of the methods that have been proposed for music emotion recognition was prepared by Yang et al. [14]. Another paper surveying state-of-the-art automatic emotion recognition was presented by Kim et al. in [15]. The issue of mood tracking is not only limited to music. The paper by Mohammad [16] is an interesting extension of the topic; the author investigated the development of emotions in literary texts. Yeh et al. [17] tracked the continuous changes of emotional expressions in Mandarin speech. A method of profiling radio stations was described by Lidy and Rauber [18]. They used a technique of Self-Organizing Maps to organize the program coverage of radio stations on a two-dimensional map. This approach allows to profile the complete program of a radio station. 2 Music Data To conduct the study of emotion detection of radio stations, we prepared two sets of data. One of them was used for music/speech discrimination, and the other for the detection of emotion in music. A set of training data for music/speech discrimination consisted of 128 wav files, including 64 designated as speech and 64 marked as music. The training data were taken from the generally accessible data collection project MARSYAS ( The training data set for emotion detection consisted of 374 six-second fragments of different genres of music: classical, jazz, blues, country, disco, hip-hop, metal, pop, reggae, and rock. The tracks were all 22050Hz Mono 16-bit audio files in wav format. In this research we use 4 emotion classes: energetic-positive, energetic-negative, calm-negative, calm-positive, presented with their abbreviation in Table 1. They

3 Mood Tracking of Radio Station Broadcasts 3 cover the four quadrants of the 2 dimensional Thayer model of emotion [19]. They correspond to four basic emotion classes: happy, angry, sad and relaxed. Abbreviation Description e1 energetic-positive e2 energetic-negative e3 calm-negative e4 calm-positive Table 1. Description of mood labels Music samples were labeled by the author of this paper, a music expert with a university musical education. Six-second music samples were listened to and then labeled with one of the emotions (e1, e2, e3, e4). In the case when the music expert was not certain which emotion to assign, such a sample was rejected. In this way, the created labels were associated with only one emotion in the file. As a result, we obtained 4 sets of files: 101 files labeled e1, 107 files labeled e2, 78 files labeled e3, and 88 files labeled e4. To study changes in emotions, we used recorded broadcasts from 4 selected European radio stations: Polish Radio Dwojka (Classical/Culture), recorded on ; Polish Radio Trojka (Pop/Rock), recorded on ; BBC Radio 3 (Classical), recorded on ; ORF OE1 (Information/Culture), recorded on For each station we recorded 10 hours beginning at 10 A.M. The recorded broadcasts were segmented into 6-second fragments using sfplay.exe from MARSYAS software. For example, we obtained 6000 segments from one 10 h broadcast. 3 Features Extraction For features extraction, we used the framework for audio analysis of MARSYAS software, written by George Tzanetakis [20]. MARSYAS is implemented in C++ and retains the ability to output feature extraction data to ARFF format [21]. With the tool bextract.exe, the following features can be extracted: Zero Crossings, Spectral Centroid, Spectral Flux, Spectral Rolloff, Mel-Frequency Cepstral Coefficients (MFCC), and chroma features - 31 features in total. For each of these basic features, four statistic features were calculated: 1. The mean of the mean (calculate mean over the 20 frames, and then calculate the mean of this statistic over the entire segment); 2. The mean of the standard deviation (calculate the standard deviation of the feature over 20 frames, and then calculate the mean these standard deviations over the entire segment); 3. The standard deviation of the mean (calculate the mean of the feature over

4 4 Jacek Grekow 20 frames, and then calculate the standard deviation of these values over the entire segment); 4. The standard deviation of the standard deviation (calculate the standard deviation of the feature over 20 frames, and then calculate the standard deviation of these values over the entire segment). In this way, we obtained 124 features. The input data during features extraction were 6-second segments in wav format, sample rate 22050, channels: 1, Bits: 16. An example of using bextract.exe from the MARSYAS v0.2 package to extract features: bextract.exe -fe -sv colllection -w outputfile where collection is file with list of input files and outputfile is name of output file in ARFF format. For each 6-second file we obtained a representative single feature vector. The obtained vectors were used for building classifiers and for predicting new instances. 4 Classification 4.1 The Construction of Classifiers We built two classifiers, one for music/speech discrimination and the second for emotion detection, using the WEKA package [21]. During the construction of the classifier for music/speech discrimination, we tested the following algorithms: J48, RandomForest, BayesNet, SMO [22]. The classification results were calculated using a cross validation evaluation CV-10. The best accuracy (98%) was achieved using SMO algorithm, which is an implementation of support vector machines (SVM) algorithm. The second best algorithm was Random-Forest (94% accuracy). During the construction of the classifier for emotion detection, we tested the following algorithms: J48, RandomForest, BayesNet, IBk (K-nn), SMO (SVM). The highest accuracy (55.61%) was obtained for SMO algorithm. SMO was trained using polynominal kernel. The classification results were calculated using a cross validation evaluation CV-10. After applying attribute selection (attribute evaluator: WrapperSubsetEval, search method BestFirst), classifier accuracy improved to 60.69%. classified as a b c d a = e b = e c = e d = e Table 2. Confusion matrix The confusion matrix (Table 2) obtained during classifier evaluation shows that the most recognized emotion was e2 (Precision 0.706, Recall 0.72, F-measure

5 Mood Tracking of Radio Station Broadcasts ), and the next emotion was e1 (Precision 0.542, Recall 0.644, F-measure 0.588). We may notice a considerable amount of mistakes between the emotions of the left and right quadrants of the Thayer model, that is between e1 and e2, and analogously between e3 and e4. This is confirmed by the fact that detection on the arousal axis of Thayers model is easier. There are less mistakes made between the top and bottom quadrants. At the same time, recognition of emotions on the valence axis (positive-negative) is more difficult. 4.2 Analysis of Recordings During the analysis of the recorded radio broadcasts, we conducted a two-phase classification. The recorded radio program was divided into 6-second segments. For each segment, we extracted a feature vector. This feature vector was first used to detect if the given segment is speech or music. If the current segment was music, then we used a second classifier to predict what type of emotion it contained. For features extraction, file segmentation, use of classifiers to predict new instances, and visualization of results, we wrote a Java application that connected different software products: MARSYAS, MATLAB and WEKA package. 5 Results of Mood Tracking in Radio Stations The percentages of speech, music, and emotion in music obtained during the segment classification of 10-hour broadcasts of four radio stations are presented in Table 3. On the basis of these results, radio stations can be compared in two ways. The first way is to compare the amount of music and speech in the radio broadcasts, and the second is to compare the occurrence of individual emotions. PR Dwojka PR Trojka BBC Radio 3 ORF OE1 speech 59.37% 73.35% 32.25% 69.10% music 40.63% 26.65% 67.75% 30.90% e1 4.78% 4.35% 2.43% 2.48% e2 5.35% 14.43% 1.00% 0.92% e % 6.02% 56.19% 22.53% e % 1.85% 8.13% 4.97% e1 in music 11.76% 16.32% 3.58% 8.02% e2 in music 13.16% 54.14% 1.47% 2.98% e3 in music 49.89% 22.59% 82.93% 72.91% e4 in music 25.17% 6.94% 12.00% 16.08% Table 3. Percentage of speech, music, and emotion in music in 10-hour broadcasts of four radio stations

6 6 Jacek Grekow 5.1 Comparison of Radio Stations The dominant station in the amount of music presented was BBC Radio 3 (67.75%). We noted a similar ratio of speech to music in the broadcasts of PR Trojka and ORF OE1, in both of which speech dominated (73.35% and 69.10%, respectively). A more balanced amount of speech and music was noted on PR Dwojka (59.37% and 40.63%, respectively). Comparing the content of emotions, we can see that PR Trojka clearly differs from the other radio stations, because the dominant emotion is e2 energeticnegative (54.14%) and e4 calm-positive occurs the least often (6.94%). We noted a clear similarity between BBC Radio 3 and ORF OE1, where the dominant emotion was e3 calm-negative (82.93% and 72.91%, respectively). Also, the proportions of the other emotions (e1, e2, e4) were similar for these stations. We could say that emotionally these stations are similar, except that considering the speech to music ratio, BBC Radio 3 had much more music. The dominant emotion for PR Dwojka was e3, which is somewhat similar to BBC Radio 3 and ORF OE1. Compared to the other stations, PR Dwojka had the most (25.17%) e4 calm-positive music. 5.2 Emotion Maps The figures (Fig. 1, Fig. 2, Fig. 3, Fig. 4) present speech and emotion maps for each radio broadcast. Each point on the map is the value obtained from the classification of a 6-second segment. These show which emotions occurred at given hours of the broadcasts. For PR Dwojka (Fig. 1), there are clear musical segments ( , ) during which e3 dominated. At the end of the day ( ), emotion e2 occurs sporadically. It is interesting that e1 and e4 (from right half of the Thayer model) did not occur in the morning. For PR Trojka (Fig. 2), emotion e4 did not occur in the morning, and e2 and e3 dominated (segments and ). For BBC Radio 3 (Fig. 3), we observed almost a complete lack of energetic emotions (e1 and e2) in the afternoon (segments after 3200). For ORF OE1 (Fig. 4), e3 dominated up to segment 3600, and then broadcasts without music dominated. The presented analyses of maps of emotions could be developed by examining the quantity of changes of emotions or the distribution of daily emotions. 6 Conclusions This paper presents an example of a system for the analysis of emotions contained within radio broadcasts. The collected data allowed to determine the dominant emotion in the radio broadcast and present the amount of speech and music. The obtained results provide a new interesting view of the emotional content of radio stations. The precision of the constructed maps visualizing the distribution of emotions in time obviously depends on the precision of the classifiers of emotion detection.

7 Mood Tracking of Radio Station Broadcasts 7 Fig. 1. Map of speech and music emotion in PR Dwojka 10h broadcast Fig. 2. Map of speech and music emotion in PR Trojka 10h broadcast

8 8 Jacek Grekow Fig. 3. Map of speech and music emotion in BBC Radio 3 10h broadcast Fig. 4. Map of speech and music emotion in ORF OE1 10h broadcast

9 Mood Tracking of Radio Station Broadcasts 9 Their accuracy could be better. This is still associated with the imperfection of features for audio analysis. In this matter, there is still much to be done. We could also test audio features extracted by other software for feature extraction, such as jaudio or MIR toolbox. Also, musical file labeling, which are input data for learning classifiers, could be made by a bigger number of music experts; this would enhance the reliability of the classifiers. Development of the presented system to include emotion detection in speech also seems to be a logical prospect in the future. A system for the analysis of emotions contained within radio broadcasts could be a helpful tool for people planning radio programs enabling them to consciously plan the emotional distribution in the broadcast music. Another example of applying this system could be an additional tool for radio station searching. Because the perception of emotions can be subjective and different people perceive emotions slightly differently, the emotional analysis of radio stations could be dependent on the user s preferences. Search profiling of radio stations taking into consideration the user would be an interesting solution. Acknowledgments. This paper is supported by the S/WI/3/2013. References 1. Li, T. and Ogihara, M.: Detecting emotion in music. Proceedings of the Fifth International Symposium on Music Information Retrieval, pp (2003) 2. Grekow, J. and Ras, Z.: Detecting emotions in classical music from MIDI files. Foundations of Intelligent Systems: ISMIS 2009, LNAI, Vol. 5722, pp (2009) 3. Lu, L., Liu, D. and Zhang, H.J.: Automatic mood detection and tracking of music audio signals. IEEE Transactions on Audio, Speech and Language Processing, vol. 14, no. 1, pp (2006) 4. Song, Y., Dixon, S. and Pearce, M.: Evaluation of Musical Features for Emotion Classifica-tion. In Proceedings of the 13th International Society for Music Information Retrieval Con-ference (2012) 5. Yang,Y.-H., Lin,Y.C., Su,Y.F. and Chen, H.H.: A regression approach to music emotion recognition. IEEE Transactions on Audio, Speech, and Language Processing, Volume 16, Issue 2, pp (2008) 6. Schmidt, E. and Kim, Y.: Modeling Musical Emotion Dynamics with Conditional Random Fields. In Proceedings of the 12th International Society for Music Information Retrieval Conference, pp (2011) 7. Schmidt, E.M., Turnbull, D. and Kim Y.E.: Feature Selection for Content-Based, Time-Varying Musical Emotion Regression. Proc. ACM SIGMM International Conference on Multimedia Information Retrieval, Philadelphia, PA (2010) 8. Schmidt, E.M. and Kim, Y.E.: Prediction of time-varying musical mood distributions from audio. Proceedings of the 2010 International Society for Music Information Retrieval Conference, Utrecht, Netherlands (2010) 9. Grekow, J.: Mood tracking of musical compositions. Foundations of Intelligent Systems: ISMIS 2012, Lecture Notes in Computer Science, pp , eds. Li Chen, Alexander Felfernig, Jiming Liu, Zbigniew Ras; 20th International Symposium, Macau, China (2012)

10 10 Jacek Grekow 10. Grekow, J. and Ras, Z.: Emotion Based MIDI Files Retrieval System. Advances in Music Information Retrieval, Studies in Computational Intelligence, Springer (2010) 11. Bachorik, J.P., Bangert, M., Loui, P., Larke, K., Berger, J., Rowe, R. and Schlaug, G.: Emotion in motion: Investigating the time-course of emotional judgments of musical stimuli. Music Perception, vol. 26, no. 4, pp (2009) 12. Xiao, Z., Dellandrea, E., Dou, W. and Chen, L.: What is the best segment duration for music mood analysis? International Workshop on Content-Based Multimedia Indexing (CBMI 2008), pp (2008) 13. Schmidt, E. M., Scott, J.J. and Kim, Y.E.: Feature Learning in Dynamic Environments: Modeling the Acoustic Structure of Musical Emotion. In Proceedings of the 12th International Society for Music Information Retrieval Conference, pp (2012) 14. Yang Y.H. and Homer H. Chen, H.H.: Machine Recognition of Music Emotion: A Re-view. ACM Transactions on Intelligent Systems and Technology, Volume 3, Issue 3, Article No. 40 (2012) 15. Kim, Y., Schmidt, E., Migneco, R., Morton, B., Richardson, P., Scott, J., Speck, J. and Turnbull, D.: State of the Art Report: Music Emotion Recognition: A State of the Art Review. In Proceedings of the 11th International Society for Music Information Retrieval Conference, pp (2010) 16. Mohammad, S.: From Once Upon a Time to Happily Ever After: Tracking Emotions in Novels and Fairy Tales. Proceedings of the ACL 2011 Workshop on Language Technology for Cultural Heritage, Social Sciences, and Humanities, pp , Portland, OR, USA (2011) 17. Yeh, J., Pao, T., Pai, Ch., Cheng, Y.: Tracking and Visualizing the Changes of Mandarin Emotional Expression. ICIC 2008, LNCS 5226, pp (2008) 18. Lidy, T. and Rauber, A.: Visually Profiling Radio Stations. In Proceedings of the 7th Inter-national Conference on Music Information Retrieval (2006) 19. Thayer, R.E.: The biopsychology arousal. Oxford University Press (1989) 20. Tzanetakis, G. and Cook, P.: Marsyas: A framework for audio analysis. Organized Sound, 10: (2000) 21. Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P. and Witten, I.H.: The WEKA Data Mining Software: An Update; SIGKDD Explorations, Volume 11, Issue 1 (2009) 22. Witten, I.H. and Frank, E.: Data Mining: Practical machine learning tools and techniques. Morgan Kaufmann, San Francisco, CA, USA (2005)

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION ULAŞ BAĞCI AND ENGIN ERZIN arxiv:0907.3220v1 [cs.sd] 18 Jul 2009 ABSTRACT. Music genre classification is an essential tool for

More information

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

Music Emotion Recognition. Jaesung Lee. Chung-Ang University Music Emotion Recognition Jaesung Lee Chung-Ang University Introduction Searching Music in Music Information Retrieval Some information about target music is available Query by Text: Title, Artist, or

More information

Music Genre Classification and Variance Comparison on Number of Genres

Music Genre Classification and Variance Comparison on Number of Genres Music Genre Classification and Variance Comparison on Number of Genres Miguel Francisco, miguelf@stanford.edu Dong Myung Kim, dmk8265@stanford.edu 1 Abstract In this project we apply machine learning techniques

More information

Predicting Time-Varying Musical Emotion Distributions from Multi-Track Audio

Predicting Time-Varying Musical Emotion Distributions from Multi-Track Audio Predicting Time-Varying Musical Emotion Distributions from Multi-Track Audio Jeffrey Scott, Erik M. Schmidt, Matthew Prockup, Brandon Morton, and Youngmoo E. Kim Music and Entertainment Technology Laboratory

More information

Exploring Relationships between Audio Features and Emotion in Music

Exploring Relationships between Audio Features and Emotion in Music Exploring Relationships between Audio Features and Emotion in Music Cyril Laurier, *1 Olivier Lartillot, #2 Tuomas Eerola #3, Petri Toiviainen #4 * Music Technology Group, Universitat Pompeu Fabra, Barcelona,

More information

A Categorical Approach for Recognizing Emotional Effects of Music

A Categorical Approach for Recognizing Emotional Effects of Music A Categorical Approach for Recognizing Emotional Effects of Music Mohsen Sahraei Ardakani 1 and Ehsan Arbabi School of Electrical and Computer Engineering, College of Engineering, University of Tehran,

More information

Music Information Retrieval with Temporal Features and Timbre

Music Information Retrieval with Temporal Features and Timbre Music Information Retrieval with Temporal Features and Timbre Angelina A. Tzacheva and Keith J. Bell University of South Carolina Upstate, Department of Informatics 800 University Way, Spartanburg, SC

More information

ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC

ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC Vaiva Imbrasaitė, Peter Robinson Computer Laboratory, University of Cambridge, UK Vaiva.Imbrasaite@cl.cam.ac.uk

More information

Dimensional Music Emotion Recognition: Combining Standard and Melodic Audio Features

Dimensional Music Emotion Recognition: Combining Standard and Melodic Audio Features Dimensional Music Emotion Recognition: Combining Standard and Melodic Audio Features R. Panda 1, B. Rocha 1 and R. P. Paiva 1, 1 CISUC Centre for Informatics and Systems of the University of Coimbra, Portugal

More information

Subjective Similarity of Music: Data Collection for Individuality Analysis

Subjective Similarity of Music: Data Collection for Individuality Analysis Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp

More information

MUSI-6201 Computational Music Analysis

MUSI-6201 Computational Music Analysis MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)

More information

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset Ricardo Malheiro, Renato Panda, Paulo Gomes, Rui Paiva CISUC Centre for Informatics and Systems of the University of Coimbra {rsmal,

More information

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM A QUER B EAMPLE MUSIC RETRIEVAL ALGORITHM H. HARB AND L. CHEN Maths-Info department, Ecole Centrale de Lyon. 36, av. Guy de Collongue, 69134, Ecully, France, EUROPE E-mail: {hadi.harb, liming.chen}@ec-lyon.fr

More information

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis Fengyan Wu fengyanyy@163.com Shutao Sun stsun@cuc.edu.cn Weiyao Xue Wyxue_std@163.com Abstract Automatic extraction of

More information

VECTOR REPRESENTATION OF EMOTION FLOW FOR POPULAR MUSIC. Chia-Hao Chung and Homer Chen

VECTOR REPRESENTATION OF EMOTION FLOW FOR POPULAR MUSIC. Chia-Hao Chung and Homer Chen VECTOR REPRESENTATION OF EMOTION FLOW FOR POPULAR MUSIC Chia-Hao Chung and Homer Chen National Taiwan University Emails: {b99505003, homer}@ntu.edu.tw ABSTRACT The flow of emotion expressed by music through

More information

Outline. Why do we classify? Audio Classification

Outline. Why do we classify? Audio Classification Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification Implementation Future Work Why do we classify

More information

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Aric Bartle (abartle@stanford.edu) December 14, 2012 1 Background The field of composer recognition has

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

Multimodal Music Mood Classification Framework for Christian Kokborok Music

Multimodal Music Mood Classification Framework for Christian Kokborok Music Journal of Engineering Technology (ISSN. 0747-9964) Volume 8, Issue 1, Jan. 2019, PP.506-515 Multimodal Music Mood Classification Framework for Christian Kokborok Music Sanchali Das 1*, Sambit Satpathy

More information

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? NICHOLAS BORG AND GEORGE HOKKANEN Abstract. The possibility of a hit song prediction algorithm is both academically interesting and industry motivated.

More information

Singer Identification

Singer Identification Singer Identification Bertrand SCHERRER McGill University March 15, 2007 Bertrand SCHERRER (McGill University) Singer Identification March 15, 2007 1 / 27 Outline 1 Introduction Applications Challenges

More information

Analytic Comparison of Audio Feature Sets using Self-Organising Maps

Analytic Comparison of Audio Feature Sets using Self-Organising Maps Analytic Comparison of Audio Feature Sets using Self-Organising Maps Rudolf Mayer, Jakob Frank, Andreas Rauber Institute of Software Technology and Interactive Systems Vienna University of Technology,

More information

Chord Classification of an Audio Signal using Artificial Neural Network

Chord Classification of an Audio Signal using Artificial Neural Network Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------

More information

Music Mood Classification - an SVM based approach. Sebastian Napiorkowski

Music Mood Classification - an SVM based approach. Sebastian Napiorkowski Music Mood Classification - an SVM based approach Sebastian Napiorkowski Topics on Computer Music (Seminar Report) HPAC - RWTH - SS2015 Contents 1. Motivation 2. Quantification and Definition of Mood 3.

More information

A Study on Cross-cultural and Cross-dataset Generalizability of Music Mood Regression Models

A Study on Cross-cultural and Cross-dataset Generalizability of Music Mood Regression Models A Study on Cross-cultural and Cross-dataset Generalizability of Music Mood Regression Models Xiao Hu University of Hong Kong xiaoxhu@hku.hk Yi-Hsuan Yang Academia Sinica yang@citi.sinica.edu.tw ABSTRACT

More information

Research & Development. White Paper WHP 232. A Large Scale Experiment for Mood-based Classification of TV Programmes BRITISH BROADCASTING CORPORATION

Research & Development. White Paper WHP 232. A Large Scale Experiment for Mood-based Classification of TV Programmes BRITISH BROADCASTING CORPORATION Research & Development White Paper WHP 232 September 2012 A Large Scale Experiment for Mood-based Classification of TV Programmes Jana Eggink, Denise Bland BRITISH BROADCASTING CORPORATION White Paper

More information

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC Scientific Journal of Impact Factor (SJIF): 5.71 International Journal of Advance Engineering and Research Development Volume 5, Issue 04, April -2018 e-issn (O): 2348-4470 p-issn (P): 2348-6406 MUSICAL

More information

Toward Multi-Modal Music Emotion Classification

Toward Multi-Modal Music Emotion Classification Toward Multi-Modal Music Emotion Classification Yi-Hsuan Yang 1, Yu-Ching Lin 1, Heng-Tze Cheng 1, I-Bin Liao 2, Yeh-Chin Ho 2, and Homer H. Chen 1 1 National Taiwan University 2 Telecommunication Laboratories,

More information

Singer Traits Identification using Deep Neural Network

Singer Traits Identification using Deep Neural Network Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic

More information

A Music Retrieval System Using Melody and Lyric

A Music Retrieval System Using Melody and Lyric 202 IEEE International Conference on Multimedia and Expo Workshops A Music Retrieval System Using Melody and Lyric Zhiyuan Guo, Qiang Wang, Gang Liu, Jun Guo, Yueming Lu 2 Pattern Recognition and Intelligent

More information

A Large Scale Experiment for Mood-Based Classification of TV Programmes

A Large Scale Experiment for Mood-Based Classification of TV Programmes 2012 IEEE International Conference on Multimedia and Expo A Large Scale Experiment for Mood-Based Classification of TV Programmes Jana Eggink BBC R&D 56 Wood Lane London, W12 7SB, UK jana.eggink@bbc.co.uk

More information

Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections

Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections 1/23 Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections Rudolf Mayer, Andreas Rauber Vienna University of Technology {mayer,rauber}@ifs.tuwien.ac.at Robert Neumayer

More information

MODELING MUSICAL MOOD FROM AUDIO FEATURES AND LISTENING CONTEXT ON AN IN-SITU DATA SET

MODELING MUSICAL MOOD FROM AUDIO FEATURES AND LISTENING CONTEXT ON AN IN-SITU DATA SET MODELING MUSICAL MOOD FROM AUDIO FEATURES AND LISTENING CONTEXT ON AN IN-SITU DATA SET Diane Watson University of Saskatchewan diane.watson@usask.ca Regan L. Mandryk University of Saskatchewan regan.mandryk@usask.ca

More information

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors Priyanka S. Jadhav M.E. (Computer Engineering) G. H. Raisoni College of Engg. & Mgmt. Wagholi, Pune, India E-mail:

More information

Supervised Learning in Genre Classification

Supervised Learning in Genre Classification Supervised Learning in Genre Classification Introduction & Motivation Mohit Rajani and Luke Ekkizogloy {i.mohit,luke.ekkizogloy}@gmail.com Stanford University, CS229: Machine Learning, 2009 Now that music

More information

Automatic Laughter Detection

Automatic Laughter Detection Automatic Laughter Detection Mary Knox Final Project (EECS 94) knoxm@eecs.berkeley.edu December 1, 006 1 Introduction Laughter is a powerful cue in communication. It communicates to listeners the emotional

More information

Automatic Rhythmic Notation from Single Voice Audio Sources

Automatic Rhythmic Notation from Single Voice Audio Sources Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung

More information

Lyric-Based Music Mood Recognition

Lyric-Based Music Mood Recognition Lyric-Based Music Mood Recognition Emil Ian V. Ascalon, Rafael Cabredo De La Salle University Manila, Philippines emil.ascalon@yahoo.com, rafael.cabredo@dlsu.edu.ph Abstract: In psychology, emotion is

More information

Features for Audio and Music Classification

Features for Audio and Music Classification Features for Audio and Music Classification Martin F. McKinney and Jeroen Breebaart Auditory and Multisensory Perception, Digital Signal Processing Group Philips Research Laboratories Eindhoven, The Netherlands

More information

AN EMOTION MODEL FOR MUSIC USING BRAIN WAVES

AN EMOTION MODEL FOR MUSIC USING BRAIN WAVES AN EMOTION MODEL FOR MUSIC USING BRAIN WAVES Rafael Cabredo 1,2, Roberto Legaspi 1, Paul Salvador Inventado 1,2, and Masayuki Numao 1 1 Institute of Scientific and Industrial Research, Osaka University,

More information

Quality of Music Classification Systems: How to build the Reference?

Quality of Music Classification Systems: How to build the Reference? Quality of Music Classification Systems: How to build the Reference? Janto Skowronek, Martin F. McKinney Digital Signal Processing Philips Research Laboratories Eindhoven {janto.skowronek,martin.mckinney}@philips.com

More information

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular Music Mood Sheng Xu, Albert Peyton, Ryan Bhular What is Music Mood A psychological & musical topic Human emotions conveyed in music can be comprehended from two aspects: Lyrics Music Factors that affect

More information

Capturing the Temporal Domain in Echonest Features for Improved Classification Effectiveness

Capturing the Temporal Domain in Echonest Features for Improved Classification Effectiveness Capturing the Temporal Domain in Echonest Features for Improved Classification Effectiveness Alexander Schindler 1,2 and Andreas Rauber 1 1 Department of Software Technology and Interactive Systems Vienna

More information

Melody classification using patterns

Melody classification using patterns Melody classification using patterns Darrell Conklin Department of Computing City University London United Kingdom conklin@city.ac.uk Abstract. A new method for symbolic music classification is proposed,

More information

Some Experiments in Humour Recognition Using the Italian Wikiquote Collection

Some Experiments in Humour Recognition Using the Italian Wikiquote Collection Some Experiments in Humour Recognition Using the Italian Wikiquote Collection Davide Buscaldi and Paolo Rosso Dpto. de Sistemas Informáticos y Computación (DSIC), Universidad Politécnica de Valencia, Spain

More information

Creating a Feature Vector to Identify Similarity between MIDI Files

Creating a Feature Vector to Identify Similarity between MIDI Files Creating a Feature Vector to Identify Similarity between MIDI Files Joseph Stroud 2017 Honors Thesis Advised by Sergio Alvarez Computer Science Department, Boston College 1 Abstract Today there are many

More information

Automatic Music Genre Classification

Automatic Music Genre Classification Automatic Music Genre Classification Nathan YongHoon Kwon, SUNY Binghamton Ingrid Tchakoua, Jackson State University Matthew Pietrosanu, University of Alberta Freya Fu, Colorado State University Yue Wang,

More information

Automatic Mood Detection of Music Audio Signals: An Overview

Automatic Mood Detection of Music Audio Signals: An Overview Automatic Mood Detection of Music Audio Signals: An Overview Sonal P.Sumare 1 Mr. D.G.Bhalke 2 1.(PG Student Department of Electronics and Telecommunication Rajarshi Shahu College of Engineering Pune)

More information

Discovering Similar Music for Alpha Wave Music

Discovering Similar Music for Alpha Wave Music Discovering Similar Music for Alpha Wave Music Yu-Lung Lo ( ), Chien-Yu Chiu, and Ta-Wei Chang Department of Information Management, Chaoyang University of Technology, 168, Jifeng E. Road, Wufeng District,

More information

Music Information Retrieval

Music Information Retrieval Music Information Retrieval Automatic genre classification from acoustic features DANIEL RÖNNOW and THEODOR TWETMAN Bachelor of Science Thesis Stockholm, Sweden 2012 Music Information Retrieval Automatic

More information

MusCat: A Music Browser Featuring Abstract Pictures and Zooming User Interface

MusCat: A Music Browser Featuring Abstract Pictures and Zooming User Interface MusCat: A Music Browser Featuring Abstract Pictures and Zooming User Interface 1st Author 1st author's affiliation 1st line of address 2nd line of address Telephone number, incl. country code 1st author's

More information

Automatic Detection of Emotion in Music: Interaction with Emotionally Sensitive Machines

Automatic Detection of Emotion in Music: Interaction with Emotionally Sensitive Machines Automatic Detection of Emotion in Music: Interaction with Emotionally Sensitive Machines Cyril Laurier, Perfecto Herrera Music Technology Group Universitat Pompeu Fabra Barcelona, Spain {cyril.laurier,perfecto.herrera}@upf.edu

More information

The Role of Time in Music Emotion Recognition

The Role of Time in Music Emotion Recognition The Role of Time in Music Emotion Recognition Marcelo Caetano 1 and Frans Wiering 2 1 Institute of Computer Science, Foundation for Research and Technology - Hellas FORTH-ICS, Heraklion, Crete, Greece

More information

Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting

Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting Dalwon Jang 1, Seungjae Lee 2, Jun Seok Lee 2, Minho Jin 1, Jin S. Seo 2, Sunil Lee 1 and Chang D. Yoo 1 1 Korea Advanced

More information

Improving Music Mood Annotation Using Polygonal Circular Regression. Isabelle Dufour B.Sc., University of Victoria, 2013

Improving Music Mood Annotation Using Polygonal Circular Regression. Isabelle Dufour B.Sc., University of Victoria, 2013 Improving Music Mood Annotation Using Polygonal Circular Regression by Isabelle Dufour B.Sc., University of Victoria, 2013 A Thesis Submitted in Partial Fulfillment of the Requirements for the Degree of

More information

TOWARD UNDERSTANDING EXPRESSIVE PERCUSSION THROUGH CONTENT BASED ANALYSIS

TOWARD UNDERSTANDING EXPRESSIVE PERCUSSION THROUGH CONTENT BASED ANALYSIS TOWARD UNDERSTANDING EXPRESSIVE PERCUSSION THROUGH CONTENT BASED ANALYSIS Matthew Prockup, Erik M. Schmidt, Jeffrey Scott, and Youngmoo E. Kim Music and Entertainment Technology Laboratory (MET-lab) Electrical

More information

IMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM

IMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM IMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM Thomas Lidy, Andreas Rauber Vienna University of Technology, Austria Department of Software

More information

Psychophysiological measures of emotional response to Romantic orchestral music and their musical and acoustic correlates

Psychophysiological measures of emotional response to Romantic orchestral music and their musical and acoustic correlates Psychophysiological measures of emotional response to Romantic orchestral music and their musical and acoustic correlates Konstantinos Trochidis, David Sears, Dieu-Ly Tran, Stephen McAdams CIRMMT, Department

More information

arxiv: v1 [cs.ai] 30 Nov 2016

arxiv: v1 [cs.ai] 30 Nov 2016 Fusion of EEG and Musical Features in Continuous Music-emotion Recognition Nattapong Thammasan 1,*, Ken-ichi Fukui 2, and Masayuki Numao 2 1 Graduate school of Information Science and Technology, Osaka

More information

MELODY ANALYSIS FOR PREDICTION OF THE EMOTIONS CONVEYED BY SINHALA SONGS

MELODY ANALYSIS FOR PREDICTION OF THE EMOTIONS CONVEYED BY SINHALA SONGS MELODY ANALYSIS FOR PREDICTION OF THE EMOTIONS CONVEYED BY SINHALA SONGS M.G.W. Lakshitha, K.L. Jayaratne University of Colombo School of Computing, Sri Lanka. ABSTRACT: This paper describes our attempt

More information

Multi-label classification of emotions in music

Multi-label classification of emotions in music Multi-label classification of emotions in music Alicja Wieczorkowska 1, Piotr Synak 1, and Zbigniew W. Raś 2,1 1 Polish-Japanese Institute of Information Technology, Koszykowa 86, 02-008 Warsaw, Poland

More information

Mood Classification Using Lyrics and Audio: A Case-Study in Greek Music

Mood Classification Using Lyrics and Audio: A Case-Study in Greek Music Mood Classification Using Lyrics and Audio: A Case-Study in Greek Music Spyros Brilis, Evagelia Gkatzou, Antonis Koursoumis, Karolos Talvis, Katia Kermanidis, Ioannis Karydis To cite this version: Spyros

More information

Detecting Musical Key with Supervised Learning

Detecting Musical Key with Supervised Learning Detecting Musical Key with Supervised Learning Robert Mahieu Department of Electrical Engineering Stanford University rmahieu@stanford.edu Abstract This paper proposes and tests performance of two different

More information

Homework 2 Key-finding algorithm

Homework 2 Key-finding algorithm Homework 2 Key-finding algorithm Li Su Research Center for IT Innovation, Academia, Taiwan lisu@citi.sinica.edu.tw (You don t need any solid understanding about the musical key before doing this homework,

More information

Analysing Musical Pieces Using harmony-analyser.org Tools

Analysing Musical Pieces Using harmony-analyser.org Tools Analysing Musical Pieces Using harmony-analyser.org Tools Ladislav Maršík Dept. of Software Engineering, Faculty of Mathematics and Physics Charles University, Malostranské nám. 25, 118 00 Prague 1, Czech

More information

Recommending Music for Language Learning: The Problem of Singing Voice Intelligibility

Recommending Music for Language Learning: The Problem of Singing Voice Intelligibility Recommending Music for Language Learning: The Problem of Singing Voice Intelligibility Karim M. Ibrahim (M.Sc.,Nile University, Cairo, 2016) A THESIS SUBMITTED FOR THE DEGREE OF MASTER OF SCIENCE DEPARTMENT

More information

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST) Computational Models of Music Similarity 1 Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST) Abstract The perceived similarity of two pieces of music is multi-dimensional,

More information

arxiv: v1 [cs.ir] 16 Jan 2019

arxiv: v1 [cs.ir] 16 Jan 2019 It s Only Words And Words Are All I Have Manash Pratim Barman 1, Kavish Dahekar 2, Abhinav Anshuman 3, and Amit Awekar 4 1 Indian Institute of Information Technology, Guwahati 2 SAP Labs, Bengaluru 3 Dell

More information

Lyrics Classification using Naive Bayes

Lyrics Classification using Naive Bayes Lyrics Classification using Naive Bayes Dalibor Bužić *, Jasminka Dobša ** * College for Information Technologies, Klaićeva 7, Zagreb, Croatia ** Faculty of Organization and Informatics, Pavlinska 2, Varaždin,

More information

MPEG-7 AUDIO SPECTRUM BASIS AS A SIGNATURE OF VIOLIN SOUND

MPEG-7 AUDIO SPECTRUM BASIS AS A SIGNATURE OF VIOLIN SOUND MPEG-7 AUDIO SPECTRUM BASIS AS A SIGNATURE OF VIOLIN SOUND Aleksander Kaminiarz, Ewa Łukasik Institute of Computing Science, Poznań University of Technology. Piotrowo 2, 60-965 Poznań, Poland e-mail: Ewa.Lukasik@cs.put.poznan.pl

More information

HOW COOL IS BEBOP JAZZ? SPONTANEOUS

HOW COOL IS BEBOP JAZZ? SPONTANEOUS HOW COOL IS BEBOP JAZZ? SPONTANEOUS CLUSTERING AND DECODING OF JAZZ MUSIC Antonio RODÀ *1, Edoardo DA LIO a, Maddalena MURARI b, Sergio CANAZZA a a Dept. of Information Engineering, University of Padova,

More information

WHEN LYRICS OUTPERFORM AUDIO FOR MUSIC MOOD CLASSIFICATION: A FEATURE ANALYSIS

WHEN LYRICS OUTPERFORM AUDIO FOR MUSIC MOOD CLASSIFICATION: A FEATURE ANALYSIS WHEN LYRICS OUTPERFORM AUDIO FOR MUSIC MOOD CLASSIFICATION: A FEATURE ANALYSIS Xiao Hu J. Stephen Downie Graduate School of Library and Information Science University of Illinois at Urbana-Champaign xiaohu@illinois.edu

More information

World Academy of Science, Engineering and Technology International Journal of Computer and Information Engineering Vol:6, No:12, 2012

World Academy of Science, Engineering and Technology International Journal of Computer and Information Engineering Vol:6, No:12, 2012 A method for Music Classification based on Perceived Mood Detection for Indian Bollywood Music Vallabha Hampiholi Abstract A lot of research has been done in the past decade in the field of audio content

More information

EVALUATION OF FEATURE EXTRACTORS AND PSYCHO-ACOUSTIC TRANSFORMATIONS FOR MUSIC GENRE CLASSIFICATION

EVALUATION OF FEATURE EXTRACTORS AND PSYCHO-ACOUSTIC TRANSFORMATIONS FOR MUSIC GENRE CLASSIFICATION EVALUATION OF FEATURE EXTRACTORS AND PSYCHO-ACOUSTIC TRANSFORMATIONS FOR MUSIC GENRE CLASSIFICATION Thomas Lidy Andreas Rauber Vienna University of Technology Department of Software Technology and Interactive

More information

Multi-modal Analysis for Person Type Classification in News Video

Multi-modal Analysis for Person Type Classification in News Video Multi-modal Analysis for Person Type Classification in News Video Jun Yang, Alexander G. Hauptmann School of Computer Science, Carnegie Mellon University, 5000 Forbes Ave, PA 15213, USA {juny, alex}@cs.cmu.edu,

More information

Hidden Markov Model based dance recognition

Hidden Markov Model based dance recognition Hidden Markov Model based dance recognition Dragutin Hrenek, Nenad Mikša, Robert Perica, Pavle Prentašić and Boris Trubić University of Zagreb, Faculty of Electrical Engineering and Computing Unska 3,

More information

Improving Frame Based Automatic Laughter Detection

Improving Frame Based Automatic Laughter Detection Improving Frame Based Automatic Laughter Detection Mary Knox EE225D Class Project knoxm@eecs.berkeley.edu December 13, 2007 Abstract Laughter recognition is an underexplored area of research. My goal for

More information

HUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH

HUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH Proc. of the th Int. Conference on Digital Audio Effects (DAFx-), Hamburg, Germany, September -8, HUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH George Tzanetakis, Georg Essl Computer

More information

A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION

A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION Graham E. Poliner and Daniel P.W. Ellis LabROSA, Dept. of Electrical Engineering Columbia University, New York NY 127 USA {graham,dpwe}@ee.columbia.edu

More information

MUSICAL MOODS: A MASS PARTICIPATION EXPERIMENT FOR AFFECTIVE CLASSIFICATION OF MUSIC

MUSICAL MOODS: A MASS PARTICIPATION EXPERIMENT FOR AFFECTIVE CLASSIFICATION OF MUSIC 12th International Society for Music Information Retrieval Conference (ISMIR 2011) MUSICAL MOODS: A MASS PARTICIPATION EXPERIMENT FOR AFFECTIVE CLASSIFICATION OF MUSIC Sam Davies, Penelope Allen, Mark

More information

Speech To Song Classification

Speech To Song Classification Speech To Song Classification Emily Graber Center for Computer Research in Music and Acoustics, Department of Music, Stanford University Abstract The speech to song illusion is a perceptual phenomenon

More information

Topics in Computer Music Instrument Identification. Ioanna Karydi

Topics in Computer Music Instrument Identification. Ioanna Karydi Topics in Computer Music Instrument Identification Ioanna Karydi Presentation overview What is instrument identification? Sound attributes & Timbre Human performance The ideal algorithm Selected approaches

More information

Speech and Speaker Recognition for the Command of an Industrial Robot

Speech and Speaker Recognition for the Command of an Industrial Robot Speech and Speaker Recognition for the Command of an Industrial Robot CLAUDIA MOISA*, HELGA SILAGHI*, ANDREI SILAGHI** *Dept. of Electric Drives and Automation University of Oradea University Street, nr.

More information

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES 12th International Society for Music Information Retrieval Conference (ISMIR 2011) A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES Erdem Unal 1 Elaine Chew 2 Panayiotis Georgiou

More information

GRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM

GRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM 19th European Signal Processing Conference (EUSIPCO 2011) Barcelona, Spain, August 29 - September 2, 2011 GRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM Tomoko Matsui

More information

Perceptual dimensions of short audio clips and corresponding timbre features

Perceptual dimensions of short audio clips and corresponding timbre features Perceptual dimensions of short audio clips and corresponding timbre features Jason Musil, Budr El-Nusairi, Daniel Müllensiefen Department of Psychology, Goldsmiths, University of London Question How do

More information

Automatic Laughter Detection

Automatic Laughter Detection Automatic Laughter Detection Mary Knox 1803707 knoxm@eecs.berkeley.edu December 1, 006 Abstract We built a system to automatically detect laughter from acoustic features of audio. To implement the system,

More information

Automatic Musical Pattern Feature Extraction Using Convolutional Neural Network

Automatic Musical Pattern Feature Extraction Using Convolutional Neural Network Automatic Musical Pattern Feature Extraction Using Convolutional Neural Network Tom LH. Li, Antoni B. Chan and Andy HW. Chun Abstract Music genre classification has been a challenging yet promising task

More information

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval DAY 1 Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval Jay LeBoeuf Imagine Research jay{at}imagine-research.com Rebecca

More information

Coimbra, Coimbra, Portugal Published online: 18 Apr To link to this article:

Coimbra, Coimbra, Portugal Published online: 18 Apr To link to this article: This article was downloaded by: [Professor Rui Pedro Paiva] On: 14 May 2015, At: 03:23 Publisher: Taylor & Francis Informa Ltd Registered in England and Wales Registered Number: 1072954 Registered office:

More information

Audio-Based Video Editing with Two-Channel Microphone

Audio-Based Video Editing with Two-Channel Microphone Audio-Based Video Editing with Two-Channel Microphone Tetsuya Takiguchi Organization of Advanced Science and Technology Kobe University, Japan takigu@kobe-u.ac.jp Yasuo Ariki Organization of Advanced Science

More information

Acoustic Scene Classification

Acoustic Scene Classification Acoustic Scene Classification Marc-Christoph Gerasch Seminar Topics in Computer Music - Acoustic Scene Classification 6/24/2015 1 Outline Acoustic Scene Classification - definition History and state of

More information

Melody Retrieval On The Web

Melody Retrieval On The Web Melody Retrieval On The Web Thesis proposal for the degree of Master of Science at the Massachusetts Institute of Technology M.I.T Media Laboratory Fall 2000 Thesis supervisor: Barry Vercoe Professor,

More information

HIT SONG SCIENCE IS NOT YET A SCIENCE

HIT SONG SCIENCE IS NOT YET A SCIENCE HIT SONG SCIENCE IS NOT YET A SCIENCE François Pachet Sony CSL pachet@csl.sony.fr Pierre Roy Sony CSL roy@csl.sony.fr ABSTRACT We describe a large-scale experiment aiming at validating the hypothesis that

More information

Polyphonic Audio Matching for Score Following and Intelligent Audio Editors

Polyphonic Audio Matching for Score Following and Intelligent Audio Editors Polyphonic Audio Matching for Score Following and Intelligent Audio Editors Roger B. Dannenberg and Ning Hu School of Computer Science, Carnegie Mellon University email: dannenberg@cs.cmu.edu, ninghu@cs.cmu.edu,

More information

Drum Stroke Computing: Multimodal Signal Processing for Drum Stroke Identification and Performance Metrics

Drum Stroke Computing: Multimodal Signal Processing for Drum Stroke Identification and Performance Metrics Drum Stroke Computing: Multimodal Signal Processing for Drum Stroke Identification and Performance Metrics Jordan Hochenbaum 1, 2 New Zealand School of Music 1 PO Box 2332 Wellington 6140, New Zealand

More information

Speech Recognition Combining MFCCs and Image Features

Speech Recognition Combining MFCCs and Image Features Speech Recognition Combining MFCCs and Image Featres S. Karlos from Department of Mathematics N. Fazakis from Department of Electrical and Compter Engineering K. Karanikola from Department of Mathematics

More information

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION Halfdan Rump, Shigeki Miyabe, Emiru Tsunoo, Nobukata Ono, Shigeki Sagama The University of Tokyo, Graduate

More information

A Survey Of Mood-Based Music Classification

A Survey Of Mood-Based Music Classification A Survey Of Mood-Based Music Classification Sachin Dhande 1, Bhavana Tiple 2 1 Department of Computer Engineering, MIT PUNE, Pune, India, 2 Department of Computer Engineering, MIT PUNE, Pune, India, Abstract

More information

Music Recommendation from Song Sets

Music Recommendation from Song Sets Music Recommendation from Song Sets Beth Logan Cambridge Research Laboratory HP Laboratories Cambridge HPL-2004-148 August 30, 2004* E-mail: Beth.Logan@hp.com music analysis, information retrieval, multimedia

More information