Automatic Music Genre Classification

Similar documents
WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

Using Genre Classification to Make Content-based Music Recommendations

Automatic Music Clustering using Audio Attributes

Music Genre Classification

Can Song Lyrics Predict Genre? Danny Diekroeger Stanford University

Detecting Musical Key with Supervised Learning

Music Genre Classification and Variance Comparison on Number of Genres

MUSI-6201 Computational Music Analysis

Music Mood Classication Using The Million Song Dataset

The Million Song Dataset

Supervised Learning in Genre Classification

Chord Classification of an Audio Signal using Artificial Neural Network

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

Music Information Retrieval

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

Automatic Labelling of tabla signals

Automatic Piano Music Transcription

Release Year Prediction for Songs

Predicting Hit Songs with MIDI Musical Features

Singer Traits Identification using Deep Neural Network

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

Capturing the Temporal Domain in Echonest Features for Improved Classification Effectiveness

CS229 Project Report Polyphonic Piano Transcription

Analysing Musical Pieces Using harmony-analyser.org Tools

Effects of acoustic degradations on cover song recognition

Analytic Comparison of Audio Feature Sets using Self-Organising Maps

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

A Survey Of Mood-Based Music Classification

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

MODELS of music begin with a representation of the

Week 14 Music Understanding and Classification

SONG-LEVEL FEATURES AND SUPPORT VECTOR MACHINES FOR MUSIC CLASSIFICATION

Acoustic Scene Classification

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)

Composer Style Attribution

Predicting Time-Varying Musical Emotion Distributions from Multi-Track Audio

HIDDEN MARKOV MODELS FOR SPECTRAL SIMILARITY OF SONGS. Arthur Flexer, Elias Pampalk, Gerhard Widmer

Music Composition with RNN

Automatic Rhythmic Notation from Single Voice Audio Sources

arxiv: v1 [cs.ir] 16 Jan 2019

Music Mood Classification - an SVM based approach. Sebastian Napiorkowski

A Survey of Audio-Based Music Classification and Annotation

Singer Identification

Large-Scale Pattern Discovery in Music. Thierry Bertin-Mahieux

Neural Network Predicating Movie Box Office Performance

Improving Frame Based Automatic Laughter Detection

Lyrics Classification using Naive Bayes

COMBINING FEATURES REDUCES HUBNESS IN AUDIO SIMILARITY

Automatic Music Similarity Assessment and Recommendation. A Thesis. Submitted to the Faculty. Drexel University. Donald Shaul Williamson

hit), and assume that longer incidental sounds (forest noise, water, wind noise) resemble a Gaussian noise distribution.

Visual Encoding Design

A System for Automatic Chord Transcription from Audio Using Genre-Specific Hidden Markov Models

A CLASSIFICATION-BASED POLYPHONIC PIANO TRANSCRIPTION APPROACH USING LEARNED FEATURE REPRESENTATIONS

A Study on Music Genre Recognition and Classification Techniques

MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES

Creating a Feature Vector to Identify Similarity between MIDI Files

A Language Modeling Approach for the Classification of Audio Music

A Survey on: Sound Source Separation Methods

ISMIR 2008 Session 2a Music Recommendation and Organization

Lecture 15: Research at LabROSA

Music Information Retrieval

Neural Network for Music Instrument Identi cation

Topics in Computer Music Instrument Identification. Ioanna Karydi

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

Music Information Retrieval with Temporal Features and Timbre

Computational Modelling of Harmony

TIMBRAL MODELING FOR MUSIC ARTIST RECOGNITION USING I-VECTORS. Hamid Eghbal-zadeh, Markus Schedl and Gerhard Widmer

USING ARTIST SIMILARITY TO PROPAGATE SEMANTIC INFORMATION

GRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM

Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections

LEARNING SPECTRAL FILTERS FOR SINGLE- AND MULTI-LABEL CLASSIFICATION OF MUSICAL INSTRUMENTS. Patrick Joseph Donnelly

Automatic Musical Pattern Feature Extraction Using Convolutional Neural Network

Contextual music information retrieval and recommendation: State of the art and challenges

Kent Academic Repository

Content-based music retrieval

Exploring the Design Space of Symbolic Music Genre Classification Using Data Mining Techniques Ortiz-Arroyo, Daniel; Kofod, Christian

CTP431- Music and Audio Computing Music Information Retrieval. Graduate School of Culture Technology KAIST Juhan Nam

Automatic Laughter Detection

The Effect of DJs Social Network on Music Popularity

Image Steganalysis: Challenges

Automatic Laughter Detection

Feature-Based Analysis of Haydn String Quartets

Classification of Timbre Similarity

ALF-200k: Towards Extensive Multimodal Analyses of Music Tracks and Playlists

Analyzing the Relationship Among Audio Labels Using Hubert-Arabie adjusted Rand Index

Hidden Markov Model based dance recognition

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

Semi-supervised Musical Instrument Recognition

Joint Image and Text Representation for Aesthetics Analysis

EVALUATION OF FEATURE EXTRACTORS AND PSYCHO-ACOUSTIC TRANSFORMATIONS FOR MUSIC GENRE CLASSIFICATION

Limitations of interactive music recommendation based on audio content

MODELING GENRE WITH THE MUSIC GENOME PROJECT: COMPARING HUMAN-LABELED ATTRIBUTES AND AUDIO FEATURES

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

Audio Structure Analysis

Recognising Cello Performers using Timbre Models

Multiple classifiers for different features in timbre estimation

Transcription:

Automatic Music Genre Classification Nathan YongHoon Kwon, SUNY Binghamton Ingrid Tchakoua, Jackson State University Matthew Pietrosanu, University of Alberta Freya Fu, Colorado State University Yue Wang, University of Alberta Victoria Sabo, Georgetown University Group Mentor: Sercan Yildiz

Outline Introduction Background Previous work General approach Methods Results Conclusions Acknowledgments https://musicmachinery.com/2013/09/22/5025/

Context/motivation Automatic music genre classification Pandora s Genome Project and Spotify s Echo Nest API Music sorting/searching, recommendations Human labor is expensive Investigate algorithms for genre classification - accurate and efficient https://thenextweb.com/insider/2013/05/29/spotify-finally-discovers-socialrecommends-new-music-and-more-to-users-but-only-for-the-webnow/#.tnw_wy2w27nc

Data Openly available through Million Song Dataset - 300GB Each song has song-wide information and segment-based information segments- song broken up these.25 second-long pieces which are screened for their own information features that cannot be quantified by the human ear segments_loudness song_hotttnesss tempo loudness *https://labrosa.ee.columbia.edu/millionsong/

Early work Early attempt: Tzanetakis and Cook (2002) Further attempts since then: features: timbre texture (MFCCs), rhythmic content, pitch... methods: neural network, Gaussian classifier, SVM, K-nearest neighbor classifier... http://newatlas.com/automatic-music-genre-classificationsystem/38240/

Methods Supervised learning with genre-labeled training data Feature extraction/ selection Tempo, loudness, key, pitch, etc. Dimension reduction to focus on specific features/genres Correlation Matrix between characteristics

General approach and machine learning techniques Logistic regression Dimension reduction Gaussian naive Bayes k-means clustering k-nn classifier Support vector machine Decision tree Neural network http://docs.opencv.org/2.4/doc/tutorials/ml/introduction_to _svm/introduction_to_svm.html

Best Subset Selection Gaussian Naive Bayes: Accuracy: training 33.78%, testing 31.75% Support Vector Machine (SVM) Accuracy: training 64.14%, testing 55.36% Logistic Regression Accuracy : training 54.24%, testing 53.19% Decision Trees Accuracy: training 100%, testing 35.70% https://www.analyticsvidhya.com/blog/2016/04/completetutorial-tree-based-modeling-scratch-in-python/ https://www.researchgate.net/figure/255695722_fig1_figure- 1-Illustration-of-how-a-Gaussian-Naive-Bayes-GNBclassifier-works-For-each

Gaussian Classifier Used on MFC coefficients (timbre) Test accuracy: 61.75% http://modelai.gettysburg.edu/2012/music/index.html Confusion matrix: Column labels are actual genres.

Gaussian Classifier Used on MFC coefficients (timbre) Test accuracy: 61.75% Removing pop increased accuracy to 66.93% http://modelai.gettysburg.edu/2012/music/index.html Confusion matrix w/o pop: Column labels are actual genres

Dimension Reduction Helpful in figuring out how to further classify the data scatter plot showing results of t- SNE

k-means Clustering t-sne suggested genres that could be easily separated Standard approach: Euclidean Distance each song represented as the sample mean of its 12-dimensional MFCCs (Symmetrized) KL divergence: each song represented as an MVN distribution KL divergence methods more successful: rap vs. blues: 98.8%, rap vs. blues vs. reggae: 93.5% accuracy Euclidean distance: blues vs. reggae vs rap KL divergence: blues, reggae, rap using

Segment-Level: k-nn Classifier Model most prevalent pitch chroma and timbre waveform as a Markov chain Expect different genres to differ in transition matrices Example heatmap for rock transition matrix (pitch) Example heatmap for blues transition matrix (pitch)

Segment-Level: k-nn Classifier k-nn classifier (k=50) Accuracy (pitch): training 48.5%, testing 52.7% Accuracy (timbre): training 51.24%, testing 54.75% Combined accuracy (pitch+timbre): training 60.2%, testing 59% Confusion matrix for combined test

Neural Network Artificial neural network inspired by the human brain Single layer used to classify all 14 genres (17 hidden units with 30,000 iterations) Test accuracy : 62.28% Confusion matrix for combined features: Columns labels are actual genres.

Accuracy of Different Test- Feature Combinations

Limitations Considers a limited number of genres, not subgenres or new genres Training set comprises mostly rock Missing data on danceability and song_hotttnesss (Time) https://sites.google.com/site/kfrawleymusicgenresincontext/

Future Work Combining both song-level and segment-level attributes into the same test for more methods Better methods for variable screening for this kind of data Pattern recognition and other machine learning techniques (eg: convolutional network, residual learning) Scaling methods to full (300 GB) dataset https://clipartfest.com/categori es/view/d0342e4e212661cf48 756958b25b170b904a6a8b/pa ttern-recognition.html

And to finish... https://xkcd.com/1838/

Acknowledgments Sercan Yildiz Thomas Gehrmann

References Tzanetakis, G., and P. Cook. "Musical genre classification of audio signals." IEEE Transactions on Speech and Audio Processing 10.5 (2002): 293-302. Web. Thierry Bertin-Mahieux, Daniel P. W. Ellis, Brian Whitman, and Paul Lamere. The Million Song Dataset. In Proceedings of the 12th International Society for Music Information Retrieval Conference (ISMIR 2011), pages 591-596, Oct. 2011. Alexander Schindler, Rudolf Mayer, and Andreas Rauber. Facilitating comprehensive benchmarking experiments on the million song dataset. In Proceedings of the 13th International Society for Music Information Retrieval Conference (ISMIR 2012), pages 469-474, Oct. 2012.