Using Genre Classification to Make Content-based Music Recommendations

Size: px
Start display at page:

Download "Using Genre Classification to Make Content-based Music Recommendations"

Transcription

1 Using Genre Classification to Make Content-based Music Recommendations Robbie Jones and Karen Lu CS 221, Autumn 2016 Stanford University I. Introduction Our project goal is to create personalized music recommendations based on song similarity. Popular music recommender systems like Pandora, itunes Genius, and Spotify Discover use a variety of information filtering techniques to engage users through the discovery of new music that they would enjoy. These recommender systems go beyond surface-level classifications such as pop songs, songs from the year 2010, or songs by Bruce Springsteen. Rather, they consider similarities in user profiles, including demographic data, geographic information, and similar likes and dislikes, as well as similarities in the audio files themselves. In this project, we choose to focus specifically on developing effective content-based recommendations by looking at song similarity through the lenses of lyrics and audio metadata. II. Task Definition We broke down our project into two main tasks. Firstly, we chose to focus on multi-class genre classification as our method for training song features relevant to identifying song similarity. We decided to concentrate on predicting pre-selected genres because humans generally accept these genres to be a useful method of clustering music according to similarity. Most of us would consider two country songs, for example, to be more similar than a country song and an electronic song. In the process of classifying songs by genre, we create a high-dimensional feature space model that we can then use to approach our second task: given a set of songs that a user likes, we create a content-based recommendation system to deliver additional songs suitable to their taste. III. Literature Review We initially discovered our dataset (the Million Song Dataset, described in detail in Section IV) by researching similar projects that predict song popularity based on audio metadata [4, 6, 10]. These projects define a song s features based on metadata such as tempo, key, time signature, etc. and output a popularity index that can be ranked against other songs. The evaluation of popularity can be achieved via predefined rankings such as Billboard Top 10. Although the models and algorithms vary widely across projects, the task is more or less the same. 1

2 Although our initial project idea was to predict song popularity, we soon became more interested in song recommender systems. Some previous recommendation work has been done using our dataset; the creators of the dataset even released the Million Song Dataset Challenge in 2012, encouraging participants to create an offline music recommendation system to be evaluated against real users listening history [8]. Some popular music recommendation services in industry include Spotify [1, 5] and Pandora [7]. Of course, neither of these companies make their specific models or algorithms publically available, but we can still get an overview of the general frameworks. Spotify has historically relied on collaborative filtering as their main method of generating recommendations. Collaborative filtering determines user preferences based on the usages of other users -- users that have listened to similar sets of songs are likely to have similar taste. Issues with collaborative filtering approaches include the cold-start problem (the system doesn t recommend songs that are new) and the fact that popular songs are much more likely to be recommended than unpopular songs. After acquiring The Echo Nest, Spotify has start to incorporate content-based recommendation into its product. Pandora, on the other hand, focuses on content-based recommendation. Their Music Genome Project documents 400 musical attributes per song, hand-labeled by music experts. IV. Data Song Metadata Our main dataset was downloaded from the Million Song Dataset [2], which was compiled by Daniel P.W. Ellis and Thierry Bertin-Mahieux from the Laboratory for the Recognition and Organization of Speech and Audio (LabROSA) of Columbia University. This publically available dataset contains audio features and metadata information for one million individual songs. Most of the fields were extracted using the Echo Nest API, although it also contains song IDs from other websites, like musicbrainz.org and playme.com (Echo Nest was acquired by Spotify for $100 million in March 2014). Fields include basic metadata like loudness, artist location, and duration as well as more subjective or algorithmically calculated values like danceability and time signature. The data is formatted in individual HDF5 files per song, and a single reduced summary file of non-array fields for all one million songs is also available. A SQLite database of most metadata from each track is also available for download. Lyrics Data Since the Million Song Dataset was created and published in February 2011, several additional compatible datasets have also been compiled and shared with the research community. One dataset that we found particularly interesting was a lyrics dataset developed in a partnership between Million Song Dataset and musixmatch [9]. The team 2

3 was able to resolve lyrics for over 77% of the dataset, of which the lyrics for 237,662 tracks are released. The 5,000 most common words from the dataset are included in bag-of-words format (unordered word counts). Genre Data For our genre classification task, we required labeled ground truth genre data for each track in our training set. We chose to use a dataset of genre annotations developed by tagtraum industries incorporated because the dataset uses a majority voting method for combining crowd-sourced genre labels from multiple sources, including the Last.fm datase and the beatunes Genre Dataset (BGD) [3]. In total, the dataset contains genre labels for 280,831 tracks. The genre dataset defines fifteen ground truth genres: Rock, Pop, Metal, Country, Rap, RnB, Electronic, Punk, Latin, Folk, Reggae, Jazz, Blues, World, and New. Creating Our Dataset In order select tracks at the intersection of all three datasets, we created a Python parser for the lyrics and genre text files and loaded the relevant information from each dataset into tables in a SQLite database. By joining the tables on track_id, we found that a subset of 122,447 tracks included both tagtraum genre annotations and musixmatch dataset of song lyrics. We ran some queries to examine the number of tracks for each genre (Figure 1). Figure 1. Count of Tracks with Genre and Lyric Data, By Genre. Genre Track Count Rock Pop Metal Country Rap RnB* Electronic Punk Latin Folk Reggae Jazz Blues World New * In this dataset, RnB refers to the genre Rhythm and Blues. 3

4 Next, we created balanced datasets with equal examples from each of ten most common genres: Rock, Pop, Metal, Country, Rap, RnB, Electronic, Punk, Latin, and Folk. Our training set included 2500 songs of each genre, or 25,000 songs in total, while our test set included 100 songs of each genre, or 1,000 songs in total. V. Approach Model We aim to classify the genre of a song since it seems to be the most natural way to cluster groups of songs. We model individual songs in classic machine learning fashion as a point in a high-dimensional space, where each dimension represents a value in the corresponding feature vector. We model our problem as a multi-class classification problem: Learn a classifier F that accurately classifies a song by genre, such that f( Φ (x)) = y i, where Φ (x) is the feature vector of x and y i is the ground truth genre. Input: x = track_id from Million Song Dataset. Output: y { y 1, y 2,, y K }, where y i are all of the possible predefined genres. Algorithms Our goal was to choose a feature extractor and algorithm that would incorporate information from both audio metadata and word counts. In order to accomplish this task, we chose to use a combination of two machine learning algorithms: one-vs-rest and k-nearest neighbors. First, we focused on word features only and derived a list of ten confidence scores per song, one for each of our ten predefined genres. One-vs-rest is a strategy that trains a binary classifier for each class (genre), using tracks from that genre as positive training examples and tracks from all other genres as negative training examples. In order to evaluate a new track, we run it through all ten genre classifiers and choose the genre with the highest confidence score. Instead of simply outputting the final genre classification, we chose to keep the ten confidence scores and use them as features in the next step of our project. We chose to do this instead of using the 5,000 word count features directly in order to reduce our feature space and also prevent recommending songs based primarily on word overlap. The next step in our implementation is to use our word feature confidence scores along with metadata features to train a multiclass classifier for genre. Because of its simplicity, effectiveness and easy training stage, we decided to start by implementing the k-nearest neighbors algorithm. With k-nearest neighbors, the training set s confidence score and metadata can be used directly as feature vectors without needing to tune parameters. At test time, we simply find the k points with the smallest Euclidian distance and choose the genre based on a simple majority vote. 4

5 Once we ve selected our features in order to optimize our genre classifier s accuracy, we can use the same vector space model to recommend similar songs by finding the closest neighbors in vector space. If done effectively, this allows us to define more fine-grained similarity than a genre label could capture. V. Data and Experimentation Baseline and Oracle For our project baseline, we ran the k-nearest neighbors using k = 1 with 5 floating-point features. These features, taken directly from the Million Song Dataset format, were: duration: length of the song in seconds tempo: speed of the song in beats per minute loudness: loudness in decibels derived from a model of human listening song hotttnesss: an aggregate measure of how much buzz the song is getting (derived from online mentions, play counts, etc.) danceability: a derived measure of the ease with which a person could dance to a song We trained our model using a very small training set of 40 tracks for each of three genres: Rock, Pop, and Rap. We tested our model using a test set of 30 tracks comprised of 10 tracks each for each of the same three genres. Our baseline method s accuracy was 37%. Our oracle consisted simply of using our ground truth genre labels, resulting in an accuracy of 100%. One-vs-rest Classification on Word Features We started by training a single classifier for each class, such that training examples labeled with that genre are considered positive, while all other training examples are considered negative. Using the one-vs-rest strategy, we achieved an overall accuracy of 47.2% using a Naive Bayes classifier and 44.0% using a logistic regression classifier (Figure 2). In Figure 3, we present more fine-grained accuracies based on the ground truth genre of each test set example. The chart in Figure 4 displays these accuracies for the Naive Bayes classifier. 5

6 Figure 2. Overall Accuracies of One-vs-rest Classification using Word Count Features. Binary Classification Algorithm Naive Bayes Logistic Regression Overall Test Set Accuracy 47.2% 44.0% Figure 3. Accuracies of One-vs-rest Classification using Word Count Features, by Genre. Genre Latin Rap Country RnB Punk Metal Folk Electronic Pop Rock Test Set Accuracy using Naive Bayes 91.0% 82.0% 69.0% 68.0% 54.0% 53.0% 34.0% 9.0% 6.0% 5.0% Test Set Accuracy using Logistic Regression Figure 4. Accuracies of One-vs-rest Classification with Naive Bayes Classifier using Word Count Features. 6

7 k-nearest Neighbors for Classification and Recommendations We ran our k-nearest neighbors algorithm on word count confidence scores only, song metadata features only, and both sets of features combined. These results are documented in Figure 5. Figure 5. Comparison of k-nearest Neighbors Classification Accuracies Using Word Count Confidence Scores vs. Acoustic Metadata (k = 21). Word Count Confidence Scores Only Acoustic Metadata Only Both Word Count Confidence Scores and Acoustic Metadata Accuracy on Training Set 67.1% 22.5% 25.5% Accuracy on Test Set 46.7% 12.5% 14.0% Since our acoustic metadata accuracies were so low, we also ran our algorithm individually on each feature to see if we could find individual features that worked well (Figure 6). Figure 6. Comparison of k-nearest Neighbors Classification Accuracies for Individual Acoustic Metadata Features (k = 21). duration end_of_fade_in key loudness mode start_of_fade_in tempo time_signature Accuracy on Training Set 10.0% 10.1% 9.8% 10.0% 9.7% 21.5% 21.8% 10.2% Accuracy on Test Set 10.0% 9.5% 8.5% 10.0% 12.6% 10.1% 10.3% 10.1% We also tried various values of k (k = 1, k = 11, k = 21, k = 31), but the results for each were similar to the results for k = 21 (Figure 5 above). 7

8 VI. Analysis Interpretation and Discussion For our initial task of training a one-vs-rest classifier with word count features, we were able to achieve an accuracy of 47.2% across all ten genres. While this accuracy is not fantastic, it is much higher than random (10% for classification across all ten genres). Based on our results in Section V, it is evident that using a bag-of-word model on song lyrics is much more informative for some genres than for others. For example, when using Naive Bayes, we were able to achieve 91% accuracy on tracks classified as Latin and 82% accuracy on tracks classified as Rap, but we were only able to achieve 6% and 5% on Pop and Rock, respectively. We surmise that this large variation in accuracy between genres is due both to the more specific slang and vernacular of some genres, as well as more focused themes and subjects. Our Nearest-Neighbors algorithm also did not perform as well as we had hoped. While we initially assumed that songs from different from genres would have noticeably different metadata, it s possible that the variance of metadata of songs within individual genres was enough to throw off the algorithm. In addition, there s a good chance that whatever intrinsic qualities in genres that allow our ears to distinguish one from the other is simply not captured within the metadata features that we used for our classification. Challenges Many of the challenges we faced with this project involved our inexperience with working with datasets of this size. Since the main Million Song Dataset contains metadata for 1,000,000 songs, our feature extraction process was slow. Even though we selected a relatively small training set of 25,000 songs, for example, we needed to go through all 1,000,000 songs in order to find the subset of track_ids that we desired. We started from an unbalanced dataset; for example, in the intersection of the Million Song Database, our genre database, and our lyrics database, there are 61,531 rock songs but only 2,762 folk songs. In addition, our data came from multiple sources in multiple formats, including plain text files, HDF5 files, and SQLite databases. We also faced several other challenges in the realms of feature engineering and evaluation. We lacked some domain-specific knowledge and intuition for how to engineer features that work well with our algorithms to classify effectively. In addition, while evaluation of genre classification is simple given our dataset of ground-truth genre labels, evaluation of our unsupervised learning task of song recommendation is much more open-ended. Even once we have achieved reasonable genre classification, it is difficult to determine whether user-specific recommendations are effective. 8

9 VII. Conclusion In order to classify songs by ten predefined genres, we took a two-pronged approach. First, we used one-vs-rest classifier powered by Naive Bayes to find confidence scores for each genre based simply on bag-of-words word count features. Next, we combined word count features and audio metadata features to train a k-nearest neighbors genre classifier, with limited success. In order to recommend songs based on user preferences, we searched our training set examples for the nearest songs in this high-dimensional vector space. Our research indicated that word count features performed much better than the acoustic metadata features available to us. With more domain-specific knowledge and better feature engineering, perhaps our accuracies would improve. VIII. References [1] Dieleman, Sander. "Recommending Music on Spotify with Deep Learning." Sander Dieleman. N.p., 5 Aug Web. 15 Dec [2] "Getting the Dataset." Million Song Dataset. LabROSA, Echo Nest, n.d. Web. 26 Oct [3] Hendrik Schreiber. Improving genre annotations for the million song dataset. In Proceedings of the 16th International Society for Music Information Retrieval Conference (ISMIR), pages , Málaga, Spain, Oct [4] Herremans, Dorien, David Martens, and Kenneth Sörensen. "Dance Hit Song Prediction." Journal of New Music Research 43.3 (2014): Web. 26 Oct [5] Johnson, Chris. "Algorithmic Music Recommendations at Spotify." LinkedIn SlideShare. N.p., 13 Jan Web. 15 Dec [6] Koenigstein, Noam, Yuval Shavitt, and Noa Zilberman. "Predicting Billboard Success Using Data-Mining in P2P Networks - Semantic Scholar." Predicting Billboard Success Using Data-Mining in P2P Networks. N.p., n.d. Web. 26 Oct [7] Layton, Julia. "How Pandora Radio Works." HowStuffWorks. N.p., 23 May Web. 16 Dec [8] Mcfee, Brian, Thierry Bertin-Mahieux, Daniel P.w. Ellis, and Gert R.g. Lanckriet. "Million Song Dataset Challenge." N.p., n.d. Web. 26 Oct [9] "The MusiXmatch Dataset: Connecting Lyrics." Million Song Dataset. LabROSA, Echo Nest, 11 Apr Web. 15 Dec [10] Pham, James, Edric Kyauk, and Edwin Park. "Predicting Song Popularity." (n.d.): n. pag. Web. 26 Oct

Can Song Lyrics Predict Genre? Danny Diekroeger Stanford University

Can Song Lyrics Predict Genre? Danny Diekroeger Stanford University Can Song Lyrics Predict Genre? Danny Diekroeger Stanford University danny1@stanford.edu 1. Motivation and Goal Music has long been a way for people to express their emotions. And because we all have a

More information

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? NICHOLAS BORG AND GEORGE HOKKANEN Abstract. The possibility of a hit song prediction algorithm is both academically interesting and industry motivated.

More information

Automatic Music Genre Classification

Automatic Music Genre Classification Automatic Music Genre Classification Nathan YongHoon Kwon, SUNY Binghamton Ingrid Tchakoua, Jackson State University Matthew Pietrosanu, University of Alberta Freya Fu, Colorado State University Yue Wang,

More information

Automatic Music Clustering using Audio Attributes

Automatic Music Clustering using Audio Attributes Automatic Music Clustering using Audio Attributes Abhishek Sen BTech (Electronics) Veermata Jijabai Technological Institute (VJTI), Mumbai, India abhishekpsen@gmail.com Abstract Music brings people together,

More information

The Million Song Dataset

The Million Song Dataset The Million Song Dataset AUDIO FEATURES The Million Song Dataset There is no data like more data Bob Mercer of IBM (1985). T. Bertin-Mahieux, D.P.W. Ellis, B. Whitman, P. Lamere, The Million Song Dataset,

More information

Music Genre Classification

Music Genre Classification Music Genre Classification chunya25 Fall 2017 1 Introduction A genre is defined as a category of artistic composition, characterized by similarities in form, style, or subject matter. [1] Some researchers

More information

Detecting Musical Key with Supervised Learning

Detecting Musical Key with Supervised Learning Detecting Musical Key with Supervised Learning Robert Mahieu Department of Electrical Engineering Stanford University rmahieu@stanford.edu Abstract This paper proposes and tests performance of two different

More information

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

Music Emotion Recognition. Jaesung Lee. Chung-Ang University Music Emotion Recognition Jaesung Lee Chung-Ang University Introduction Searching Music in Music Information Retrieval Some information about target music is available Query by Text: Title, Artist, or

More information

Lecture 15: Research at LabROSA

Lecture 15: Research at LabROSA ELEN E4896 MUSIC SIGNAL PROCESSING Lecture 15: Research at LabROSA 1. Sources, Mixtures, & Perception 2. Spatial Filtering 3. Time-Frequency Masking 4. Model-Based Separation Dan Ellis Dept. Electrical

More information

MUSI-6201 Computational Music Analysis

MUSI-6201 Computational Music Analysis MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)

More information

http://www.xkcd.com/655/ Audio Retrieval David Kauchak cs160 Fall 2009 Thanks to Doug Turnbull for some of the slides Administrative CS Colloquium vs. Wed. before Thanksgiving producers consumers 8M artists

More information

Music Information Retrieval

Music Information Retrieval CTP 431 Music and Audio Computing Music Information Retrieval Graduate School of Culture Technology (GSCT) Juhan Nam 1 Introduction ü Instrument: Piano ü Composer: Chopin ü Key: E-minor ü Melody - ELO

More information

Music Information Retrieval

Music Information Retrieval Music Information Retrieval Automatic genre classification from acoustic features DANIEL RÖNNOW and THEODOR TWETMAN Bachelor of Science Thesis Stockholm, Sweden 2012 Music Information Retrieval Automatic

More information

Music Genre Classification and Variance Comparison on Number of Genres

Music Genre Classification and Variance Comparison on Number of Genres Music Genre Classification and Variance Comparison on Number of Genres Miguel Francisco, miguelf@stanford.edu Dong Myung Kim, dmk8265@stanford.edu 1 Abstract In this project we apply machine learning techniques

More information

Singer Traits Identification using Deep Neural Network

Singer Traits Identification using Deep Neural Network Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic

More information

Predicting Hit Songs with MIDI Musical Features

Predicting Hit Songs with MIDI Musical Features Predicting Hit Songs with MIDI Musical Features Keven (Kedao) Wang Stanford University kvw@stanford.edu ABSTRACT This paper predicts hit songs based on musical features from MIDI files. The task is modeled

More information

A Music Recommendation System Based on User Behaviors and Genre Classification

A Music Recommendation System Based on User Behaviors and Genre Classification University of Miami Scholarly Repository Open Access Theses Electronic Theses and Dissertations --7 A Music Recommendation System Based on User Behaviors and Genre Classification Yajie Hu University of

More information

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval DAY 1 Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval Jay LeBoeuf Imagine Research jay{at}imagine-research.com Rebecca

More information

Automatic Piano Music Transcription

Automatic Piano Music Transcription Automatic Piano Music Transcription Jianyu Fan Qiuhan Wang Xin Li Jianyu.Fan.Gr@dartmouth.edu Qiuhan.Wang.Gr@dartmouth.edu Xi.Li.Gr@dartmouth.edu 1. Introduction Writing down the score while listening

More information

Supervised Learning in Genre Classification

Supervised Learning in Genre Classification Supervised Learning in Genre Classification Introduction & Motivation Mohit Rajani and Luke Ekkizogloy {i.mohit,luke.ekkizogloy}@gmail.com Stanford University, CS229: Machine Learning, 2009 Now that music

More information

Music Mood Classification - an SVM based approach. Sebastian Napiorkowski

Music Mood Classification - an SVM based approach. Sebastian Napiorkowski Music Mood Classification - an SVM based approach Sebastian Napiorkowski Topics on Computer Music (Seminar Report) HPAC - RWTH - SS2015 Contents 1. Motivation 2. Quantification and Definition of Mood 3.

More information

Release Year Prediction for Songs

Release Year Prediction for Songs Release Year Prediction for Songs [CSE 258 Assignment 2] Ruyu Tan University of California San Diego PID: A53099216 rut003@ucsd.edu Jiaying Liu University of California San Diego PID: A53107720 jil672@ucsd.edu

More information

Music Mood Classication Using The Million Song Dataset

Music Mood Classication Using The Million Song Dataset Music Mood Classication Using The Million Song Dataset Bhavika Tekwani December 12, 2016 Abstract In this paper, music mood classication is tackled from an audio signal analysis perspective. There's an

More information

Outline. Why do we classify? Audio Classification

Outline. Why do we classify? Audio Classification Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification Implementation Future Work Why do we classify

More information

CTP431- Music and Audio Computing Music Information Retrieval. Graduate School of Culture Technology KAIST Juhan Nam

CTP431- Music and Audio Computing Music Information Retrieval. Graduate School of Culture Technology KAIST Juhan Nam CTP431- Music and Audio Computing Music Information Retrieval Graduate School of Culture Technology KAIST Juhan Nam 1 Introduction ü Instrument: Piano ü Genre: Classical ü Composer: Chopin ü Key: E-minor

More information

arxiv: v1 [cs.ir] 16 Jan 2019

arxiv: v1 [cs.ir] 16 Jan 2019 It s Only Words And Words Are All I Have Manash Pratim Barman 1, Kavish Dahekar 2, Abhinav Anshuman 3, and Amit Awekar 4 1 Indian Institute of Information Technology, Guwahati 2 SAP Labs, Bengaluru 3 Dell

More information

Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections

Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections 1/23 Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections Rudolf Mayer, Andreas Rauber Vienna University of Technology {mayer,rauber}@ifs.tuwien.ac.at Robert Neumayer

More information

Creating a Feature Vector to Identify Similarity between MIDI Files

Creating a Feature Vector to Identify Similarity between MIDI Files Creating a Feature Vector to Identify Similarity between MIDI Files Joseph Stroud 2017 Honors Thesis Advised by Sergio Alvarez Computer Science Department, Boston College 1 Abstract Today there are many

More information

Topics in Computer Music Instrument Identification. Ioanna Karydi

Topics in Computer Music Instrument Identification. Ioanna Karydi Topics in Computer Music Instrument Identification Ioanna Karydi Presentation overview What is instrument identification? Sound attributes & Timbre Human performance The ideal algorithm Selected approaches

More information

GENDER IDENTIFICATION AND AGE ESTIMATION OF USERS BASED ON MUSIC METADATA

GENDER IDENTIFICATION AND AGE ESTIMATION OF USERS BASED ON MUSIC METADATA GENDER IDENTIFICATION AND AGE ESTIMATION OF USERS BASED ON MUSIC METADATA Ming-Ju Wu Computer Science Department National Tsing Hua University Hsinchu, Taiwan brian.wu@mirlab.org Jyh-Shing Roger Jang Computer

More information

Analyzing the Relationship Among Audio Labels Using Hubert-Arabie adjusted Rand Index

Analyzing the Relationship Among Audio Labels Using Hubert-Arabie adjusted Rand Index Analyzing the Relationship Among Audio Labels Using Hubert-Arabie adjusted Rand Index Kwan Kim Submitted in partial fulfillment of the requirements for the Master of Music in Music Technology in the Department

More information

Enhancing Music Maps

Enhancing Music Maps Enhancing Music Maps Jakob Frank Vienna University of Technology, Vienna, Austria http://www.ifs.tuwien.ac.at/mir frank@ifs.tuwien.ac.at Abstract. Private as well as commercial music collections keep growing

More information

Exploring the Design Space of Symbolic Music Genre Classification Using Data Mining Techniques Ortiz-Arroyo, Daniel; Kofod, Christian

Exploring the Design Space of Symbolic Music Genre Classification Using Data Mining Techniques Ortiz-Arroyo, Daniel; Kofod, Christian Aalborg Universitet Exploring the Design Space of Symbolic Music Genre Classification Using Data Mining Techniques Ortiz-Arroyo, Daniel; Kofod, Christian Published in: International Conference on Computational

More information

Lyrics Classification using Naive Bayes

Lyrics Classification using Naive Bayes Lyrics Classification using Naive Bayes Dalibor Bužić *, Jasminka Dobša ** * College for Information Technologies, Klaićeva 7, Zagreb, Croatia ** Faculty of Organization and Informatics, Pavlinska 2, Varaždin,

More information

SIGNAL + CONTEXT = BETTER CLASSIFICATION

SIGNAL + CONTEXT = BETTER CLASSIFICATION SIGNAL + CONTEXT = BETTER CLASSIFICATION Jean-Julien Aucouturier Grad. School of Arts and Sciences The University of Tokyo, Japan François Pachet, Pierre Roy, Anthony Beurivé SONY CSL Paris 6 rue Amyot,

More information

Music Recommendation from Song Sets

Music Recommendation from Song Sets Music Recommendation from Song Sets Beth Logan Cambridge Research Laboratory HP Laboratories Cambridge HPL-2004-148 August 30, 2004* E-mail: Beth.Logan@hp.com music analysis, information retrieval, multimedia

More information

ALF-200k: Towards Extensive Multimodal Analyses of Music Tracks and Playlists

ALF-200k: Towards Extensive Multimodal Analyses of Music Tracks and Playlists ALF-200k: Towards Extensive Multimodal Analyses of Music Tracks and Playlists Eva Zangerle, Michael Tschuggnall, Stefan Wurzinger, Günther Specht Department of Computer Science Universität Innsbruck firstname.lastname@uibk.ac.at

More information

Assigning and Visualizing Music Genres by Web-based Co-Occurrence Analysis

Assigning and Visualizing Music Genres by Web-based Co-Occurrence Analysis Assigning and Visualizing Music Genres by Web-based Co-Occurrence Analysis Markus Schedl 1, Tim Pohle 1, Peter Knees 1, Gerhard Widmer 1,2 1 Department of Computational Perception, Johannes Kepler University,

More information

Week 14 Music Understanding and Classification

Week 14 Music Understanding and Classification Week 14 Music Understanding and Classification Roger B. Dannenberg Professor of Computer Science, Music & Art Overview n Music Style Classification n What s a classifier? n Naïve Bayesian Classifiers n

More information

Analysis and Clustering of Musical Compositions using Melody-based Features

Analysis and Clustering of Musical Compositions using Melody-based Features Analysis and Clustering of Musical Compositions using Melody-based Features Isaac Caswell Erika Ji December 13, 2013 Abstract This paper demonstrates that melodic structure fundamentally differentiates

More information

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC Scientific Journal of Impact Factor (SJIF): 5.71 International Journal of Advance Engineering and Research Development Volume 5, Issue 04, April -2018 e-issn (O): 2348-4470 p-issn (P): 2348-6406 MUSICAL

More information

Semi-supervised Musical Instrument Recognition

Semi-supervised Musical Instrument Recognition Semi-supervised Musical Instrument Recognition Master s Thesis Presentation Aleksandr Diment 1 1 Tampere niversity of Technology, Finland Supervisors: Adj.Prof. Tuomas Virtanen, MSc Toni Heittola 17 May

More information

Sarcasm Detection in Text: Design Document

Sarcasm Detection in Text: Design Document CSC 59866 Senior Design Project Specification Professor Jie Wei Wednesday, November 23, 2016 Sarcasm Detection in Text: Design Document Jesse Feinman, James Kasakyan, Jeff Stolzenberg 1 Table of contents

More information

Learning Word Meanings and Descriptive Parameter Spaces from Music. Brian Whitman, Deb Roy and Barry Vercoe MIT Media Lab

Learning Word Meanings and Descriptive Parameter Spaces from Music. Brian Whitman, Deb Roy and Barry Vercoe MIT Media Lab Learning Word Meanings and Descriptive Parameter Spaces from Music Brian Whitman, Deb Roy and Barry Vercoe MIT Media Lab Music intelligence Structure Structure Genre Genre / / Style Style ID ID Song Song

More information

Analysing Musical Pieces Using harmony-analyser.org Tools

Analysing Musical Pieces Using harmony-analyser.org Tools Analysing Musical Pieces Using harmony-analyser.org Tools Ladislav Maršík Dept. of Software Engineering, Faculty of Mathematics and Physics Charles University, Malostranské nám. 25, 118 00 Prague 1, Czech

More information

Lyric-Based Music Genre Classification. Junru Yang B.A.Honors in Management, Nanjing University of Posts and Telecommunications, 2014

Lyric-Based Music Genre Classification. Junru Yang B.A.Honors in Management, Nanjing University of Posts and Telecommunications, 2014 Lyric-Based Music Genre Classification by Junru Yang B.A.Honors in Management, Nanjing University of Posts and Telecommunications, 2014 A Project Submitted in Partial Fulfillment of the Requirements for

More information

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM A QUER B EAMPLE MUSIC RETRIEVAL ALGORITHM H. HARB AND L. CHEN Maths-Info department, Ecole Centrale de Lyon. 36, av. Guy de Collongue, 69134, Ecully, France, EUROPE E-mail: {hadi.harb, liming.chen}@ec-lyon.fr

More information

Capturing the Temporal Domain in Echonest Features for Improved Classification Effectiveness

Capturing the Temporal Domain in Echonest Features for Improved Classification Effectiveness Capturing the Temporal Domain in Echonest Features for Improved Classification Effectiveness Alexander Schindler 1,2 and Andreas Rauber 1 1 Department of Software Technology and Interactive Systems Vienna

More information

Lyric-Based Music Mood Recognition

Lyric-Based Music Mood Recognition Lyric-Based Music Mood Recognition Emil Ian V. Ascalon, Rafael Cabredo De La Salle University Manila, Philippines emil.ascalon@yahoo.com, rafael.cabredo@dlsu.edu.ph Abstract: In psychology, emotion is

More information

Subjective Similarity of Music: Data Collection for Individuality Analysis

Subjective Similarity of Music: Data Collection for Individuality Analysis Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp

More information

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset Ricardo Malheiro, Renato Panda, Paulo Gomes, Rui Paiva CISUC Centre for Informatics and Systems of the University of Coimbra {rsmal,

More information

Supporting Information

Supporting Information Supporting Information I. DATA Discogs.com is a comprehensive, user-built music database with the aim to provide crossreferenced discographies of all labels and artists. As of April 14, more than 189,000

More information

Automatic Labelling of tabla signals

Automatic Labelling of tabla signals ISMIR 2003 Oct. 27th 30th 2003 Baltimore (USA) Automatic Labelling of tabla signals Olivier K. GILLET, Gaël RICHARD Introduction Exponential growth of available digital information need for Indexing and

More information

Feature-Based Analysis of Haydn String Quartets

Feature-Based Analysis of Haydn String Quartets Feature-Based Analysis of Haydn String Quartets Lawson Wong 5/5/2 Introduction When listening to multi-movement works, amateur listeners have almost certainly asked the following situation : Am I still

More information

Music Similarity and Cover Song Identification: The Case of Jazz

Music Similarity and Cover Song Identification: The Case of Jazz Music Similarity and Cover Song Identification: The Case of Jazz Simon Dixon and Peter Foster s.e.dixon@qmul.ac.uk Centre for Digital Music School of Electronic Engineering and Computer Science Queen Mary

More information

ISMIR 2008 Session 2a Music Recommendation and Organization

ISMIR 2008 Session 2a Music Recommendation and Organization A COMPARISON OF SIGNAL-BASED MUSIC RECOMMENDATION TO GENRE LABELS, COLLABORATIVE FILTERING, MUSICOLOGICAL ANALYSIS, HUMAN RECOMMENDATION, AND RANDOM BASELINE Terence Magno Cooper Union magno.nyc@gmail.com

More information

WHAT'S HOT: LINEAR POPULARITY PREDICTION FROM TV AND SOCIAL USAGE DATA Jan Neumann, Xiaodong Yu, and Mohamad Ali Torkamani Comcast Labs

WHAT'S HOT: LINEAR POPULARITY PREDICTION FROM TV AND SOCIAL USAGE DATA Jan Neumann, Xiaodong Yu, and Mohamad Ali Torkamani Comcast Labs WHAT'S HOT: LINEAR POPULARITY PREDICTION FROM TV AND SOCIAL USAGE DATA Jan Neumann, Xiaodong Yu, and Mohamad Ali Torkamani Comcast Labs Abstract Large numbers of TV channels are available to TV consumers

More information

CS229 Project Report Polyphonic Piano Transcription

CS229 Project Report Polyphonic Piano Transcription CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project

More information

CS 1674: Intro to Computer Vision. Intro to Recognition. Prof. Adriana Kovashka University of Pittsburgh October 24, 2016

CS 1674: Intro to Computer Vision. Intro to Recognition. Prof. Adriana Kovashka University of Pittsburgh October 24, 2016 CS 1674: Intro to Computer Vision Intro to Recognition Prof. Adriana Kovashka University of Pittsburgh October 24, 2016 Plan for today Examples of visual recognition problems What should we recognize?

More information

Indexing local features. Wed March 30 Prof. Kristen Grauman UT-Austin

Indexing local features. Wed March 30 Prof. Kristen Grauman UT-Austin Indexing local features Wed March 30 Prof. Kristen Grauman UT-Austin Matching local features Kristen Grauman Matching local features? Image 1 Image 2 To generate candidate matches, find patches that have

More information

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu

More information

An Introduction to Deep Image Aesthetics

An Introduction to Deep Image Aesthetics Seminar in Laboratory of Visual Intelligence and Pattern Analysis (VIPA) An Introduction to Deep Image Aesthetics Yongcheng Jing College of Computer Science and Technology Zhejiang University Zhenchuan

More information

Singer Recognition and Modeling Singer Error

Singer Recognition and Modeling Singer Error Singer Recognition and Modeling Singer Error Johan Ismael Stanford University jismael@stanford.edu Nicholas McGee Stanford University ndmcgee@stanford.edu 1. Abstract We propose a system for recognizing

More information

Multi-modal Analysis of Music: A large-scale Evaluation

Multi-modal Analysis of Music: A large-scale Evaluation Multi-modal Analysis of Music: A large-scale Evaluation Rudolf Mayer Institute of Software Technology and Interactive Systems Vienna University of Technology Vienna, Austria mayer@ifs.tuwien.ac.at Robert

More information

Composer Style Attribution

Composer Style Attribution Composer Style Attribution Jacqueline Speiser, Vishesh Gupta Introduction Josquin des Prez (1450 1521) is one of the most famous composers of the Renaissance. Despite his fame, there exists a significant

More information

Audio Feature Extraction for Corpus Analysis

Audio Feature Extraction for Corpus Analysis Audio Feature Extraction for Corpus Analysis Anja Volk Sound and Music Technology 5 Dec 2017 1 Corpus analysis What is corpus analysis study a large corpus of music for gaining insights on general trends

More information

Music Composition with RNN

Music Composition with RNN Music Composition with RNN Jason Wang Department of Statistics Stanford University zwang01@stanford.edu Abstract Music composition is an interesting problem that tests the creativity capacities of artificial

More information

Ameliorating Music Recommendation

Ameliorating Music Recommendation Ameliorating Music Recommendation Integrating Music Content, Music Context, and User Context for Improved Music Retrieval and Recommendation MoMM 2013, Dec 3 1 Why is music recommendation important? Nowadays

More information

Automatic Rhythmic Notation from Single Voice Audio Sources

Automatic Rhythmic Notation from Single Voice Audio Sources Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung

More information

Quality of Music Classification Systems: How to build the Reference?

Quality of Music Classification Systems: How to build the Reference? Quality of Music Classification Systems: How to build the Reference? Janto Skowronek, Martin F. McKinney Digital Signal Processing Philips Research Laboratories Eindhoven {janto.skowronek,martin.mckinney}@philips.com

More information

Production. Old School. New School. Personal Studio. Professional Studio

Production. Old School. New School. Personal Studio. Professional Studio Old School Production Professional Studio New School Personal Studio 1 Old School Distribution New School Large Scale Physical Cumbersome Small Scale Virtual Portable 2 Old School Critics Promotion New

More information

Melody Retrieval On The Web

Melody Retrieval On The Web Melody Retrieval On The Web Thesis proposal for the degree of Master of Science at the Massachusetts Institute of Technology M.I.T Media Laboratory Fall 2000 Thesis supervisor: Barry Vercoe Professor,

More information

Data Driven Music Understanding

Data Driven Music Understanding Data Driven Music Understanding Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Engineering, Columbia University, NY USA http://labrosa.ee.columbia.edu/ 1. Motivation:

More information

Modeling memory for melodies

Modeling memory for melodies Modeling memory for melodies Daniel Müllensiefen 1 and Christian Hennig 2 1 Musikwissenschaftliches Institut, Universität Hamburg, 20354 Hamburg, Germany 2 Department of Statistical Science, University

More information

Shades of Music. Projektarbeit

Shades of Music. Projektarbeit Shades of Music Projektarbeit Tim Langer LFE Medieninformatik 28.07.2008 Betreuer: Dominikus Baur Verantwortlicher Hochschullehrer: Prof. Dr. Andreas Butz LMU Department of Media Informatics Projektarbeit

More information

USING ARTIST SIMILARITY TO PROPAGATE SEMANTIC INFORMATION

USING ARTIST SIMILARITY TO PROPAGATE SEMANTIC INFORMATION USING ARTIST SIMILARITY TO PROPAGATE SEMANTIC INFORMATION Joon Hee Kim, Brian Tomasik, Douglas Turnbull Department of Computer Science, Swarthmore College {joonhee.kim@alum, btomasi1@alum, turnbull@cs}.swarthmore.edu

More information

INFORMATION-THEORETIC MEASURES OF MUSIC LISTENING BEHAVIOUR

INFORMATION-THEORETIC MEASURES OF MUSIC LISTENING BEHAVIOUR INFORMATION-THEORETIC MEASURES OF MUSIC LISTENING BEHAVIOUR Daniel Boland, Roderick Murray-Smith School of Computing Science, University of Glasgow, United Kingdom daniel@dcs.gla.ac.uk; roderick.murray-smith@glasgow.ac.uk

More information

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES 12th International Society for Music Information Retrieval Conference (ISMIR 2011) A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES Erdem Unal 1 Elaine Chew 2 Panayiotis Georgiou

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

Music Information Retrieval Community

Music Information Retrieval Community Music Information Retrieval Community What: Developing systems that retrieve music When: Late 1990 s to Present Where: ISMIR - conference started in 2000 Why: lots of digital music, lots of music lovers,

More information

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION ULAŞ BAĞCI AND ENGIN ERZIN arxiv:0907.3220v1 [cs.sd] 18 Jul 2009 ABSTRACT. Music genre classification is an essential tool for

More information

YOU ARE WHAT YOU LIKE INFORMATION LEAKAGE THROUGH USERS INTERESTS

YOU ARE WHAT YOU LIKE INFORMATION LEAKAGE THROUGH USERS INTERESTS NDSS Symposium 2012 YOU ARE WHAT YOU LIKE INFORMATION LEAKAGE THROUGH USERS INTERESTS Abdelberi (Beri) Chaabane, Gergely Acs, Mohamed Ali Kaafar Internet = Online Social Networks? Most visited websites:

More information

Autotagger: A Model For Predicting Social Tags from Acoustic Features on Large Music Databases

Autotagger: A Model For Predicting Social Tags from Acoustic Features on Large Music Databases Autotagger: A Model For Predicting Social Tags from Acoustic Features on Large Music Databases Thierry Bertin-Mahieux University of Montreal Montreal, CAN bertinmt@iro.umontreal.ca François Maillet University

More information

Recommending Music for Language Learning: The Problem of Singing Voice Intelligibility

Recommending Music for Language Learning: The Problem of Singing Voice Intelligibility Recommending Music for Language Learning: The Problem of Singing Voice Intelligibility Karim M. Ibrahim (M.Sc.,Nile University, Cairo, 2016) A THESIS SUBMITTED FOR THE DEGREE OF MASTER OF SCIENCE DEPARTMENT

More information

Melody classification using patterns

Melody classification using patterns Melody classification using patterns Darrell Conklin Department of Computing City University London United Kingdom conklin@city.ac.uk Abstract. A new method for symbolic music classification is proposed,

More information

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular Music Mood Sheng Xu, Albert Peyton, Ryan Bhular What is Music Mood A psychological & musical topic Human emotions conveyed in music can be comprehended from two aspects: Lyrics Music Factors that affect

More information

Music Information Retrieval with Temporal Features and Timbre

Music Information Retrieval with Temporal Features and Timbre Music Information Retrieval with Temporal Features and Timbre Angelina A. Tzacheva and Keith J. Bell University of South Carolina Upstate, Department of Informatics 800 University Way, Spartanburg, SC

More information

Part IV: Personalization, Context-awareness, and Hybrid Methods

Part IV: Personalization, Context-awareness, and Hybrid Methods RuSSIR 2013: Content- and Context-based Music Similarity and Retrieval Titelmasterformat durch Klicken bearbeiten Part IV: Personalization, Context-awareness, and Hybrid Methods Markus Schedl Peter Knees

More information

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the

More information

Structured training for large-vocabulary chord recognition. Brian McFee* & Juan Pablo Bello

Structured training for large-vocabulary chord recognition. Brian McFee* & Juan Pablo Bello Structured training for large-vocabulary chord recognition Brian McFee* & Juan Pablo Bello Small chord vocabularies Typically a supervised learning problem N C:maj C:min C#:maj C#:min D:maj D:min......

More information

An ecological approach to multimodal subjective music similarity perception

An ecological approach to multimodal subjective music similarity perception An ecological approach to multimodal subjective music similarity perception Stephan Baumann German Research Center for AI, Germany www.dfki.uni-kl.de/~baumann John Halloran Interact Lab, Department of

More information

Automatic Construction of Synthetic Musical Instruments and Performers

Automatic Construction of Synthetic Musical Instruments and Performers Ph.D. Thesis Proposal Automatic Construction of Synthetic Musical Instruments and Performers Ning Hu Carnegie Mellon University Thesis Committee Roger B. Dannenberg, Chair Michael S. Lewicki Richard M.

More information

Music Radar: A Web-based Query by Humming System

Music Radar: A Web-based Query by Humming System Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,

More information

Supplementary Note. Supplementary Table 1. Coverage in patent families with a granted. all patent. Nature Biotechnology: doi: /nbt.

Supplementary Note. Supplementary Table 1. Coverage in patent families with a granted. all patent. Nature Biotechnology: doi: /nbt. Supplementary Note Of the 100 million patent documents residing in The Lens, there are 7.6 million patent documents that contain non patent literature citations as strings of free text. These strings have

More information

jsymbolic 2: New Developments and Research Opportunities

jsymbolic 2: New Developments and Research Opportunities jsymbolic 2: New Developments and Research Opportunities Cory McKay Marianopolis College and CIRMMT Montreal, Canada 2 / 30 Topics Introduction to features (from a machine learning perspective) And how

More information

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis Fengyan Wu fengyanyy@163.com Shutao Sun stsun@cuc.edu.cn Weiyao Xue Wyxue_std@163.com Abstract Automatic extraction of

More information

Enabling editors through machine learning

Enabling editors through machine learning Meta Follow Meta is an AI company that provides academics & innovation-driven companies with powerful views of t Dec 9, 2016 9 min read Enabling editors through machine learning Examining the data science

More information

Retrieval of textual song lyrics from sung inputs

Retrieval of textual song lyrics from sung inputs INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Retrieval of textual song lyrics from sung inputs Anna M. Kruspe Fraunhofer IDMT, Ilmenau, Germany kpe@idmt.fraunhofer.de Abstract Retrieving the

More information

Improving Frame Based Automatic Laughter Detection

Improving Frame Based Automatic Laughter Detection Improving Frame Based Automatic Laughter Detection Mary Knox EE225D Class Project knoxm@eecs.berkeley.edu December 13, 2007 Abstract Laughter recognition is an underexplored area of research. My goal for

More information

HIT SONG SCIENCE IS NOT YET A SCIENCE

HIT SONG SCIENCE IS NOT YET A SCIENCE HIT SONG SCIENCE IS NOT YET A SCIENCE François Pachet Sony CSL pachet@csl.sony.fr Pierre Roy Sony CSL roy@csl.sony.fr ABSTRACT We describe a large-scale experiment aiming at validating the hypothesis that

More information