Investigating Web-Based Approaches to Revealing Prototypical Music Artists in Genre Taxonomies

Similar documents
Assigning and Visualizing Music Genres by Web-based Co-Occurrence Analysis

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

Automatically Detecting Members and Instrumentation of Music Bands via Web Content Mining

Investigating Different Term Weighting Functions for Browsing Artist-Related Web Pages by Means of Term Co-Occurrences

Music Recommendation from Song Sets

OVER the past few years, electronic music distribution

Subjective Similarity of Music: Data Collection for Individuality Analysis

Context-based Music Similarity Estimation

Enhancing Music Maps

Music Genre Classification and Variance Comparison on Number of Genres

A Survey of Music Similarity and Recommendation from Music Context Data

Outline. Why do we classify? Audio Classification

Computational Modelling of Harmony

An ecological approach to multimodal subjective music similarity perception

An Innovative Three-Dimensional User Interface for Exploring Music Collections Enriched with Meta-Information from the Web

Quality of Music Classification Systems: How to build the Reference?

The ubiquity of digital music is a characteristic

ARTIST CLASSIFICATION WITH WEB-BASED DATA

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

COMBINING FEATURES REDUCES HUBNESS IN AUDIO SIMILARITY

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video

HIDDEN MARKOV MODELS FOR SPECTRAL SIMILARITY OF SONGS. Arthur Flexer, Elias Pampalk, Gerhard Widmer

Lyrics Classification using Naive Bayes

Toward Evaluation Techniques for Music Similarity

Detecting Musical Key with Supervised Learning

MUSI-6201 Computational Music Analysis

The song remains the same: identifying versions of the same piece using tonal descriptors

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

Melody classification using patterns

Personalization in Multimodal Music Retrieval

Supervised Learning in Genre Classification

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

HIT SONG SCIENCE IS NOT YET A SCIENCE

A Computational Model for Discriminating Music Performers

A TEXT RETRIEVAL APPROACH TO CONTENT-BASED AUDIO RETRIEVAL

Measuring Musical Rhythm Similarity: Further Experiments with the Many-to-Many Minimum-Weight Matching Distance

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur

Hidden Markov Model based dance recognition

Adaptive Key Frame Selection for Efficient Video Coding

Limitations of interactive music recommendation based on audio content

Modeling memory for melodies

Tool-based Identification of Melodic Patterns in MusicXML Documents

Using Genre Classification to Make Content-based Music Recommendations

Ameliorating Music Recommendation

Effects of acoustic degradations on cover song recognition

Bibliometric analysis of the field of folksonomy research

FULL-AUTOMATIC DJ MIXING SYSTEM WITH OPTIMAL TEMPO ADJUSTMENT BASED ON MEASUREMENT FUNCTION OF USER DISCOMFORT

BIBLIOGRAPHIC DATA: A DIFFERENT ANALYSIS PERSPECTIVE. Francesca De Battisti *, Silvia Salini

Can Song Lyrics Predict Genre? Danny Diekroeger Stanford University

Music Information Retrieval with Temporal Features and Timbre

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

USING ARTIST SIMILARITY TO PROPAGATE SEMANTIC INFORMATION

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University

Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors *

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

A Framework for Segmentation of Interview Videos

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

GENDER IDENTIFICATION AND AGE ESTIMATION OF USERS BASED ON MUSIC METADATA

Mood Tracking of Radio Station Broadcasts

Automatic Laughter Detection

Chord Classification of an Audio Signal using Artificial Neural Network

Analysis of MPEG-2 Video Streams

Figures in Scientific Open Access Publications

MUSICAL MOODS: A MASS PARTICIPATION EXPERIMENT FOR AFFECTIVE CLASSIFICATION OF MUSIC

Music Information Retrieval. Juan P Bello

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset

Music Information Retrieval Community

EE373B Project Report Can we predict general public s response by studying published sales data? A Statistical and adaptive approach

Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections

CS229 Project Report Polyphonic Piano Transcription

Creating a Feature Vector to Identify Similarity between MIDI Files

COSC282 BIG DATA ANALYTICS FALL 2015 LECTURE 11 - OCT 21

ON INTER-RATER AGREEMENT IN AUDIO MUSIC SIMILARITY

EVALUATING THE IMPACT FACTOR: A CITATION STUDY FOR INFORMATION TECHNOLOGY JOURNALS

BIBLIOMETRIC REPORT. Bibliometric analysis of Mälardalen University. Final Report - updated. April 28 th, 2014

UC San Diego UC San Diego Previously Published Works

EVALUATION OF FEATURE EXTRACTORS AND PSYCHO-ACOUSTIC TRANSFORMATIONS FOR MUSIC GENRE CLASSIFICATION

Normalized Cumulative Spectral Distribution in Music

EVALUATING THE GENRE CLASSIFICATION PERFORMANCE OF LYRICAL FEATURES RELATIVE TO AUDIO, SYMBOLIC AND CULTURAL FEATURES

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

ISSN: ISO 9001:2008 Certified International Journal of Engineering Science and Innovative Technology (IJESIT) Volume 3, Issue 2, March 2014

Analysing Musical Pieces Using harmony-analyser.org Tools

m RSC Chromatographie Integration Methods Second Edition CHROMATOGRAPHY MONOGRAPHS Norman Dyson Dyson Instruments Ltd., UK

Adaptive decoding of convolutional codes

Music Genre Classification

TOWARDS CHARACTERISATION OF MUSIC VIA RHYTHMIC PATTERNS

Enabling editors through machine learning

Statistical Modeling and Retrieval of Polyphonic Music

Automatic Music Genre Classification

Syddansk Universitet. The data sharing advantage in astrophysics Dorch, Bertil F.; Drachen, Thea Marie; Ellegaard, Ole

ASSOCIATIONS BETWEEN MUSICOLOGY AND MUSIC INFORMATION RETRIEVAL

Why t? TEACHER NOTES MATH NSPIRED. Math Objectives. Vocabulary. About the Lesson

Automatic Music Clustering using Audio Attributes

Wipe Scene Change Detection in Video Sequences

Interactive Visualization for Music Rediscovery and Serendipity

Professor Birger Hjørland and associate professor Jeppe Nicolaisen hereby endorse the proposal by

Singer Traits Identification using Deep Neural Network

Subjective evaluation of common singing skills using the rank ordering method

SIGNAL + CONTEXT = BETTER CLASSIFICATION

Automatic Rhythmic Notation from Single Voice Audio Sources

Transcription:

Investigating Web-Based Approaches to Revealing Prototypical Music Artists in Genre Taxonomies Markus Schedl markus.schedl@jku.at Peter Knees peter.knees@jku.at Department of Computational Perception Johannes Kepler University Linz, Austria Austrian Research Institute for Artificial Intelligence Vienna, Austria Gerhard Widmer, gerhard.widmer@jku.at Abstract We present three general approaches to detecting prototypical entities in a given taxonomy and apply them to a music information retrieval (MIR) problem. More precisely, we try to find prototypical music artists for each genre in a given real-world taxonomy. The three approaches rely on web-based data mining techniques and derive prototypicality rankings from properties based on the number of web pages found for given entity names. We illustrate the approaches using a genre taxonomy created by music experts and present results of extensive evaluations. In detail, three evaluation approaches have been applied. First, we model and evaluate a classification task to determine accuracies. Taking the ordinal character of the prototypicality rankings into account, we further calculate rank order correlation according to Spearman and to Kendall. Interesting insights concerning the performance of the respective approaches when confronting them to the expert rankings are given. I. INTRODUCTION Prototypical entities play an essential role in cognitive processes. Thus, detecting such entities in given taxonomies is of high interest for a wide variety of fields. Some examples are given in the following list. Biology: most prominent representative of a breed Science: most prestigious researchers in a research field Music: most typical artists for a genre Prototypical entities are of vital importance for learning, e.g. []. Thus, information about them can be applied in various areas, especially in the context of information representation and visualization. In this paper, we present three methods to compute prototypicality rankings. We apply them to the problem of determining music artists that are typical representatives of a genre. Such prototypical artists can be used, for example, in music information systems and online music stores to support users in finding music more efficiently than with conventional text-based search methods. Since prototypical artists are very well-known, they can also be used to enrich visualizations of, and user interfaces to, music repositories like those presented in [], [], [4]. In this context, prototypical artists may serve as reference points to discover similar but less known artists. To measure the prototypicality of music artists in a given genre taxonomy, we make use of the world wide web. This offers the advantage of incorporating the knowledge and opinions of a large number of people. Thus, web-based data mining approaches reflect a kind of cultural knowledge that we extract and use for prototype detection. Nevertheless, web mining approaches also face some problems. The most obvious one is that they rely on the existence of web pages dealing with the topic under consideration. Therefore, our approaches can only be applied to areas for which enough information is available on the web. However, since the web is still growing rapidly, new areas of application arise every day. Another issue is to find the requested information. For example, searching for web pages related to the music artist Bush will probably result in a large number of web pages not dealing with the band, but with politics and botany. We alleviate this problem by adding music-related terms to the search query. In addition, we will present an approach that corrects prototypicality rankings that are distorted by common speech words by penalizing exorbitant popularity. Despite these challenges of web-based data mining, it has already been shown that exploiting the world wide web for MIR tasks yields promising results, e.g. [5], [6], [7]. In this paper, we investigate three different approaches to prototypical music artist detection. Two are based on co-occurrence analysis, the third one simply on the number of web pages found for the entity (the artist) under consideration. The remainder of this paper is structured as follows. In Section II, related literature is briefly discussed. In Section III, the three approaches to prototype detection are presented. Hereafter, we describe in detail the setup of the evaluations performed as well as the obtained results (Section IV). Finally, we summarize our work and point out some future directions in Section V. II. RELATED WORK Since the approaches we present in this paper are strongly related to co-occurrence analysis, we first give a short overview of this topic. In [8], playlists of radio stations and

databases of CD compilations were used to derive co-occurrences between tracks and between artists. In [5], [6], first attempts to web-based MIR were made. To this end, user collections of the music sharing service OpenNap were analyzed, co-occurrences were extracted and used to build a similarity measure based on community metadata. Co-occurrences of artist names on web pages were first investigated in [9], where the aim was to automatically retrieve related artists to a given seed artist. In [], artist co-occurrences on web pages were used to create complete similarity matrices which were evaluated for genre classification. As for the topic of automatic prototypical entity detection for music artists, in [], an approach based on co-occurrence analysis is presented. Furthermore, a visualization method that illustrates similarities between artists using the most prototypical artist of every genre as reference point is elaborated. In [], the approach of [] is refined by downranking artist names that equal common speech terms. Unlike in [], [], where prototype detection approaches for music artists are demonstrated on a quite small set of 4 artists, we use a much larger set of 995 artists here. A further weakness of the test set used in [], [] is its high number of very popular artists. That makes a serious validation of the obtained prototypicality rankings very difficult. In contrast, this paper presents the first quantitative evaluation of web-based prototypicality ranking approaches performed on a large test collection which comprises nearly well-known as well as less popular music artists and checked against expert rankings. III. METHODS We consider prototypicality as being strongly related to how often web pages related to the topic under consideration (music, in our case) refer to the entities (artists, in our case). Two of the approaches to prototypicality estimation we evaluate in this paper rely on co-occurrences of entity names (i.e. artist names) on web pages, the third one simply uses page counts. Given a list of artist names, we use Google to obtain the URLs of the top-ranked web pages containing each of the respective strings. Google was chosen since it is the most popular search engine and provides a Web API. As for the number of retrieved URLs, preliminary experiments have shown that web pages per artist seem to be a good trade-off between retrieval costs and quality of the results. Addressing the issue of finding only music-related web pages, we add additional keywords to the search query. More precisely, we use the scheme artist name +music+review since it was already successfully applied in [5], [7]. Subsequently, we crawl the top-ranked web pages of every artist and compute a cooccurrence matrix C. To this end, we successively analyze the textual content of each artist s web pages and count how many of them mention the names of the other artists. Performing this procedure for every artist yields a matrix C, where element http://www.google.com/apis c ij gives the number of web pages returned for artist i that also mention artist j. The diagonal elements c ii represent the total number of web pages retrieved for artist i, which does not necessarily equal as some pages were not accessible. This calculation method for co-occurrences differs from the one used in [], [] in that it restricts queries to Google to a minimum. Raising a query for every pair of artists would be unfeasible for a test collection of nearly items. In the following, we show how we obtain prototypicality rankings based on the co-occurrence matrix C. A. Backlink/Forward Link (BL/FL) Ratio The first method to infer prototypicality is based on an idea similar to the PageRank mechanism used by Google, where backlinks and forward links of a web page are used to measure relevancy, cf. []. Since we investigate co-occurrences rather than hyperlinks, we call any co-occurrence of artist a i and artist a j (unequal to a i ) on a web page that is known to contain artist a j a backlink of a i (from a j ). A forward link of an artist of interest a i to another artist a j, in contrast, is given by any occurrence of artist a j on a web page that is known to mention artist a i. Using this interpretation of a backlink and a forward link, we obtain the prototypicality of an artist a g i for genre g by counting for how many of the artists a g j,j i the number of backlinks of a g i (from ag j ) exceeds the number of forward links of a g i (to ag j ). The larger this count, the higher the probability for artist a g i being mentioned in the context of other artists from the same genre g and thus, the higher the prototypicality of a g i for genre g. Formally, the ranking function r(a g i ) that describes the prototypicality of an artist a g i for genre g is given by Formula, where n g is the total number of artists in genre g and bl(i, j) and f l(i, j) are functions that return a boolean value according to Formulas and respectively. n g,j i r(a g i ) = j= bl(i, j) n g,j i j= fl(i, j) bl(i, j) = fl(i, j) = { if c ij c ii < cji c jj otherwise { if c ij c ii cji c jj otherwise bl(i, j) returns the value if artist a g i has more backlinks from artist a g j (relative to the total number of web pages retrieved for a g j ) than forward links to artist ag j (relative to the total number of web pages retrieved for a g i ). fl(i, j) is defined analogously. We call r(a g i ) the backlink/forward link (bl/fl) ratio of artist r(a g i ) since it counts how often the relative frequency of backlinks for a g i exceeds the relative frequency of its forward links and relates these two counts. () () ()

B. BL/FL Ratio with Popularity Penalization A drawback of the BL/FL approach is that artist names that equal common speech terms, e.g. Kiss, Prince, or Hole, are always top-ranked. The reason for this is that such words frequently occur on arbitrary web pages, regardless of their relatedness to the topic. Therefore, they create a lot of unjustified backlinks for artists with the respective names, what could distort the prototypicality ranking. To avoid such distortions, we introduce a mechanism that basically pursues the idea of the commonly used information retrieval approach tf idf (term frequency inverse document frequency), cf. [4]. In this approach, the importance of a term is higher if it occurs frequently (high tf). On the other hand, a term is penalized if it occurs in many documents and hence, does not contain much relevant information (high df leads to low idf). In the modified BL/FL approach, we adapt this principle to penalize the prototypicality of an artist if it is high over all genres (following the naming scheme of tf idf, we call this approach gp iop for genre prototypicality inverse overall prototypicality). This is reasonable since even very popular and important artists are unlikely to be prototypes for all genres. To emphasize this, we take a look at the 4-artist-set used in []. Those artists whose names equal common speech words yield by far the highest overall bl/fl ratios, i.e. Bush (/), Prince (/), Kiss (/), Madonna (/), and Nirvana (8/5). Incorporating information about overall prototypicality, the second ranking function we propose is shown in Formula 4. The used penalization term is given by Formula 5, where n is the total number of artists in the collection. The functions bl(i, j) and fl(i, j) are defined as in Formulas and. norm is a function that shifts all values in the positive range by subtracting the smallest (non negative infinite) value, replaces infinite numbers by, and normalizes the values by division by the maximum (in the order mentioned). r(a g i ) = n g,j i j= bl(i, j) ng,j i j= fl(i, j) + penalty(ag i ) (4) penalty(a i ) = norm C. Simple Page Counts ( log n,j i ) j= fl(i, j) n,j i j= bl(i, j) + The third approach we investigate is very straightforward. We simply query Google using the scheme artist name + genre name and retrieve the page count value, i.e. the number of found web pages returned for the query. Since in our test collection (cf. Section IV-A), every artist is assigned a single genre, we need to perform this step only once for every artist. For each genre, we then rank its artists according to the page counts to obtain a popularity ranking. Since prototypicality is strongly related to popularity, we simply use this as a prototypicality ranking. (5) TABLE I THE DISTRIBUTION OF THE TEST SET AMONG THE TIERS GIVEN BY THE AMG. THE ABSOLUTE NUMBER OF ARTISTS ARE GIVEN FOR EVERY GENRE AS WELL AS THE RELATIVE FREQUENCIES AMONG THE AMG TIERS. absolute relative AMG tier AMG tier Genre Blues 7 95 56 88..5. Electronica 5 68 95.6.7. Reggae 8 6.47.5. Jazz 9 4 8 8..49.9 Folk 44 6 8.54.44. Heavy Metal 4 59 98 7.5..7 RnB 47 8 7..4.6 Country 9 75 46.6.54. Rap 8 4.8.. Total 6 9 7 995.8.46.6 IV. EVALUATION Evaluating the quality of the prototypicality ranking approaches is a difficult task for various reasons. First, prototypicality is influenced by personal taste and cultural opinions. Thus, if we had asked a number of people which artists they considered prototypical for a certain genre, they might have named largely their favorites (maybe also those from their own country of origin). Another issue is that prototypical artists may also change over time. For example, formerly unknown artists may become very popular overnight. This raises the question in which way time should be considered in a prototypicality ranking. Should artists be downranked because they were very popular for a genre years ago? Since our aim was to perform evaluations on a large artist set, conducting a web survey to obtain a ground truth against which the approaches are evaluated was out of the question as this would have included ranking every artist with respect to all other artists of the respective genre. Alternatively, presenting only a subset of artists would have resulted in incomplete rankings. A. Test Collection and Ground Truth We finally decided to use a test collection of 995 artists from nine common genres, which were extracted from the popular music information system All Music Guide (AMG). The collection comprises very popular as well as less known artists. A list of the artists and their assigned genres can be downloaded from http://www.cp.jku.at/people/schedl/music/ C995a artists genres.txt. As ground truth against which we evaluated the prototypicality ranking approaches, we used the tiers given by the AMG. The artists of each genre are usually clustered in three tiers according to their importance for the respective genre which is defined by experts: The Tier value indicates a ranking of the choices in the list according to the AMG Editors determination http://www.allmusic.com

6.8.5 BL/FL.6 6.. 6.4 4. 66.4 6..7.5 BL/FL Penalized 4. 6..8 5.6 5 65.7 56.4 4.7. Page Counts 7.5 59.4.5 6. 5.9 64. Fig.. Confusion matrices for the classification task for each of the three approaches. The columns indicate the tiers to which the approaches map their rankings, the rows indicate the actual AMG tiers. The values are given in percent. 59.5.5 8.9 Blues 9.7 64. 4..8 5. 5 5.6 9..5 Jazz 4 6 6.5 5.4 9.8 6 68.8 Electronica 88. 6.6 4.7 Folk 6.4 55.6.8 64.. Reggae 5.7 68.8 64. 5. Heavy Metal.4 4.7 6. 4. 54. 8.8 estimation on a classification task, Spearman s rank-order correlation, and Kendall s tau. ) Classification Accuracy: To gain an overall impression of the performance of the investigated approaches, we interpret the AMG tiers as classes and simulate a classification task using our prototypicality ratings as classifiers. To this end, we map the rankings obtained by the prototypicality detection approaches to the ones given by the AMG tiers and determine the concordances. More precisely, given that our prototypicality algorithm has produced a specific ranking R of the artists of a genre and assuming the three AMG tiers for this genre contain n, n, and n artists, respectively, we assign the first n elements of R to tier, the next n to tier, and the last n to tier. We can then view these assignments as classification decisions and calculate classification accuracy values. ) Spearman s Rank-Order Correlation: To measure the correlation between the ground truth ranking of the AMG and the rankings obtained with our prototypicality detection approaches, we use the well-established Spearman s rankorder correlation coefficient, e.g. [5]. Since the rankings by the AMG are strongly tied, using the standard formula would spuriously inflate the correlation values. Therefore, we apply a tie-corrected version according to [6] as shown in Formulas 6 9, where r Sc gives the rank-order correlation coefficient, n is the total number of ranked data items, X and Y represent the two rankings under consideration, s X and s Y are the numbers of sets of ties present in X and Y respectively, and t Xi and t Yi are the numbers of X and Y scores that are tied for a given rank. RnB 6.7.4 4.9 Country 5. 4.6 5. Rap 78.8. x + y d r Sc = x (6) y 9.5.7 56. 4. 4.4 6 4.4 6.4 45. 4. 54.7 87.5.5 Fig.. Confusion matrices for the classification task shown for every genre. The BL/FL approach with penalization of exorbitant popularity was used. The values are given in percent. of importance, quality, and relevance to the selected category. The composition of the test collection can be seen in Table I, where for each genre and each tier, the absolute and relative numbers of included artists are shown. B. Evaluation Methods We investigated the quality of the prototypicality rankings using three different evaluation methods simple accuracy http://www.allmusic.com/cg/amg.dll?p=amg&sql=: amg/info pages/a siteglossary.html x = n n T X T X = s X i= (t X i t Xi ) T Y = d = y = n n T Y s Y i= (7) (t Y i t Yi ) (8) n (X i Y i ) (9) i= ) Kendall s Tau: We further calculated rank-order correlations according to Kendall s τ. Again, we used the tiecorrected version which is elaborated, for example, in [6]. However, since the Kendall s τ values yielded no new insights when compared to the Spearman s rank-order correlation values, we do not elaborate on them here. C. Results and Discussion The overall results of the classification task are depicted in Figure, where a confusion matrix for each of the three investigated approaches is shown. It can be seen that the BL/FLbased approaches, in general, perform better than the Simple Page Counts approach, especially for predicting first-tierartists. Comparing the BL/FL to the BL/FL Penalized approach

TABLE II THE TEN TOP-RANKED ARTISTS FOR THE GENRES HEAVY METAL AND FOLK FOR EACH OF THE THREE APPROACHES. Heavy Metal BL/FL BL/FL Penalized Page Counts Death Metallica Metallica Europe AC/DC Death Tool Black Sabbath Kiss Metallica Death Tool Kiss Led Zeppelin Extreme Filter Riot Europe AC/DC Iron Maiden Trouble Led Zeppelin Judas Priest Iron Maiden Black Sabbath Slayer Filter Alice Cooper Marilyn Manson Rainbow Folk BL/FL BL/FL Penalized Page Counts Woody Guthrie Woody Guthrie Woody Guthrie Joan Baez Joan Baez Joan Baez Lucinda Williams Judy Collins Pete Seeger Pete Seeger Pete Seeger Lucinda Williams Judy Collins Lucinda Williams Arlo Guthrie Leadbelly Doc Watson Doc Watson Doc Watson Leadbelly Judy Collins Townes Van Zandt Phil Ochs Alan Lomax Gordon Lightfoot Gordon Lightfoot Leadbelly Phil Ochs Townes Van Zandt Gordon Lightfoot reveals slightly significant better results for the version using penalization of exorbitant popularity when predicting first-tierartists, but slightly worse results for predicting tiers two and three. This becomes particularly obvious when considering Table II, where the top-ranked artists for the genres Heavy Metal and Folk are shown. In this table, the penalization of artists whose names equal common speech terms can be seen very well when regarding the results for the genre Heavy Metal. In fact, the BL/FL approach (and also the Simple Page Counts approach) top-ranks artists like Death, Europe, Tool, Kiss, and Filter. The same artists are considerably downranked by the BL/FL Penalized approach. In contrast, the rankings for the genre Folk remain almost unmodified since the artists of this genre are usually known by their real name, cf. Table II. To get an impression of the impact of the genre on the quality of the results, Figure shows a confusion matrix for each of the nine genres for the best-performing BL/FL Penalized approach. It can be seen that the overall results for the genre Electronica are by far the best (weighted with the number of artists in every tier, we obtain an accuracy of 8%, which is % above the baseline, cf. Table I). The remarkable wrong confusion for in genre Folk is due to only one single artist which is incorrectly classified as belonging to tier instead of and therefore, does not considerably influence the overall performance of the approach. Comparing Table III to Table I (for the baseline) reveals that the overall accuracies, except those for the genre Rap, considerably exceed the baseline. In the case of Electronica, Reggae, Jazz, and RnB they are even between % and more than % above the baseline. In contrast, the results for the genre Rap are very TABLE III OVERALL GENRE-SPECIFIC ACCURACIES FOR THE THREE APPROACHES, OBTAINED BY WEIGHTING THE GENRE-SPECIFIC ACCURACIES GIVEN BY FIGURE WITH THE NUMBER OF ARTISTS IN EVERY TIER. acc DENOTES THE ACCURACY THAT THE EVALUATED RANKING APPROACH MAPS AN ARTIST EXACTLY TO THE SAME AMG TIER IT SHOULD FALL INTO ACCORDING TO AMG S RANKING. acc DENOTES THE ACCURACY WHEN DEVIATIONS OF UP TO ONE TIER ARE ALLOWED. BL/FL BL/FL Pen Page Counts Genre acc acc acc acc acc acc Blues.59.95.59.95.57.94 Electronica.8..8..7.98 Reggae.67..67..7. Jazz.6.98.6.98.6.99 Folk.6.99.6.99.6.99 Heavy Metal.7.98.7.99.67.98 RnB.6.95.6.96.5.9 Country.59.99.58.99.59.99 Rap.66..66..66. poor. Taking a closer look at the AMG tiers let us assume that this may be caused by subjective and time-dependent opinions of the experts at AMG since very popular Rap artists like Eminem and Snoop Dogg are assigned to the second tier, whereas many artists that were very popular some years ago are still assigned to tier. As for the results of the correlation analysis, in Table IV, the Spearman s rank-order correlations between the ground truth ranking given by the AMG and the rankings obtained with our prototypicality detection approaches are shown for every genre. For all genres except Rap, the rank-order correlation coefficient is at least., for the genres Electronica, Jazz, Heavy Metal, and RnB it is about.5, and for Country it almost reaches.6. We also performed a significance test for the results of the Spearman s rank-order correlation according to [6]. Since we do not have any previous knowledge to predict the direction of the difference, we used a two-tailed test with a significance interval of 95%. We proved significance for all obtained correlations, despite those for the genre Rap. For this genre, we obtained a weak negative correlation, which was not stated significant. V. CONCLUSIONS AND FUTURE WORK In this paper, we presented and investigated three webbased approaches to ranking entities according to their prototypicality and demonstrated them on the problem of finding prototypical music artists in a given genre taxonomy. Two of the three approaches rely on co-occurrence analysis of entity names on web pages, the third one simply uses page counts returned by Google when searching for the entity name (artist) together with the corresponding category (genre). We used a test collection of nearly artists from nine genres for evaluation. As ground truth, we relied on expert opinions taken from the music information system All Music Guide. We assessed the quality of the prototypicality rankings using three different evaluation methods accuracy estimation

TABLE IV SPEARMAN S RANK-ORDER CORRELATIONS BETWEEN THE GROUND TRUTH RANKING BY AMG AND THE RANKINGS OBTAINED WITH THE PROTOTYPICALITY RANKING APPROACHES. Genre BL/FL BL/FL Pen Page Counts Blues.4.9.5 Electronica.45.48.7 Reggae... Jazz.49.49.5 Folk..4. Heavy Metal.48.48.4 RnB.55.55.4 Country.58.59.58 Rap -. -. -.5 Mean.7.8.5 on a classification task, Spearman s rank-order correlation, and Kendall s tau. To summarize the results, we have shown that the approaches based on Backlink/Forward Link (BL/FL) Ratios perform better than the Simple Page Counts approach. We further showed that penalization of exorbitant popularity improves results in some cases, e.g. for the genre Heavy Metal, where many artist names equal common speech terms. However, for genres like Folk or Jazz, where most artists use their real names, or at least pseudonyms that sound like real names, no significant improvements could be made out when using the BL/FL approach with penalization of exorbitant popularity. As for future work, we plan to create a user interface that incorporates information about prototypical music artists. Our aim is to provide the user with reference points (the prototypical artists), so that he/she will be able to browse music repositories more efficiently than with conventional user interfaces. We further intend to apply our prototype detection approaches to domains other than music. We are currently investigating web-based approaches to determining the period of activity of an artist. This information could help refining prototypicality by weighting an artist according to his/her principal years of musical activity. [] D. Gleich, M. Rasmussen, K. Lang, and L. Zhukov, The World of Music: SDP Layout of High Dimensional Data, in Proceedings of the IEEE Symposium on Information Visualization 5, Minneapolis, Minnesota, USA, October 5. [4] M. Goto and T. Goto, Musicream: New Music Playback Interface for Streaming, Sticking, Sorting, and Recalling Musical Pieces, in Proceedings of the Sixth International Conference on Music Information Retrieval (ISMIR 5), London, UK, September 5. [5] B. Whitman and S. Lawrence, Inferring Descriptions and Similarity for Music from Community Metadata, in Proceedings of the International Computer Music Conference, Goeteborg, Sweden, September, pp. 59 598. [6] D. P. W. Ellis, B. Withman, A. Berenzweig, and S. Lawrence, The Quest for Ground Truth in Musical Artist Similarity, in Proceedings of the rd International Symposium on Music Information Retrieval (ISMIR ), Paris, France,. [7] P. Knees, E. Pampalk, and G. Widmer, Artist Classification with Webbased Data, in Proceedings of the 5th International Symposium on Music Information Retrieval (ISMIR 4), Barcelona, Spain, October 4, pp. 57 54. [8] F. Pachet, G. Westerman, and D. Laigre, Musical Data Mining for Electronic Music Distribution, in Proceedings of the st WedelMusic Conference,. [9] M. Zadel and I. Fujinaga, Web Services for Music Information Retrieval, in Proceedings of the 5th International Symposium on Music Information Retrieval (ISMIR 4), Barcelona, Spain, October 4. [] M. Schedl, P. Knees, and G. Widmer, A Web-Based Approach to Assessing Artist Similarity using Co-Occurrences, in Proceedings of the Fourth International Workshop on Content-Based Multimedia Indexing (CBMI 5), Riga, Latvia, June 5. [], Discovering and Visualizing Prototypical Artists by Web-based Co-Occurrence Analysis, in Proceedings of the Sixth International Conference on Music Information Retrieval (ISMIR 5), London, UK, September 5. [], Improving Prototypical Artist Detection by Penalizing Exorbitant Popularity, in Proceedings of the Third International Symposium on Computer Music Modeling and Retrieval (CMMR 5), Pisa, Italy, September 5. [] L. Page, S. Brin, R. Motwani, and T. Winograd, The PageRank Citation Ranking: Bringing Order to the Web, in Proceedings of the Annual Meeting of the American Society for Information Science (ASIS 98), January 998, pp. 6 7. [4] G. Salton and C. Buckley, Term-weighting Approaches in Automatic Text Retrieval, Information Processing and Management, vol. 4, no. 5, pp. 5 5, 988. [5] R. V. Hogg, A. Craig, and J. W. McKean, Introduction to Mathematical Statistics, 6th ed. Prentice Hall, June 4. [6] D. J. Sheskin, Handbook of Parametric and Nonparametric Statistical Procedures, rd ed. Boca Raton, London, New York, Washington, D.C.: Chapman and Hall/CRC, 4. ACKNOWLEDGMENTS This research is supported by the Austrian Fonds zur Förderung der Wissenschaftlichen Forschung (FWF) under project number L-N4 and by the Vienna Science and Technology Fund (WWTF) under project number CI (Interfaces to Music). The Austrian Research Institute for Artificial Intelligence acknowledges financial support by the Austrian ministries BMBWK and BMVIT. REFERENCES [] R. T. Kellogg, Cognitive Psychology, nd ed. Thousand Oaks, California, USA: Sage Publications, Inc.,. [] E. Pampalk, S. Dixon, and G. Widmer, Exploring Music Collections by Browsing Different Views, in Proceedings of the Fourth International Conference on Music Information Retrieval (ISMIR ), Washington, D.C., USA, October.