Pattern Based Melody Matching Approach to Music Information Retrieval

Size: px
Start display at page:

Download "Pattern Based Melody Matching Approach to Music Information Retrieval"

Transcription

1 Pattern Based Melody Matching Approach to Music Information Retrieval 1 D.Vikram and 2 M.Shashi 1,2 Department of CSSE, College of Engineering, Andhra University, India 1 daravikram@yahoo.co.in, 2 smogalla2000@yahoo.com ABSTRACT Digitization of music and advancements in information technology for sharing information on World Wide Web paved way for its availability in enormous quantities anywhere any time. Rather than retrieving annotated music in response to query given in terms of Metadata such as name of the composer/singer, genre etc modern researchers are challenged towards content based music information retrieval systems (CBMIR). CBMIR systems differ while representing the main melody either as a note sequence or as an analog acoustic signal; the note sequence representation is explored in this research work. Based on the observation that repeating patterns of the note sequences representing the main melody capture the essence of the music object, this research work developed a framework to investigate the feasibility and effectiveness of pattern based melody matching approach to music information retrieval. Experimentation is conducted on a real world dataset of musical objects belonging to South Indian classical music and the performance of the framework is estimated in terms of Mean Reciprocal Ranking. Keywords CBMIR, DTW, MIDI, MRR, QBH. 1 Introduction Popular music objects are digitized and hence are available anywhere and anytime. Due to the huge volume of available music objects, retrieval of a specific object based on the user request is becoming more and more complex. Existing Music Information Retrieval Systems [2] accept user request expressed as a logical combination of metadata items like title, composer name, genre, singers name, movie name in terms of which the music databases are indexed and maintained. Such representations of musical objects cannot support content based retrieval of music objects as they are limited to matching based on metadata. However the researchers are challenged by the demand to retrieve music in response to query given in terms of content rather than metadata leading to content based representation of music objects. In the context of information retrieval the content of a music object is often captured by its main melody. The main melody of a music object is a series of musical notes (semi tones in an octave) played whose frequencies are represented in terms of MIDI note numbers. The musical frequencies are divided into 11 octaves numbered from -1 to 9 each containing 12 semitones named C C# D D# E F F# G G# A A# B. The range of frequencies [11] encompassed by an octave doubles as you go for higher octaves successively. The name of the note reflects the semitone and the octave like A4 represents semitone A in 4th octave and it has a distinct MIDI note number 60. The MIDI note number is determined by the DOI: /tmlai Publication Date: 1 st January, 2016 URL:

2 T r a n s a c t i o n s o n M a c h i n e L e a r n i n g a n d A r t i f i c i a l I n t e l l i g e n c e V o l u m e 4, I s s u e 6, D e c following formula that transforms hummed tune frequency into the representation of MIDI values (semitones): MIDI Value = 69 + [12 X log2 ( freq 440 )] Where freq is the frequency of hummed note and the operator [ ] calculates the nearest integer value, 12 leads to the classic dodecaphonic musical scale, and 69 is the MIDI note number that corresponds to central A with pitch equal to 440 Hz. By convention middle C (MIDI note Number 60) is C4. A MIDI note number of 69 is used for A 440 tuning, that is the note A above middle C. Content Based Music Information Retrieval (CBMIR) Systems rely on the frequent patterns extracted from the main melody for identifying the music objects in response to query given in terms of content referred to as "Content Query". Such a retrieval scenario belongs to query-by-example paradigm in which deals two types of content queries namely Query by Patterns and Query by Humming. In the context of Query by Patterns the CBMIR finds the music objects that contain repeated occurrences of the query patterns and ranks them based on the prominence of query patterns in the music objects. The second type of content query called QBH [3,6] refers to accepting user s music requirement in the form of audio input representing a part of the required song either sung or hummed at the user interface. The main melody of the query should be represented as a note sequence and segmented into phrases before proceeding for matching. Irrespective of the way the query is received, a CBMIR have to deal with approximate matching of patterns/phrases [4] while building the database as well as responding to a query. Hence modern MIRs have to adopt intelligent pattern extraction and matching techniques in support of content based music information retrieval [7,8,14]. In this paper the authors proposed an effective framework for CBMIR is synthesized applying various concepts and techniques of sequential mining, approximate pattern mining, DTW and other relevant concepts of Content Based Information Retrieval. The Mean Reciprocal Ranking (MRR) [1] is found suitable to assess the performance of the system for different inputs. Section 2 presents the recent research outcomes on various issues of CBMIR. Section 3 discusses the experimentation and results followed by conclusion. 2 Related Work Anssi Klapuri [5] discusses various methods for dividing musical audio signal into shorter sequences if they occur repeatedly applying pattern segmentation and clustering approaches. A Self Distance Matrix (SDM) is often used for audio based analysis of segments of music expressed as acoustic signals. Repeated sequences occur as off-diagonal-strips of SDM and hence they are recognized. Rifki Afinai Putri [1] et al discussed the representation of main melody as note sequence to transform 250 Indonesian pop songs in MIDI format and used DTW for matching the query with the whole song to take care of imperfect queries. The Query by Humming System could retrieve appropriate songs provided there is no difference in sruthi (key) of query and the song. Alexios [9] et al proposed a subsequence matching algorithm for identifying a matching subsequence for a given query from a large database of music objects. The paper highlights the effectiveness of the method for a query by humming application. C o p y r i g h t S o c i e t y f o r S c i e n c e a n d E d u c a t i o n U n i t e d K i n g d o m 79

3 D.Vikram and M.Shashi; Pattern Based Melody Matching Approach to Music Information Retrieval. Transactions on Machine Learning and Artificial Intelligence, Volume 4 No 6 December (2016); pp: Methodology In this research a framework is developed for Content Based Music Information Retrieval Systems which consists of five modules namely main melody representation as note sequences, finding approximate repeating patterns, and pattern based indexing of music objects, segmentation of the input query and pattern matching and ranking based on the relevance of music objects. Figure.1 depicts the components of the proposed framework. Figure.1: Framework of Proposed Content Based Music Information Retrieval System Representation of a music object as a note sequence, after converting it into MIDI file format and extracting the main melody was discussed in detailed in [10]. Based on the user specified or predetermined thresholds on number of repeats of a pattern and tolerance of approximate matching, the module 2 of the framework extracts approximate sequential patterns by assembling the exact repeating patterns possibly separated by a tolerable gap. The details of pattern extraction from the music objects are discussed in [11]. Module 3 represents the music objects in terms of constituent repeating patterns along with their prominences to construct a pattern base of inverted lists which are used for estimating the matching scores of music objects. Module 4 preprocesses query to extract query patterns and finds matching patterns from the pattern base. The matching scores of music objects to the query are estimated to produce a ranked list of music objects in module 5. This paper discusses the implementation of the last three modules of the framework. This framework adopts proven techniques of document retrieval for content based music information retrieval by appropriately tuning the required metrics and methods. Analogous to the words in the document text, repeating patterns of the note sequence representing the main melody capture the essence of the music object and hence the proposed framework explores pattern based melody matching approach to music information retrieval. URL: 80

4 T r a n s a c t i o n s o n M a c h i n e L e a r n i n g a n d A r t i f i c i a l I n t e l l i g e n c e V o l u m e 4, I s s u e 6, D e c The note sequence representing the main melody of a music object needs to be transformed into a list of constituent repeating patterns which calls for complex asynchronous periodic pattern mining. Similar to the concept of document representation as a term vector, each music object is represented in terms of constituent repeating patterns along with their prominence estimated using a variant of tf.idf. It is essential to allow a certain level of tolerance while matching musical patterns as against exact matching of words/terms used in the context of document retrieval. Hence approximate pattern matching techniques are implemented using Dynamic Time Warping (DTW) algorithm for matching query patterns with the repeating patterns of the vocabulary developed for the corpus. In the context of query by example the cardinality of the result set for a query is much smaller, if not one, compared to the document retrieval scenario. Hence the performance metrics for CBMIR systems differ from those of document retrieval systems. The first three modules of the framework prepare the database of music objects for intelligent content based retrieval which is performed off-line and is represented in dark colored blocks in Figure.1. Module 3 receives the set of repeating sequential patterns along with their frequency of occurrence in a song as the outcome of module 2 which applied the approximate sequential pattern mining algorithm on the note sequence representing the song/music object. Module 3 collects the repeating approximate patterns occurred in all songs/ musical objects of the corpus and maintains them as vocabulary of the corpus. It counts sf(p), the number of songs covering each pattern, p, of the vocabulary to estimate inverse song frequency similar to the concept of inverse document frequency. The prominence of a pattern, p, in a music object s is estimated as a product of its proportionate frequency in the song and the inverse song frequency as given below: Prominence (p, s) = frequency of p in s s Loge( N sf(p) ) Each musical object is represented as a vector of patterns and their prominences analogous to the concept of document representation as a term vector. For fast retrieval in response to a query, a pattern base is created as an array of vectors to maintain inverted list of songs covering each pattern. Each element of the array contains the list of ordered pairs of the form <song_id, prominence> representing the list of songs containing the pattern along with the prominence of the pattern in the song. This Pattern Base contains 1 to n vectors and the i th vector of the array is the list of songs containing i th pattern. The last three columns of Table 2 depict a portion of the pattern base. 3.1 Query Processing The query consists of fragments of a song or humming that corresponds to a part of a song which is available in the database of music objects. Each query is converted into a note sequence before extracting repeating patterns from it referred to as query patterns in module 4. Each query pattern will be compared with the patterns in the pattern base to identify matching patterns, if exists; the matching scores of the songs in the corresponding inverted list of each matching pattern will be incremented by the prominence of the pattern given in the inverted list. This process continues for each query pattern incrementing the matching scores of appropriate songs as guided by the inverted list. C o p y r i g h t S o c i e t y f o r S c i e n c e a n d E d u c a t i o n U n i t e d K i n g d o m 81

5 D.Vikram and M.Shashi; Pattern Based Melody Matching Approach to Music Information Retrieval. Transactions on Machine Learning and Artificial Intelligence, Volume 4 No 6 December (2016); pp: In the context of CBMIR, the query given by the user as a piece of main melody may not be perfect due to altered gamakas. Hence the query patterns are expected to contain additions, deletions or substitutions of musical notes from the corresponding patterns existing in the pattern base. Dynamic Time Warping (DTW) algorithm is found suitable [1,12] for identifying matching patterns for a given query pattern as it considers possible insertion, deletion and replacement of symbols contained in symbolic sequences of different lengths. DTW algorithm finds the distance between two symbolic sequences by the best possible alignment through maximal matching. Let X and Y be symbolic sequences of length n and m respectively. D(i,j) represents the distance between subsequences representing i and j long prefixes of X and Y respectively. d(i,j) is the distance between the pair of symbols i and j. DTW algorithm is described below. Input: Given two sequences X=(x 1,x 2.x i x n) and Y=(y 1, y 2.y j y m) Output: Distance between X and Y, D(n,m) 3.2 Algorithm 1 for Estimating the distance between query sequence and data sequence 1. Estimate the distance between every pair of distinct symbols constituting X and Y 2. Create distance matrix with rows corresponding to symbols of X and columns corresponding to symbols of Y so that d(i, j) represents the local distance between x i and y j for i [1, n] and j [1, m] 3. Initialize the matrix D: D(1, 1) = d(1, 1) for i [2, n], D(i, 1) = D(i 1, 1) + d(i, 1) for j [2, m], D(1, j) = D(1, j 1) + d(1, j) 4. Compute the remaining elements of matrix D 5. Return D(n, m). for i from 2 to n for j from 2 to m D(i,j)=Min{ D(i-1,j), D(i-1,j-1), D(i,j-1)}+ d(i,j) Example: Query Sequence: Consider a note sequence. Sequence : URL: 82

6 T r a n s a c t i o n s o n M a c h i n e L e a r n i n g a n d A r t i f i c i a l I n t e l l i g e n c e V o l u m e 4, I s s u e 6, D e c Figure.2: Matching Query sequence and the Sequence in Pattern Base DTW algorithm is used to find the distance between query pattern and the patterns in the pattern base. The patterns whose distances from the query pattern are less than a given threshold are selected as matching patterns. The query pattern is skipped if the distance to the closest pattern is more than the threshold. 3.3 Upon Receiving a Query: Module 4 receives the query which will be processed in a similar manner like any music object described in module 1 to extract query patterns. Each query pattern is matched with the patterns in the pattern base allowing certain level of tolerance (based on a threshold) using DTW algorithm. 3.4 Algorithm 2 for query processing online: Input: Query as a note sequence with tolerance threshold for matching Step 1: Query pattern extraction: Extract patterns in the query either by segmentation or by approximate pattern extraction algorithm depending on the context. Step 2: Create an array of matching scores whose size is equal to the number of songs in the corpus and initialize its elements to zero. # The i th element in this array maintains the matching score of the i th song to the query. Repeat Steps 3 and 4 for each query pattern. Step 3: Apply DTW on successive elements of the pattern base to find approximately matching patterns for a given query pattern. Step 4: Repeat for each matching pattern, j Process the ordered pairs contained in the j th vector of the Pattern Base one by one as detailed below: Add p, the prominence of j th pattern in song, s, as mentioned in the ordered pairs <s, p> to the s th entry of the array of matching scores. Outcome: Array of scores incremented to finally reflect the matching scores of various songs for a given query. C o p y r i g h t S o c i e t y f o r S c i e n c e a n d E d u c a t i o n U n i t e d K i n g d o m 83

7 D.Vikram and M.Shashi; Pattern Based Melody Matching Approach to Music Information Retrieval. Transactions on Machine Learning and Artificial Intelligence, Volume 4 No 6 December (2016); pp: Step 5: Sort the songs based on the matching scores to generate ranking list of songs for the query. The results of melody matching and ranking is illustrated with an example in terms of Table 1, 2 and 3. The pattern base contains musical note sequences repeatedly occurred in various songs along with their prominence. In the example the given query contained three patterns as shown in Table 1. Each query pattern has one or more matching patterns found in the pattern base by applying DTW algorithm. Table 2 depicts for each query pattern the list of matching patterns along with the song(s) in which it occurred with the prominence given in the last column. As described in the algorithm step 4 the matching scores of pertinent songs are calculated and shown in Table3 as a ranking list of songs in response to the query. 1. List of Query Patterns Identified Table.1 2. For the given query patterns details of matching patterns, songs covering patterns, and prominence: Table Ranked List of Songs: Table. 3 URL: 84

8 T r a n s a c t i o n s o n M a c h i n e L e a r n i n g a n d A r t i f i c i a l I n t e l l i g e n c e V o l u m e 4, I s s u e 6, D e c Dataset 4 Experiments and Results Raga Surabhi [13] provides a collection of 700 (approximately) songs/musical objects belonging to various ragas of Carnatic music (South Indian Classical Music) including ragaalapana, songs and signature of each raga in mp3 format. Each song is represented as a note sequence during the preprocessing steps by converting wave files into strings. The length of the songs varies extensively resulting in a range of 238 to 6144 long note sequences repeating patterns were extracted from these songs and a pattern base is maintained with inverted lists. 50 songs were selected for test set and query patterns from these test songs were given as input for retrieving ranked list of songs containing matching patterns to the query patterns. Mean Reciprocal Ranking (MRR) is a metric suitable for finding the accuracy of query-by-example based CBMIR systems aiming at a single song to be retrieved in response to a given query. Let S(q i) be the song aimed at and hence containing the query pattern(s) constituting the query q i and let Q be number of queries used in the testing process, then MRR = 1 Q i 1 Rank(Sqi) Experiments were conducted on the test set to estimate MRR for each query length by extending the queries (aiming at each song of the test set) from 1 to 5 query patterns. It was observed that MRR tends to 1 representing almost absolute accuracy even for short queries containing a small number of query patterns. Since a pattern may occur in more than a single song with different prominence like the pattern 2 and 3 given in Table 2 the accuracy of ranking order reflected in MRR improves by adding additional query patterns belonging to the song aimed at. For example the prominence of the pattern 2 D#3D#3C3C3D3D3C3C3C3C3 is comparatively less in the song aimed at (Aadikondar) and hence reduces the reciprocal ranking and same is the case with pattern 3; however as the query contained three patterns the song aimed at could achieve highest matching score and there by placed in first position in the ranked list for the query of length 3. Figure 3 depicts the graphical representation of MRR versus length of the query in terms of number of query patterns. Figure 3: Precision of Music Information Retrieval for different query lengths. C o p y r i g h t S o c i e t y f o r S c i e n c e a n d E d u c a t i o n U n i t e d K i n g d o m 85

9 D.Vikram and M.Shashi; Pattern Based Melody Matching Approach to Music Information Retrieval. Transactions on Machine Learning and Artificial Intelligence, Volume 4 No 6 December (2016); pp: Conclusion A framework for content based music information retrieval system that retrieves a ranked list of musical objects from a database in response to query by example is developed. The framework is implemented in terms of five modules for representing the main melody as a note sequence, extracting sequential patterns from the main melody, preparing pattern base of inverted list of musical objects that cover each pattern, query processing and identification of matching patterns using DTW algorithm and finally to rank the music objects after estimating their matching scores to the query expressed as one or more patterns. The established techniques for content based document retrieval were appropriately adopted to build the framework which is tested on real dataset, Raga Surabhi. The framework is found to be very effective as it resulted in very high Mean Reciprocal Ranking (almost one) even for short queries consisting of two to three query patterns. ACKNOWLEDGEMENTS This work was supported by the Council of Scientific & Industrial Research (CSIR) and Andhra University, Visakhapatnam, Andhra Pradesh, India. REFERENCES [1] Rifki Afina Putri, Dessi Puji Lestari, Music Information Retrieval using Query-by-Humming based on the Dynamic Time Warping The 5th International Conference on Electrical Engineering and Informatics 2015 August 10-11, 2015, Bali, Indonesia. [2] M. Müller, Information Retrieval for Music and Motion, New York: Springer, [3] Z. W. Ras and A. Wieczorkowska, Advances in Music Information Retrieval, New York: Springer, [4] Youxi Wu Shuai Fu He Jiang Xindong Wu Strict approximate pattern matching with general gaps Springer Science+Business Media New York [5] Anssi Klapuri Pattern induction and matching in music signals Exploring Music Contents, Springer 2011 [6] J. Paulus, M. Muller, and A. Klapuri, Audio-based music structure analysis," in Proc. of the Int. Society for Music Information Retrieval Conference, Utrecht, Netherlands, [7] J. Serra, E. Gomez, P. Herrera,, and X. Serra, Chroma binary similarity and local alignment applied to cover song identi_cation," IEEE Trans. on Audio, Speech, andlanguage Processing, vol. 16, pp. 1138{1152, [8] R. Typke, Music retrieval based on melodic similarity," Ph.D. dissertation, Universiteit Utrecht, [9] Alexios Kotsifakos, Isak Karlsson, Panagiotis Papapetrou, Vassilis Athitsos, Dimitrios Gunopulos, Embedding-based subsequence matching with gaps range tolerances: a Query-By-Humming application, The VLDB Journal August 2015, Volume 24, Issue 4, pp URL: 86

10 T r a n s a c t i o n s o n M a c h i n e L e a r n i n g a n d A r t i f i c i a l I n t e l l i g e n c e V o l u m e 4, I s s u e 6, D e c [10] D. Vikram, Dr. M. Shashi, Content Based Indexing of Music Objects using Approximate Sequential Patterns International Journal of Data Mining & Knowledge Management Process (IJDKP) Vol.5, No.2, March [11] D. Vikram, Dr. M. Shashi, Content Based Music Information Retrieval: Concepts And Techniques Journal of Multidisciplinary Engineering Science and Technology (JMEST) ISSN: Vol. 2 Issue 8, August 2015 [12] Hung-Che Shen, Chungnan Lee Whistle for music: using melody transcription and approximate string matching for content-based query over a MIDI database Multimed Tools Appl (2007) 35: [13] Raga Surabhi is a collection of audio files containing raga snippets and songs for the process of understanding and learning Carnatic music. [14] J. Stephen Downie, Andreas F. Ehmann, Mert Bay, M. Cameron Jones The Music Information Retrieval Evaluation exchange: Some Observations and Insights, Advances in Music Information Retrieval, Springer, C o p y r i g h t S o c i e t y f o r S c i e n c e a n d E d u c a t i o n U n i t e d K i n g d o m 87

CONTENT BASED INDEXING OF MUSIC OBJECTS USING APPROXIMATE SEQUENTIAL PATTERNS

CONTENT BASED INDEXING OF MUSIC OBJECTS USING APPROXIMATE SEQUENTIAL PATTERNS CONTENT BASED INDEXING OF MUSIC OBJECTS USING APPROXIMATE SEQUENTIAL PATTERNS ABSTRACT D.Vikram 1 and Dr.M.Shashi 2 1 SRF(CSIR) and 2 Professor Department of Computer Science and Systems Engineering Andhra

More information

Music Radar: A Web-based Query by Humming System

Music Radar: A Web-based Query by Humming System Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,

More information

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University Week 14 Query-by-Humming and Music Fingerprinting Roger B. Dannenberg Professor of Computer Science, Art and Music Overview n Melody-Based Retrieval n Audio-Score Alignment n Music Fingerprinting 2 Metadata-based

More information

Outline. Why do we classify? Audio Classification

Outline. Why do we classify? Audio Classification Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification Implementation Future Work Why do we classify

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

A CHROMA-BASED SALIENCE FUNCTION FOR MELODY AND BASS LINE ESTIMATION FROM MUSIC AUDIO SIGNALS

A CHROMA-BASED SALIENCE FUNCTION FOR MELODY AND BASS LINE ESTIMATION FROM MUSIC AUDIO SIGNALS A CHROMA-BASED SALIENCE FUNCTION FOR MELODY AND BASS LINE ESTIMATION FROM MUSIC AUDIO SIGNALS Justin Salamon Music Technology Group Universitat Pompeu Fabra, Barcelona, Spain justin.salamon@upf.edu Emilia

More information

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES 12th International Society for Music Information Retrieval Conference (ISMIR 2011) A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES Erdem Unal 1 Elaine Chew 2 Panayiotis Georgiou

More information

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC Vishweshwara Rao, Sachin Pant, Madhumita Bhaskar and Preeti Rao Department of Electrical Engineering, IIT Bombay {vishu, sachinp,

More information

Music Information Retrieval Using Audio Input

Music Information Retrieval Using Audio Input Music Information Retrieval Using Audio Input Lloyd A. Smith, Rodger J. McNab and Ian H. Witten Department of Computer Science University of Waikato Private Bag 35 Hamilton, New Zealand {las, rjmcnab,

More information

A Fast Alignment Scheme for Automatic OCR Evaluation of Books

A Fast Alignment Scheme for Automatic OCR Evaluation of Books A Fast Alignment Scheme for Automatic OCR Evaluation of Books Ismet Zeki Yalniz, R. Manmatha Multimedia Indexing and Retrieval Group Dept. of Computer Science, University of Massachusetts Amherst, MA,

More information

Robert Alexandru Dobre, Cristian Negrescu

Robert Alexandru Dobre, Cristian Negrescu ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Automatic Music Transcription Software Based on Constant Q

More information

Music Structure Analysis

Music Structure Analysis Lecture Music Processing Music Structure Analysis Meinard Müller International Audio Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de Book: Fundamentals of Music Processing Meinard Müller Fundamentals

More information

Audio Structure Analysis

Audio Structure Analysis Tutorial T3 A Basic Introduction to Audio-Related Music Information Retrieval Audio Structure Analysis Meinard Müller, Christof Weiß International Audio Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de,

More information

Statistical Modeling and Retrieval of Polyphonic Music

Statistical Modeling and Retrieval of Polyphonic Music Statistical Modeling and Retrieval of Polyphonic Music Erdem Unal Panayiotis G. Georgiou and Shrikanth S. Narayanan Speech Analysis and Interpretation Laboratory University of Southern California Los Angeles,

More information

Melody classification using patterns

Melody classification using patterns Melody classification using patterns Darrell Conklin Department of Computing City University London United Kingdom conklin@city.ac.uk Abstract. A new method for symbolic music classification is proposed,

More information

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS Andrew N. Robertson, Mark D. Plumbley Centre for Digital Music

More information

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC Scientific Journal of Impact Factor (SJIF): 5.71 International Journal of Advance Engineering and Research Development Volume 5, Issue 04, April -2018 e-issn (O): 2348-4470 p-issn (P): 2348-6406 MUSICAL

More information

Melody Retrieval On The Web

Melody Retrieval On The Web Melody Retrieval On The Web Thesis proposal for the degree of Master of Science at the Massachusetts Institute of Technology M.I.T Media Laboratory Fall 2000 Thesis supervisor: Barry Vercoe Professor,

More information

Music Information Retrieval with Temporal Features and Timbre

Music Information Retrieval with Temporal Features and Timbre Music Information Retrieval with Temporal Features and Timbre Angelina A. Tzacheva and Keith J. Bell University of South Carolina Upstate, Department of Informatics 800 University Way, Spartanburg, SC

More information

Retrieval of textual song lyrics from sung inputs

Retrieval of textual song lyrics from sung inputs INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Retrieval of textual song lyrics from sung inputs Anna M. Kruspe Fraunhofer IDMT, Ilmenau, Germany kpe@idmt.fraunhofer.de Abstract Retrieving the

More information

Singer Traits Identification using Deep Neural Network

Singer Traits Identification using Deep Neural Network Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic

More information

Music Genre Classification and Variance Comparison on Number of Genres

Music Genre Classification and Variance Comparison on Number of Genres Music Genre Classification and Variance Comparison on Number of Genres Miguel Francisco, miguelf@stanford.edu Dong Myung Kim, dmk8265@stanford.edu 1 Abstract In this project we apply machine learning techniques

More information

Music Processing Audio Retrieval Meinard Müller

Music Processing Audio Retrieval Meinard Müller Lecture Music Processing Audio Retrieval Meinard Müller International Audio Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de Book: Fundamentals of Music Processing Meinard Müller Fundamentals

More information

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis Automatic characterization of ornamentation from bassoon recordings for expressive synthesis Montserrat Puiggròs, Emilia Gómez, Rafael Ramírez, Xavier Serra Music technology Group Universitat Pompeu Fabra

More information

A Music Retrieval System Using Melody and Lyric

A Music Retrieval System Using Melody and Lyric 202 IEEE International Conference on Multimedia and Expo Workshops A Music Retrieval System Using Melody and Lyric Zhiyuan Guo, Qiang Wang, Gang Liu, Jun Guo, Yueming Lu 2 Pattern Recognition and Intelligent

More information

Transcription of the Singing Melody in Polyphonic Music

Transcription of the Singing Melody in Polyphonic Music Transcription of the Singing Melody in Polyphonic Music Matti Ryynänen and Anssi Klapuri Institute of Signal Processing, Tampere University Of Technology P.O.Box 553, FI-33101 Tampere, Finland {matti.ryynanen,

More information

Finding Exact and Approximate Repeating Patterns to Build Database for Content-Based Music Information Retrieval in Carnatic Music

Finding Exact and Approximate Repeating Patterns to Build Database for Content-Based Music Information Retrieval in Carnatic Music Finding Exact and Approximate Repeating Patterns to Build Database for Content-Based Music Information Retrieval in Carnatic Music 1 G Sai Naresh, 2 Archana Raghuvamshi, 3 Dara Vikram 1 M.tech (CST), Assistant

More information

Music Similarity and Cover Song Identification: The Case of Jazz

Music Similarity and Cover Song Identification: The Case of Jazz Music Similarity and Cover Song Identification: The Case of Jazz Simon Dixon and Peter Foster s.e.dixon@qmul.ac.uk Centre for Digital Music School of Electronic Engineering and Computer Science Queen Mary

More information

Algorithms for melody search and transcription. Antti Laaksonen

Algorithms for melody search and transcription. Antti Laaksonen Department of Computer Science Series of Publications A Report A-2015-5 Algorithms for melody search and transcription Antti Laaksonen To be presented, with the permission of the Faculty of Science of

More information

Automatic Piano Music Transcription

Automatic Piano Music Transcription Automatic Piano Music Transcription Jianyu Fan Qiuhan Wang Xin Li Jianyu.Fan.Gr@dartmouth.edu Qiuhan.Wang.Gr@dartmouth.edu Xi.Li.Gr@dartmouth.edu 1. Introduction Writing down the score while listening

More information

Singer Recognition and Modeling Singer Error

Singer Recognition and Modeling Singer Error Singer Recognition and Modeling Singer Error Johan Ismael Stanford University jismael@stanford.edu Nicholas McGee Stanford University ndmcgee@stanford.edu 1. Abstract We propose a system for recognizing

More information

IMPROVED MELODIC SEQUENCE MATCHING FOR QUERY BASED SEARCHING IN INDIAN CLASSICAL MUSIC

IMPROVED MELODIC SEQUENCE MATCHING FOR QUERY BASED SEARCHING IN INDIAN CLASSICAL MUSIC IMPROVED MELODIC SEQUENCE MATCHING FOR QUERY BASED SEARCHING IN INDIAN CLASSICAL MUSIC Ashwin Lele #, Saurabh Pinjani #, Kaustuv Kanti Ganguli, and Preeti Rao Department of Electrical Engineering, Indian

More information

NEW QUERY-BY-HUMMING MUSIC RETRIEVAL SYSTEM CONCEPTION AND EVALUATION BASED ON A QUERY NATURE STUDY

NEW QUERY-BY-HUMMING MUSIC RETRIEVAL SYSTEM CONCEPTION AND EVALUATION BASED ON A QUERY NATURE STUDY Proceedings of the COST G-6 Conference on Digital Audio Effects (DAFX-), Limerick, Ireland, December 6-8,2 NEW QUERY-BY-HUMMING MUSIC RETRIEVAL SYSTEM CONCEPTION AND EVALUATION BASED ON A QUERY NATURE

More information

Effects of acoustic degradations on cover song recognition

Effects of acoustic degradations on cover song recognition Signal Processing in Acoustics: Paper 68 Effects of acoustic degradations on cover song recognition Julien Osmalskyj (a), Jean-Jacques Embrechts (b) (a) University of Liège, Belgium, josmalsky@ulg.ac.be

More information

A TEXT RETRIEVAL APPROACH TO CONTENT-BASED AUDIO RETRIEVAL

A TEXT RETRIEVAL APPROACH TO CONTENT-BASED AUDIO RETRIEVAL A TEXT RETRIEVAL APPROACH TO CONTENT-BASED AUDIO RETRIEVAL Matthew Riley University of Texas at Austin mriley@gmail.com Eric Heinen University of Texas at Austin eheinen@mail.utexas.edu Joydeep Ghosh University

More information

Chord Classification of an Audio Signal using Artificial Neural Network

Chord Classification of an Audio Signal using Artificial Neural Network Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------

More information

Query By Humming: Finding Songs in a Polyphonic Database

Query By Humming: Finding Songs in a Polyphonic Database Query By Humming: Finding Songs in a Polyphonic Database John Duchi Computer Science Department Stanford University jduchi@stanford.edu Benjamin Phipps Computer Science Department Stanford University bphipps@stanford.edu

More information

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM A QUER B EAMPLE MUSIC RETRIEVAL ALGORITHM H. HARB AND L. CHEN Maths-Info department, Ecole Centrale de Lyon. 36, av. Guy de Collongue, 69134, Ecully, France, EUROPE E-mail: {hadi.harb, liming.chen}@ec-lyon.fr

More information

Semi-supervised Musical Instrument Recognition

Semi-supervised Musical Instrument Recognition Semi-supervised Musical Instrument Recognition Master s Thesis Presentation Aleksandr Diment 1 1 Tampere niversity of Technology, Finland Supervisors: Adj.Prof. Tuomas Virtanen, MSc Toni Heittola 17 May

More information

AUTOMATIC MAPPING OF SCANNED SHEET MUSIC TO AUDIO RECORDINGS

AUTOMATIC MAPPING OF SCANNED SHEET MUSIC TO AUDIO RECORDINGS AUTOMATIC MAPPING OF SCANNED SHEET MUSIC TO AUDIO RECORDINGS Christian Fremerey, Meinard Müller,Frank Kurth, Michael Clausen Computer Science III University of Bonn Bonn, Germany Max-Planck-Institut (MPI)

More information

Author Index. Absolu, Brandt 165. Montecchio, Nicola 187 Mukherjee, Bhaswati 285 Müllensiefen, Daniel 365. Bay, Mert 93

Author Index. Absolu, Brandt 165. Montecchio, Nicola 187 Mukherjee, Bhaswati 285 Müllensiefen, Daniel 365. Bay, Mert 93 Author Index Absolu, Brandt 165 Bay, Mert 93 Datta, Ashoke Kumar 285 Dey, Nityananda 285 Doraisamy, Shyamala 391 Downie, J. Stephen 93 Ehmann, Andreas F. 93 Esposito, Roberto 143 Gerhard, David 119 Golzari,

More information

Data-Driven Solo Voice Enhancement for Jazz Music Retrieval

Data-Driven Solo Voice Enhancement for Jazz Music Retrieval Data-Driven Solo Voice Enhancement for Jazz Music Retrieval Stefan Balke1, Christian Dittmar1, Jakob Abeßer2, Meinard Müller1 1International Audio Laboratories Erlangen 2Fraunhofer Institute for Digital

More information

AUDIO-BASED COVER SONG RETRIEVAL USING APPROXIMATE CHORD SEQUENCES: TESTING SHIFTS, GAPS, SWAPS AND BEATS

AUDIO-BASED COVER SONG RETRIEVAL USING APPROXIMATE CHORD SEQUENCES: TESTING SHIFTS, GAPS, SWAPS AND BEATS AUDIO-BASED COVER SONG RETRIEVAL USING APPROXIMATE CHORD SEQUENCES: TESTING SHIFTS, GAPS, SWAPS AND BEATS Juan Pablo Bello Music Technology, New York University jpbello@nyu.edu ABSTRACT This paper presents

More information

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu

More information

Comparison of Dictionary-Based Approaches to Automatic Repeating Melody Extraction

Comparison of Dictionary-Based Approaches to Automatic Repeating Melody Extraction Comparison of Dictionary-Based Approaches to Automatic Repeating Melody Extraction Hsuan-Huei Shih, Shrikanth S. Narayanan and C.-C. Jay Kuo Integrated Media Systems Center and Department of Electrical

More information

A repetition-based framework for lyric alignment in popular songs

A repetition-based framework for lyric alignment in popular songs A repetition-based framework for lyric alignment in popular songs ABSTRACT LUONG Minh Thang and KAN Min Yen Department of Computer Science, School of Computing, National University of Singapore We examine

More information

Predicting Variation of Folk Songs: A Corpus Analysis Study on the Memorability of Melodies Janssen, B.D.; Burgoyne, J.A.; Honing, H.J.

Predicting Variation of Folk Songs: A Corpus Analysis Study on the Memorability of Melodies Janssen, B.D.; Burgoyne, J.A.; Honing, H.J. UvA-DARE (Digital Academic Repository) Predicting Variation of Folk Songs: A Corpus Analysis Study on the Memorability of Melodies Janssen, B.D.; Burgoyne, J.A.; Honing, H.J. Published in: Frontiers in

More information

AUTOMATICALLY IDENTIFYING VOCAL EXPRESSIONS FOR MUSIC TRANSCRIPTION

AUTOMATICALLY IDENTIFYING VOCAL EXPRESSIONS FOR MUSIC TRANSCRIPTION AUTOMATICALLY IDENTIFYING VOCAL EXPRESSIONS FOR MUSIC TRANSCRIPTION Sai Sumanth Miryala Kalika Bali Ranjita Bhagwan Monojit Choudhury mssumanth99@gmail.com kalikab@microsoft.com bhagwan@microsoft.com monojitc@microsoft.com

More information

Chroma Binary Similarity and Local Alignment Applied to Cover Song Identification

Chroma Binary Similarity and Local Alignment Applied to Cover Song Identification 1138 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 16, NO. 6, AUGUST 2008 Chroma Binary Similarity and Local Alignment Applied to Cover Song Identification Joan Serrà, Emilia Gómez,

More information

Computational Modelling of Harmony

Computational Modelling of Harmony Computational Modelling of Harmony Simon Dixon Centre for Digital Music, Queen Mary University of London, Mile End Rd, London E1 4NS, UK simon.dixon@elec.qmul.ac.uk http://www.elec.qmul.ac.uk/people/simond

More information

FLUX-CiM: Flexible Unsupervised Extraction of Citation Metadata

FLUX-CiM: Flexible Unsupervised Extraction of Citation Metadata FLUX-CiM: Flexible Unsupervised Extraction of Citation Metadata Eli Cortez 1, Filipe Mesquita 1, Altigran S. da Silva 1 Edleno Moura 1, Marcos André Gonçalves 2 1 Universidade Federal do Amazonas Departamento

More information

Proc. of NCC 2010, Chennai, India A Melody Detection User Interface for Polyphonic Music

Proc. of NCC 2010, Chennai, India A Melody Detection User Interface for Polyphonic Music A Melody Detection User Interface for Polyphonic Music Sachin Pant, Vishweshwara Rao, and Preeti Rao Department of Electrical Engineering Indian Institute of Technology Bombay, Mumbai 400076, India Email:

More information

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon A Study of Synchronization of Audio Data with Symbolic Data Music254 Project Report Spring 2007 SongHui Chon Abstract This paper provides an overview of the problem of audio and symbolic synchronization.

More information

Music Segmentation Using Markov Chain Methods

Music Segmentation Using Markov Chain Methods Music Segmentation Using Markov Chain Methods Paul Finkelstein March 8, 2011 Abstract This paper will present just how far the use of Markov Chains has spread in the 21 st century. We will explain some

More information

Music Structure Analysis

Music Structure Analysis Overview Tutorial Music Structure Analysis Part I: Principles & Techniques (Meinard Müller) Coffee Break Meinard Müller International Audio Laboratories Erlangen Universität Erlangen-Nürnberg meinard.mueller@audiolabs-erlangen.de

More information

MELODY EXTRACTION BASED ON HARMONIC CODED STRUCTURE

MELODY EXTRACTION BASED ON HARMONIC CODED STRUCTURE 12th International Society for Music Information Retrieval Conference (ISMIR 2011) MELODY EXTRACTION BASED ON HARMONIC CODED STRUCTURE Sihyun Joo Sanghun Park Seokhwan Jo Chang D. Yoo Department of Electrical

More information

Tool-based Identification of Melodic Patterns in MusicXML Documents

Tool-based Identification of Melodic Patterns in MusicXML Documents Tool-based Identification of Melodic Patterns in MusicXML Documents Manuel Burghardt (manuel.burghardt@ur.de), Lukas Lamm (lukas.lamm@stud.uni-regensburg.de), David Lechler (david.lechler@stud.uni-regensburg.de),

More information

Music Information Retrieval

Music Information Retrieval CTP 431 Music and Audio Computing Music Information Retrieval Graduate School of Culture Technology (GSCT) Juhan Nam 1 Introduction ü Instrument: Piano ü Composer: Chopin ü Key: E-minor ü Melody - ELO

More information

Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors *

Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors * Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors * David Ortega-Pacheco and Hiram Calvo Centro de Investigación en Computación, Instituto Politécnico Nacional, Av. Juan

More information

A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION

A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION Graham E. Poliner and Daniel P.W. Ellis LabROSA, Dept. of Electrical Engineering Columbia University, New York NY 127 USA {graham,dpwe}@ee.columbia.edu

More information

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES Vishweshwara Rao and Preeti Rao Digital Audio Processing Lab, Electrical Engineering Department, IIT-Bombay, Powai,

More information

Creating a Feature Vector to Identify Similarity between MIDI Files

Creating a Feature Vector to Identify Similarity between MIDI Files Creating a Feature Vector to Identify Similarity between MIDI Files Joseph Stroud 2017 Honors Thesis Advised by Sergio Alvarez Computer Science Department, Boston College 1 Abstract Today there are many

More information

CSC475 Music Information Retrieval

CSC475 Music Information Retrieval CSC475 Music Information Retrieval Monophonic pitch extraction George Tzanetakis University of Victoria 2014 G. Tzanetakis 1 / 32 Table of Contents I 1 Motivation and Terminology 2 Psychacoustics 3 F0

More information

A Query-by-singing Technique for Retrieving Polyphonic Objects of Popular Music

A Query-by-singing Technique for Retrieving Polyphonic Objects of Popular Music A Query-by-singing Technique for Retrieving Polyphonic Objects of Popular Music Hung-Ming Yu, Wei-Ho Tsai, and Hsin-Min Wang Institute of Information Science, Academia Sinica, Taipei, Taiwan, Republic

More information

Evaluation of Melody Similarity Measures

Evaluation of Melody Similarity Measures Evaluation of Melody Similarity Measures by Matthew Brian Kelly A thesis submitted to the School of Computing in conformity with the requirements for the degree of Master of Science Queen s University

More information

Music Information Retrieval

Music Information Retrieval Music Information Retrieval Informative Experiences in Computation and the Archive David De Roure @dder David De Roure @dder Four quadrants Big Data Scientific Computing Machine Learning Automation More

More information

Content-based music retrieval

Content-based music retrieval Music retrieval 1 Music retrieval 2 Content-based music retrieval Music information retrieval (MIR) is currently an active research area See proceedings of ISMIR conference and annual MIREX evaluations

More information

Polyphonic Audio Matching for Score Following and Intelligent Audio Editors

Polyphonic Audio Matching for Score Following and Intelligent Audio Editors Polyphonic Audio Matching for Score Following and Intelligent Audio Editors Roger B. Dannenberg and Ning Hu School of Computer Science, Carnegie Mellon University email: dannenberg@cs.cmu.edu, ninghu@cs.cmu.edu,

More information

CS229 Project Report Polyphonic Piano Transcription

CS229 Project Report Polyphonic Piano Transcription CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project

More information

DISCOVERY OF REPEATED VOCAL PATTERNS IN POLYPHONIC AUDIO: A CASE STUDY ON FLAMENCO MUSIC. Univ. of Piraeus, Greece

DISCOVERY OF REPEATED VOCAL PATTERNS IN POLYPHONIC AUDIO: A CASE STUDY ON FLAMENCO MUSIC. Univ. of Piraeus, Greece DISCOVERY OF REPEATED VOCAL PATTERNS IN POLYPHONIC AUDIO: A CASE STUDY ON FLAMENCO MUSIC Nadine Kroher 1, Aggelos Pikrakis 2, Jesús Moreno 3, José-Miguel Díaz-Báñez 3 1 Music Technology Group Univ. Pompeu

More information

A Note Based Query By Humming System using Convolutional Neural Network

A Note Based Query By Humming System using Convolutional Neural Network INTERSPEECH 2017 August 20 24, 2017, Stockholm, Sweden A Note Based Query By Humming System using Convolutional Neural Network Naziba Mostafa, Pascale Fung The Hong Kong University of Science and Technology

More information

A probabilistic framework for audio-based tonal key and chord recognition

A probabilistic framework for audio-based tonal key and chord recognition A probabilistic framework for audio-based tonal key and chord recognition Benoit Catteau 1, Jean-Pierre Martens 1, and Marc Leman 2 1 ELIS - Electronics & Information Systems, Ghent University, Gent (Belgium)

More information

Evaluating Melodic Encodings for Use in Cover Song Identification

Evaluating Melodic Encodings for Use in Cover Song Identification Evaluating Melodic Encodings for Use in Cover Song Identification David D. Wickland wickland@uoguelph.ca David A. Calvert dcalvert@uoguelph.ca James Harley jharley@uoguelph.ca ABSTRACT Cover song identification

More information

Efficient Vocal Melody Extraction from Polyphonic Music Signals

Efficient Vocal Melody Extraction from Polyphonic Music Signals http://dx.doi.org/1.5755/j1.eee.19.6.4575 ELEKTRONIKA IR ELEKTROTECHNIKA, ISSN 1392-1215, VOL. 19, NO. 6, 213 Efficient Vocal Melody Extraction from Polyphonic Music Signals G. Yao 1,2, Y. Zheng 1,2, L.

More information

Lyrics Classification using Naive Bayes

Lyrics Classification using Naive Bayes Lyrics Classification using Naive Bayes Dalibor Bužić *, Jasminka Dobša ** * College for Information Technologies, Klaićeva 7, Zagreb, Croatia ** Faculty of Organization and Informatics, Pavlinska 2, Varaždin,

More information

User-Specific Learning for Recognizing a Singer s Intended Pitch

User-Specific Learning for Recognizing a Singer s Intended Pitch User-Specific Learning for Recognizing a Singer s Intended Pitch Andrew Guillory University of Washington Seattle, WA guillory@cs.washington.edu Sumit Basu Microsoft Research Redmond, WA sumitb@microsoft.com

More information

Automatic Rhythmic Notation from Single Voice Audio Sources

Automatic Rhythmic Notation from Single Voice Audio Sources Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung

More information

Low Power Estimation on Test Compression Technique for SoC based Design

Low Power Estimation on Test Compression Technique for SoC based Design Indian Journal of Science and Technology, Vol 8(4), DOI: 0.7485/ijst/205/v8i4/6848, July 205 ISSN (Print) : 0974-6846 ISSN (Online) : 0974-5645 Low Estimation on Test Compression Technique for SoC based

More information

2. AN INTROSPECTION OF THE MORPHING PROCESS

2. AN INTROSPECTION OF THE MORPHING PROCESS 1. INTRODUCTION Voice morphing means the transition of one speech signal into another. Like image morphing, speech morphing aims to preserve the shared characteristics of the starting and final signals,

More information

Music Representations. Beethoven, Bach, and Billions of Bytes. Music. Research Goals. Piano Roll Representation. Player Piano (1900)

Music Representations. Beethoven, Bach, and Billions of Bytes. Music. Research Goals. Piano Roll Representation. Player Piano (1900) Music Representations Lecture Music Processing Sheet Music (Image) CD / MP3 (Audio) MusicXML (Text) Beethoven, Bach, and Billions of Bytes New Alliances between Music and Computer Science Dance / Motion

More information

Rhythm related MIR tasks

Rhythm related MIR tasks Rhythm related MIR tasks Ajay Srinivasamurthy 1, André Holzapfel 1 1 MTG, Universitat Pompeu Fabra, Barcelona, Spain 10 July, 2012 Srinivasamurthy et al. (UPF) MIR tasks 10 July, 2012 1 / 23 1 Rhythm 2

More information

Lecture 12: Alignment and Matching

Lecture 12: Alignment and Matching ELEN E4896 MUSIC SIGNAL PROCESSING Lecture 12: Alignment and Matching 1. Music Alignment 2. Cover Song Detection 3. Echo Nest Analyze Dan Ellis Dept. Electrical Engineering, Columbia University dpwe@ee.columbia.edu

More information

Chroma-based Predominant Melody and Bass Line Extraction from Music Audio Signals

Chroma-based Predominant Melody and Bass Line Extraction from Music Audio Signals Chroma-based Predominant Melody and Bass Line Extraction from Music Audio Signals Justin Jonathan Salamon Master Thesis submitted in partial fulfillment of the requirements for the degree: Master in Cognitive

More information

Audio Cover Song Identification using Convolutional Neural Network

Audio Cover Song Identification using Convolutional Neural Network Audio Cover Song Identification using Convolutional Neural Network Sungkyun Chang 1,4, Juheon Lee 2,4, Sang Keun Choe 3,4 and Kyogu Lee 1,4 Music and Audio Research Group 1, College of Liberal Studies

More information

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval DAY 1 Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval Jay LeBoeuf Imagine Research jay{at}imagine-research.com Rebecca

More information

SHEET MUSIC-AUDIO IDENTIFICATION

SHEET MUSIC-AUDIO IDENTIFICATION SHEET MUSIC-AUDIO IDENTIFICATION Christian Fremerey, Michael Clausen, Sebastian Ewert Bonn University, Computer Science III Bonn, Germany {fremerey,clausen,ewerts}@cs.uni-bonn.de Meinard Müller Saarland

More information

Multiple instrument tracking based on reconstruction error, pitch continuity and instrument activity

Multiple instrument tracking based on reconstruction error, pitch continuity and instrument activity Multiple instrument tracking based on reconstruction error, pitch continuity and instrument activity Holger Kirchhoff 1, Simon Dixon 1, and Anssi Klapuri 2 1 Centre for Digital Music, Queen Mary University

More information

A Bootstrap Method for Training an Accurate Audio Segmenter

A Bootstrap Method for Training an Accurate Audio Segmenter A Bootstrap Method for Training an Accurate Audio Segmenter Ning Hu and Roger B. Dannenberg Computer Science Department Carnegie Mellon University 5000 Forbes Ave Pittsburgh, PA 1513 {ninghu,rbd}@cs.cmu.edu

More information

Visualizing the Chromatic Index of Music

Visualizing the Chromatic Index of Music Visualizing the Chromatic Index of Music Dionysios Politis, Dimitrios Margounakis, Konstantinos Mokos Multimedia Lab, Department of Informatics Aristotle University of Thessaloniki Greece {dpolitis, dmargoun}@csd.auth.gr,

More information

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the

More information

Automated extraction of motivic patterns and application to the analysis of Debussy s Syrinx

Automated extraction of motivic patterns and application to the analysis of Debussy s Syrinx Automated extraction of motivic patterns and application to the analysis of Debussy s Syrinx Olivier Lartillot University of Jyväskylä, Finland lartillo@campus.jyu.fi 1. General Framework 1.1. Motivic

More information

Creating Data Resources for Designing User-centric Frontends for Query by Humming Systems

Creating Data Resources for Designing User-centric Frontends for Query by Humming Systems Creating Data Resources for Designing User-centric Frontends for Query by Humming Systems Erdem Unal S. S. Narayanan H.-H. Shih Elaine Chew C.-C. Jay Kuo Speech Analysis and Interpretation Laboratory,

More information

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? NICHOLAS BORG AND GEORGE HOKKANEN Abstract. The possibility of a hit song prediction algorithm is both academically interesting and industry motivated.

More information

A LYRICS-MATCHING QBH SYSTEM FOR INTER- ACTIVE ENVIRONMENTS

A LYRICS-MATCHING QBH SYSTEM FOR INTER- ACTIVE ENVIRONMENTS A LYRICS-MATCHING QBH SYSTEM FOR INTER- ACTIVE ENVIRONMENTS Panagiotis Papiotis Music Technology Group, Universitat Pompeu Fabra panos.papiotis@gmail.com Hendrik Purwins Music Technology Group, Universitat

More information

Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment

Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment Gus G. Xia Dartmouth College Neukom Institute Hanover, NH, USA gxia@dartmouth.edu Roger B. Dannenberg Carnegie

More information

Music Alignment and Applications. Introduction

Music Alignment and Applications. Introduction Music Alignment and Applications Roger B. Dannenberg Schools of Computer Science, Art, and Music Introduction Music information comes in many forms Digital Audio Multi-track Audio Music Notation MIDI Structured

More information

Cascading Citation Indexing in Action *

Cascading Citation Indexing in Action * Cascading Citation Indexing in Action * T.Folias 1, D. Dervos 2, G.Evangelidis 1, N. Samaras 1 1 Dept. of Applied Informatics, University of Macedonia, Thessaloniki, Greece Tel: +30 2310891844, Fax: +30

More information

Automatic Reduction of MIDI Files Preserving Relevant Musical Content

Automatic Reduction of MIDI Files Preserving Relevant Musical Content Automatic Reduction of MIDI Files Preserving Relevant Musical Content Søren Tjagvad Madsen 1,2, Rainer Typke 2, and Gerhard Widmer 1,2 1 Department of Computational Perception, Johannes Kepler University,

More information

MUSIC SHAPELETS FOR FAST COVER SONG RECOGNITION

MUSIC SHAPELETS FOR FAST COVER SONG RECOGNITION MUSIC SHAPELETS FOR FAST COVER SONG RECOGNITION Diego F. Silva Vinícius M. A. Souza Gustavo E. A. P. A. Batista Instituto de Ciências Matemáticas e de Computação Universidade de São Paulo {diegofsilva,vsouza,gbatista}@icmc.usp.br

More information

Available online at ScienceDirect. Procedia Computer Science 46 (2015 )

Available online at  ScienceDirect. Procedia Computer Science 46 (2015 ) Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 46 (2015 ) 381 387 International Conference on Information and Communication Technologies (ICICT 2014) Music Information

More information