Proposal for Application of Speech Techniques to Music Analysis

Size: px
Start display at page:

Download "Proposal for Application of Speech Techniques to Music Analysis"

Transcription

1 Proposal for Application of Speech Techniques to Music Analysis 1. Research on Speech and Music Lin Zhong Dept. of Electronic Engineering Tsinghua University 1. Goal Speech research from the very beginning has aimed at benefiting the populace instead of experts. Besides the military prospects from which it has got a great deal of money, the two main goals for speech research have been 1) low bit-rate transmission of speech and 2) speech based human-machine interaction. For the former speech coding/synthesis is developed and for the later, speech recognition technology. With the development of microelectronics and network technology, these two technologies seem more and more desirable and lucrative. That's another reason why speech research attracts money from governments and industries. On the contrary, computer music began as something like personal hobbies, then aimed at benefiting music people, especially, composers and music students. Computer music research had got limited funding because it benefited a very limited community. As signal processing people and funds are gathering before the Internet, music people have noticed the tremendous need of common people for accessing, retrieving and indexing network audio resources( Vercoe et al 1998). And the same need has been noticed by speech people for long( Rose 1995). Moreover, with the development of Multi-modal Human-Machine Interfaces (Sharma et al 1998) and Integrated Media Systems( McLeod et al 1999), just like speech and gesture, music is also promising to be one mode for people to communicate with computers as we can communicate with each other through music( Here is where musical sound synthesis would be most attractive, I think). Before the Internet and Computer, Researches on Speech and Music can find common goal. And I believe this is the right place for computer music research to get funding. Many of speech and music researches have physically similar goals. ( See table 1) 2. Method Many methods for speech synthesis/coding are quite different from those for music synthesis, because they have rather different goals. The best quality at the cost of lest computation and distortion is the foremost concern for transmission aimed speech coding/synthesis, while usually authenticity is that for music synthesis. For the goal of natural mode based human machine interfaces, machines have to speak to their users. Here authenticity also become the foremost concern for speech synthesis, exactly speaking, text-to-speech synthesis. And many techniques employed here are just "transformed" versions of those for music synthesis. There are three strategies ever employed to tackle speech recognition ( Rabiner & Juang 1993), i.e., Acoustics-Phonetics, Pattern and Artificial Intelligence. However, the so far most successful one is statistics-based pattern recognition strategy, (although speech people are seeking ways to do away with the "devil" of statistics and welcome explicit modeling). When Spoken Language Understanding is concerned, AI is also incorporated. In addition, For speech, cepstral analysis is as widely employed as temporal or spectral analysis. The most widely adopted strategy for musical sound analysis seems to combine Artificial Intelligent, Psycho-acoustics, Music Theory and Signal Processing. Such strategy is more deterministic than statistic. The methods based on statistics or pattern classification hasn't yet been fully developed( Martin et al 1998b, Brown 1999 and Raphael 1999). Temporal and spectral analyses are much more employed than cepstral analysis. This proposal is preliminary, incomplete and is to undergo revision

2 However, music people should take pride in their insightful explorations of psycho-acoustics and human auditory system( Ellis 1992, Martin et al 1998a and Scheirer 1998). Such knowledge haven't so far been much exploited for speech recognition. Music people should also be proud of the diversity of the techniques they have been studying. While many technique of speech analysis become "standard" in the speech community, there is no unanimously adopted or agreed techniques for music analysis. What's more, as statistical acoustical and language modeling has for long been the core for speech analysis, the biggest share of speech recognition research has gone into collecting the speech database and training the statistical models. In this regard, music people have done much more for signal processing of musical sound. Table 1. Speech Research Topics and Their Music Counterparts Speech Music Topic Note Topics Note i)speech Synthesis from Coded Speech Synthesis speech for transmission Musical ii)text-to-speech synthesis Sound Synthesis techniques iii)synthesis for a specific speaker Speaker GMM/Cepstrum Instrument(Timbre) Language Music Genre/Type Speech Explicit Statistical and Acoustical Music Transcription Modeling Pitch Nearly the same with music Pitch Tracking Mandarin Speech Pitch Tracking and Temporal Music Tone Tone analysis, Neural Networks Cocktail Party 1) Mic Array Technique Effect/Speaker 2) Voice Separation and Tracking and Tacking Separation i)score-to-music synthesis ii)synthesis for a specific instrument See Roads' Book GMM/ Correlogram, Cepstrum No explicit statistical and acoustical modeling Nearly the same with speech 3. State of the Art The success for Speech coding/synthesis is evident in view of the speech based telecommunication( Juang edited 1998). Speech coding and compression research is now heading toward the theoretic limit and more robustness. In addition, speech people now talk about " Spoken Language Systems"( Cole et al 1995) instead of "Automatic Speech ". The later has become only a sub-area of the former. This change in terminology reflects how great advance research on speech recognition has made in the past years. Various commercialized automatic speech recognizers are available on the PC and single Chip( For example, see the homepage of Sensory, Inc.). Music synthesis for economical representation and transmission is the same successful in many respects. The standardized MIDI format is an excellent example. However, due to the inherent complexity of music and musical instrument acoustics, computer synthesized sound for such instruments as violin is far from satisfying in view of the quality and the required computation. Furthermore, unlike speech synthesis, which in most cases is only required to be understandable, musical instrumental sound synthesis must be enjoyable. Thus the big challenge for music synthesis. As for music analysis, " Today's computer systems are not capable of understanding music at the level of an average five-year-old; they cannot recognize a melody in a polyphonic recording or understand a song on a children's television program."( Martin et al 1998a). I believe there are much more interesting things to do with musical sound. 2. Speech Techniques and Their Prospects in Music As my up to now research experience centers at speech recognition, I'll stay within this domain here. There are three technologies generally boasted by speech recognition people. I would like to discuss them among many

3 other techniques. 1. Hidden Markov Models( HMM's) HMM's are most successful for statistically and dynamically modeling the speech signal. The implementation of HMM training is direct and mathematically sound, and decent generalization from training samples is usually observed. The embedded Viterbi Dynamic Programming is able to warp and align speech time variation. Moreover, HMM as both a finite automaton and probabilistic model can incorporate rules( lexics, syntax, semantics, etc.) and other probabilistic models( Neural Networks, Linear Regression etc.) easily. A system based on HMM is ready to be both knowledge-driven and data-driven. These features are desirable for music sound too. However, there are many limits for HMM's. The most serous two of them are 1. To ensure both robustness and accuracy, there are never enough speech data to train the system. If HMM's are applied to modeling musical sound, this problem would be more serious for the musical sound databases are just emerging. Speech people, while cherishing statistics, is seeking to avoid it. Researches have been conducted to model many phenomena explicitly instead of statistically. Such phenomena include speaker variation, background noise and channel distortion, etc.. Doing so will reduce the training data tremendously. 2. Distribution estimation based on HMM's is parametric and non-discriminative. The popular maximum likelihood estimation(mle) cannot ensure minimum classification error( MCE). One of the recent research foci is for training techniques that would secure most accuracy and robustness. Such training techniques as Maximum a Posteriori (MAP) training, Maximum Mutual Information(MMI) training and Generalized Probabilistic Descent/Minimum Classification Error(GPD/MCE) training have been advertised. I think that HMM's are the most promising technique that would be readily applied to music analysis. Some of the many possibilities are 1. Statistical Music Object Modeling and. Many music objects can be identified by people only through many notes or measures. They are identified as a dynamic time series. The methods ever employed are poor at dealing with the time-dynamic and integrating long segments. Moreover, knowledge based identifying systems can achieve little generalization. As HMM's are successful in modeling and recognizing isolated spoken words and phrases, I'd like to apply them to modeling and recognizing music object. Timbre recognition could be tackled by HMM's too. The current instrument recognition techniques base decision on only a few frames( Martin et al 1998b and Brown 1999). They try to catch the timbre by a glance instead of gazing at it. HMM's is superior in this respect. Music Object is significant because it paves the way for Music Object Extraction and Manipulation, thus for Content/Perception-based Music Resource Management and Access, and for model-based Musical Sound Data Reduction. 2. Statistical Model based Music Transcription. Music transcription would be another applications for statistical music object modeling. Until now, I haven't read any article about music transcription based on statistical models. Intuitively, if we statistically model the different changes of pitch( e.g., octave up or down, semitone up or down) as music objects( this is feasible just because music pitch change is quantified instead of continuous), we can track the pitch changes all through. Of course other techniques such as temporal analysis could also be incorporated. I'll continue this possibility in the following parts just to illustrate what one general speech technology could possibly do for music and how far it could go. I don't mean it is the only application speech technology could find for music or it is the best thing speech technology could do for music. 2. Mel Frequency Cepstrum Coefficients( MFCC's ) MFCC's demonstrate more robustness and accuracy while require more computation than another set of popular speech feature, Linear Predictive Cepstrum Coefficients(LPCC's). They are based on the simple knowledge for the frequency differentiating ability of the human cochlea. I don't think MFCC would be any better than the sound features already employed by most music people. As stated before, researches on computer music have employed much more psycho-acoustics. Here Musical Object refers to many kinds of musical events, such as sforzando, solo/tutti, crescendo/decrescendo, syncopation, pizzicato etc.. Such events are very important for common people to identify and describe music they've heard.

4 However, there are two issues I'd like to stress. First, most sound features employed in music analysis are resource(both computation and storage) demanding. To be implemented on PC's or inexpensive chips, which is important for them to be popular, music analysis systems must find efficient features like MFCC's and LPCC's. Another issue is Cepstral Analysis could be tried more for music analysis. For statistical acoustical modeling, acoustical features are of great importance, especially modeling music sound, which is extremely rich in perceptual information. Perceptually informative and integratable sound features are required for successful statistical modeling. Features extractable for PC's and inexpensive DSP or ASIC are very desirable for inexpensive music analysis systems. 3. N-Gram N-Gram is a statistics based language model. It is constructed by sorting and counting through the related training language materials, both spoken and written. Grammar rules can be explicitly incorporated into the language model. N-Gram has led to some successful Large Vocabulary Continuous Speech Systems ( For example, ViaVoice of IBM) and many domain/application/task related practical systems. I base my discussion on the successful music object modeling. If we have modeled music object, The way these models are organized certainly make up some kind of counterpart for the language grammar in speech recognition. Such music event/object organizing language could also be constructed in a similar way as N-Gram. Even the language could itself be the music version of N-Gram. Combining statistical music object modeling and the Music N-Gram, we could establish music analyzing, parsing and understanding systems. HMM, MFCC and N-grams have become the prototypes for statistical acoustical models, speech features and language models, respectively, for speech recognition. They are very general techniques. After examining Table 1., We could also obtain other specific speech techniques which would possible be applied to computer music. 3. Proposal Following the preceding two section, my proposal can be concise 1. Goal A graduate study oriented toward the Ph.D. should no doubt emphasize on applicable research and scholarship rather than on engineering practice. However, I believe that to get a clear idea about where one's research is heading will contribute a great deal to one's research. So, I propose to conduct research on music analysis/synthesis for Intelligent Network-based Music Resources Manipulation and Intelligent Multi-modal Human Machine Interfaces, that is, 1. Music Analysis for Content/Perception-based Musical Sound Data Indexing, Retrieval, Manipulation and Reduction, including Musical Type, Timbre and Object Identification; Musical Object Extraction and Manipulation; Resource Efficient Music Analysis Systems Perception Based Acoustical Features 2. Music Synthesis for Interactive Human-Machine Interfaces. 2. Method There is a big pool of candidate methods that could be adopted, from both speech and music research. In addition to the usually adopted methods, I propose to try the following speech techniques for the music analysis research, 1. Statistical Acoustical Modeling for Musical Sound, especially HMM's 2. Statistical Language Modeling for Music 3. Cepstral Analysis 4. Advanced Pattern Methodologies for Music Analysis We conduct research not just to be original or try new methods. Instead, we originate for we find the current

5 tools are limited; we try new methods for we expect they would yield better results. Never mind the cat is white or black, it's a good cat only when it catch mice(a Chinese saying). The case is the same in the computer music research. 3. Condition: A Personal View Four conditions have contributed most to speech research, 1) continual supports from governments and industries, 2)widely recognized and adopted standard speech database, 3)close collaboration of researchers in signal processing, statistics, linguistics and natural language understanding and 4) prompt commercialization of feasible speech technologies. Similarly, to be successful in the long run, computer music research should 1. Aim research at benefiting common people in addition to wealthy music experts or studio engineers( I firmly believe this is a common trend for signal processing and network researches.) 2. Try and Recognize the emerging musical sound databases( This need time and money). 3. Set up multidisciplinary groups which can accommodate direct communication and close collaboration( The current research groups are usually strong in some aspects while weak in others). And 4. Commercialize and Advertise any "usable", if not perfect, music technology( Doing so will not only get research better funded but also pave way for future widespread of computer music systems). 4. Reference Brown, J.(1999), "Computer Identification of Musical Instruments Using Pattern with Cepstral Coefficients as Features", J. Acoust. Soc. Am., Vol. 105, No. 3, March.. Cole, R. et al (1995), " The Challenge of Spoken Language Systems: Research Directions for the Nineties," IEEE Trans. Speech and Audio Processing, vol.3, no.1, pp.1-21 Ellis, D.P.W. (1992), A Perceptual Representation of Audio, Master's thesis, EECS dept, MIT Juang, B.-H. Edited( 1998), "The Past, Present, and Future of Speech Processing", IEEE Signal Processing Magazine, May, pp Martin, K. et al (1998a), "Music Content Analysis through Models of Audition," Presented at the 1998 ACM Multimedia Workshop on Content Processing of Music for Multimedia Applications, Bristol, England Martin, K.. et al(1998b), " 2pMU9: Music Instrument Identification: A Pattern- Approach", presented at the 136 th meeting of the J. Acoust. Soc. Am., Oct.13th McLeod, D. et al(1999), " The Move Toward Media Immersion", IEEE Signal Processing Magazine, Jan., pp Rabiner L. and Juang, B.-H.(1993), Fundamentals of Speech, Prentice-Hall, Englewood Cliff, New Jersey Raphael, C.(1999), " Automatic Segmentation of Acoustic Musical Signals Using Hidden Markov Models," IEEE Trans. on Pattern Analysis and Machine Intelligence, vol.21,no.4, April Rose, R.C.(1995), Keyword detection in conversational speech utterance using HMM based continuous speech recognition, Computer, Speech & Language,9, Scheirer, E.(1998), Music Perception Systems, Unpublished Proposal for Ph.D. Dissertation, MIT Media Laboratory Sharma, R.et al(1998), "Toward Multimodal Human-Computer Interface," Proc.IEEE, vol.86, no.5, pp Vercoe, B.L. et al(1998), " Structured Audio: Creation, Transmission, and Rendering of Parametric Sound Representations," Proc. IEEE, vol. 86, no.5, pp

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon A Study of Synchronization of Audio Data with Symbolic Data Music254 Project Report Spring 2007 SongHui Chon Abstract This paper provides an overview of the problem of audio and symbolic synchronization.

More information

Acoustic Scene Classification

Acoustic Scene Classification Acoustic Scene Classification Marc-Christoph Gerasch Seminar Topics in Computer Music - Acoustic Scene Classification 6/24/2015 1 Outline Acoustic Scene Classification - definition History and state of

More information

Automatic Laughter Detection

Automatic Laughter Detection Automatic Laughter Detection Mary Knox Final Project (EECS 94) knoxm@eecs.berkeley.edu December 1, 006 1 Introduction Laughter is a powerful cue in communication. It communicates to listeners the emotional

More information

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC Scientific Journal of Impact Factor (SJIF): 5.71 International Journal of Advance Engineering and Research Development Volume 5, Issue 04, April -2018 e-issn (O): 2348-4470 p-issn (P): 2348-6406 MUSICAL

More information

Melody Retrieval On The Web

Melody Retrieval On The Web Melody Retrieval On The Web Thesis proposal for the degree of Master of Science at the Massachusetts Institute of Technology M.I.T Media Laboratory Fall 2000 Thesis supervisor: Barry Vercoe Professor,

More information

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES PACS: 43.60.Lq Hacihabiboglu, Huseyin 1,2 ; Canagarajah C. Nishan 2 1 Sonic Arts Research Centre (SARC) School of Computer Science Queen s University

More information

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Róisín Loughran roisin.loughran@ul.ie Jacqueline Walker jacqueline.walker@ul.ie Michael O Neill University

More information

Singer Identification

Singer Identification Singer Identification Bertrand SCHERRER McGill University March 15, 2007 Bertrand SCHERRER (McGill University) Singer Identification March 15, 2007 1 / 27 Outline 1 Introduction Applications Challenges

More information

Music Radar: A Web-based Query by Humming System

Music Radar: A Web-based Query by Humming System Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,

More information

Statistical Modeling and Retrieval of Polyphonic Music

Statistical Modeling and Retrieval of Polyphonic Music Statistical Modeling and Retrieval of Polyphonic Music Erdem Unal Panayiotis G. Georgiou and Shrikanth S. Narayanan Speech Analysis and Interpretation Laboratory University of Southern California Los Angeles,

More information

Normalized Cumulative Spectral Distribution in Music

Normalized Cumulative Spectral Distribution in Music Normalized Cumulative Spectral Distribution in Music Young-Hwan Song, Hyung-Jun Kwon, and Myung-Jin Bae Abstract As the remedy used music becomes active and meditation effect through the music is verified,

More information

Automatic Laughter Detection

Automatic Laughter Detection Automatic Laughter Detection Mary Knox 1803707 knoxm@eecs.berkeley.edu December 1, 006 Abstract We built a system to automatically detect laughter from acoustic features of audio. To implement the system,

More information

Automatic music transcription

Automatic music transcription Educational Multimedia Application- Specific Music Transcription for Tutoring An applicationspecific, musictranscription approach uses a customized human computer interface to combine the strengths of

More information

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University Week 14 Query-by-Humming and Music Fingerprinting Roger B. Dannenberg Professor of Computer Science, Art and Music Overview n Melody-Based Retrieval n Audio-Score Alignment n Music Fingerprinting 2 Metadata-based

More information

Improving Frame Based Automatic Laughter Detection

Improving Frame Based Automatic Laughter Detection Improving Frame Based Automatic Laughter Detection Mary Knox EE225D Class Project knoxm@eecs.berkeley.edu December 13, 2007 Abstract Laughter recognition is an underexplored area of research. My goal for

More information

Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting

Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting Dalwon Jang 1, Seungjae Lee 2, Jun Seok Lee 2, Minho Jin 1, Jin S. Seo 2, Sunil Lee 1 and Chang D. Yoo 1 1 Korea Advanced

More information

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM A QUER B EAMPLE MUSIC RETRIEVAL ALGORITHM H. HARB AND L. CHEN Maths-Info department, Ecole Centrale de Lyon. 36, av. Guy de Collongue, 69134, Ecully, France, EUROPE E-mail: {hadi.harb, liming.chen}@ec-lyon.fr

More information

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu

More information

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC Vishweshwara Rao, Sachin Pant, Madhumita Bhaskar and Preeti Rao Department of Electrical Engineering, IIT Bombay {vishu, sachinp,

More information

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Aric Bartle (abartle@stanford.edu) December 14, 2012 1 Background The field of composer recognition has

More information

Hidden Markov Model based dance recognition

Hidden Markov Model based dance recognition Hidden Markov Model based dance recognition Dragutin Hrenek, Nenad Mikša, Robert Perica, Pavle Prentašić and Boris Trubić University of Zagreb, Faculty of Electrical Engineering and Computing Unska 3,

More information

A Fast Alignment Scheme for Automatic OCR Evaluation of Books

A Fast Alignment Scheme for Automatic OCR Evaluation of Books A Fast Alignment Scheme for Automatic OCR Evaluation of Books Ismet Zeki Yalniz, R. Manmatha Multimedia Indexing and Retrieval Group Dept. of Computer Science, University of Massachusetts Amherst, MA,

More information

A System for Acoustic Chord Transcription and Key Extraction from Audio Using Hidden Markov models Trained on Synthesized Audio

A System for Acoustic Chord Transcription and Key Extraction from Audio Using Hidden Markov models Trained on Synthesized Audio Curriculum Vitae Kyogu Lee Advanced Technology Center, Gracenote Inc. 2000 Powell Street, Suite 1380 Emeryville, CA 94608 USA Tel) 1-510-428-7296 Fax) 1-510-547-9681 klee@gracenote.com kglee@ccrma.stanford.edu

More information

A Music Retrieval System Using Melody and Lyric

A Music Retrieval System Using Melody and Lyric 202 IEEE International Conference on Multimedia and Expo Workshops A Music Retrieval System Using Melody and Lyric Zhiyuan Guo, Qiang Wang, Gang Liu, Jun Guo, Yueming Lu 2 Pattern Recognition and Intelligent

More information

Computational Modelling of Harmony

Computational Modelling of Harmony Computational Modelling of Harmony Simon Dixon Centre for Digital Music, Queen Mary University of London, Mile End Rd, London E1 4NS, UK simon.dixon@elec.qmul.ac.uk http://www.elec.qmul.ac.uk/people/simond

More information

Singer Traits Identification using Deep Neural Network

Singer Traits Identification using Deep Neural Network Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic

More information

Topics in Computer Music Instrument Identification. Ioanna Karydi

Topics in Computer Music Instrument Identification. Ioanna Karydi Topics in Computer Music Instrument Identification Ioanna Karydi Presentation overview What is instrument identification? Sound attributes & Timbre Human performance The ideal algorithm Selected approaches

More information

WE ADDRESS the development of a novel computational

WE ADDRESS the development of a novel computational IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 3, MARCH 2010 663 Dynamic Spectral Envelope Modeling for Timbre Analysis of Musical Instrument Sounds Juan José Burred, Member,

More information

Subjective Similarity of Music: Data Collection for Individuality Analysis

Subjective Similarity of Music: Data Collection for Individuality Analysis Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp

More information

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors Priyanka S. Jadhav M.E. (Computer Engineering) G. H. Raisoni College of Engg. & Mgmt. Wagholi, Pune, India E-mail:

More information

Transcription of the Singing Melody in Polyphonic Music

Transcription of the Singing Melody in Polyphonic Music Transcription of the Singing Melody in Polyphonic Music Matti Ryynänen and Anssi Klapuri Institute of Signal Processing, Tampere University Of Technology P.O.Box 553, FI-33101 Tampere, Finland {matti.ryynanen,

More information

Speech and Speaker Recognition for the Command of an Industrial Robot

Speech and Speaker Recognition for the Command of an Industrial Robot Speech and Speaker Recognition for the Command of an Industrial Robot CLAUDIA MOISA*, HELGA SILAGHI*, ANDREI SILAGHI** *Dept. of Electric Drives and Automation University of Oradea University Street, nr.

More information

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES 12th International Society for Music Information Retrieval Conference (ISMIR 2011) A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES Erdem Unal 1 Elaine Chew 2 Panayiotis Georgiou

More information

Chord Classification of an Audio Signal using Artificial Neural Network

Chord Classification of an Audio Signal using Artificial Neural Network Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------

More information

Singer Recognition and Modeling Singer Error

Singer Recognition and Modeling Singer Error Singer Recognition and Modeling Singer Error Johan Ismael Stanford University jismael@stanford.edu Nicholas McGee Stanford University ndmcgee@stanford.edu 1. Abstract We propose a system for recognizing

More information

Automatic Identification of Instrument Type in Music Signal using Wavelet and MFCC

Automatic Identification of Instrument Type in Music Signal using Wavelet and MFCC Automatic Identification of Instrument Type in Music Signal using Wavelet and MFCC Arijit Ghosal, Rudrasis Chakraborty, Bibhas Chandra Dhara +, and Sanjoy Kumar Saha! * CSE Dept., Institute of Technology

More information

Keywords Separation of sound, percussive instruments, non-percussive instruments, flexible audio source separation toolbox

Keywords Separation of sound, percussive instruments, non-percussive instruments, flexible audio source separation toolbox Volume 4, Issue 4, April 2014 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Investigation

More information

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 AN HMM BASED INVESTIGATION OF DIFFERENCES BETWEEN MUSICAL INSTRUMENTS OF THE SAME TYPE PACS: 43.75.-z Eichner, Matthias; Wolff, Matthias;

More information

Paulo V. K. Borges. Flat 1, 50A, Cephas Av. London, UK, E1 4AR (+44) PRESENTATION

Paulo V. K. Borges. Flat 1, 50A, Cephas Av. London, UK, E1 4AR (+44) PRESENTATION Paulo V. K. Borges Flat 1, 50A, Cephas Av. London, UK, E1 4AR (+44) 07942084331 vini@ieee.org PRESENTATION Electronic engineer working as researcher at University of London. Doctorate in digital image/video

More information

Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals

Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals Characteristics of Polyphonic Music Style and Markov Model of Pitch-Class Intervals Eita Nakamura and Shinji Takaki National Institute of Informatics, Tokyo 101-8430, Japan eita.nakamura@gmail.com, takaki@nii.ac.jp

More information

Robert Alexandru Dobre, Cristian Negrescu

Robert Alexandru Dobre, Cristian Negrescu ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Automatic Music Transcription Software Based on Constant Q

More information

Music Recommendation from Song Sets

Music Recommendation from Song Sets Music Recommendation from Song Sets Beth Logan Cambridge Research Laboratory HP Laboratories Cambridge HPL-2004-148 August 30, 2004* E-mail: Beth.Logan@hp.com music analysis, information retrieval, multimedia

More information

Automatic Construction of Synthetic Musical Instruments and Performers

Automatic Construction of Synthetic Musical Instruments and Performers Ph.D. Thesis Proposal Automatic Construction of Synthetic Musical Instruments and Performers Ning Hu Carnegie Mellon University Thesis Committee Roger B. Dannenberg, Chair Michael S. Lewicki Richard M.

More information

Topic 10. Multi-pitch Analysis

Topic 10. Multi-pitch Analysis Topic 10 Multi-pitch Analysis What is pitch? Common elements of music are pitch, rhythm, dynamics, and the sonic qualities of timbre and texture. An auditory perceptual attribute in terms of which sounds

More information

Speech To Song Classification

Speech To Song Classification Speech To Song Classification Emily Graber Center for Computer Research in Music and Acoustics, Department of Music, Stanford University Abstract The speech to song illusion is a perceptual phenomenon

More information

Classification of Timbre Similarity

Classification of Timbre Similarity Classification of Timbre Similarity Corey Kereliuk McGill University March 15, 2007 1 / 16 1 Definition of Timbre What Timbre is Not What Timbre is A 2-dimensional Timbre Space 2 3 Considerations Common

More information

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES Vishweshwara Rao and Preeti Rao Digital Audio Processing Lab, Electrical Engineering Department, IIT-Bombay, Powai,

More information

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? NICHOLAS BORG AND GEORGE HOKKANEN Abstract. The possibility of a hit song prediction algorithm is both academically interesting and industry motivated.

More information

Various Applications of Digital Signal Processing (DSP)

Various Applications of Digital Signal Processing (DSP) Various Applications of Digital Signal Processing (DSP) Neha Kapoor, Yash Kumar, Mona Sharma Student,ECE,DCE,Gurgaon, India EMAIL: neha04263@gmail.com, yashguptaip@gmail.com, monasharma1194@gmail.com ABSTRACT:-

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS Andrew N. Robertson, Mark D. Plumbley Centre for Digital Music

More information

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis Fengyan Wu fengyanyy@163.com Shutao Sun stsun@cuc.edu.cn Weiyao Xue Wyxue_std@163.com Abstract Automatic extraction of

More information

A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION

A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION Graham E. Poliner and Daniel P.W. Ellis LabROSA, Dept. of Electrical Engineering Columbia University, New York NY 127 USA {graham,dpwe}@ee.columbia.edu

More information

SYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS

SYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS Published by Institute of Electrical Engineers (IEE). 1998 IEE, Paul Masri, Nishan Canagarajah Colloquium on "Audio and Music Technology"; November 1998, London. Digest No. 98/470 SYNTHESIS FROM MUSICAL

More information

An Accurate Timbre Model for Musical Instruments and its Application to Classification

An Accurate Timbre Model for Musical Instruments and its Application to Classification An Accurate Timbre Model for Musical Instruments and its Application to Classification Juan José Burred 1,AxelRöbel 2, and Xavier Rodet 2 1 Communication Systems Group, Technical University of Berlin,

More information

Piano Transcription MUMT611 Presentation III 1 March, Hankinson, 1/15

Piano Transcription MUMT611 Presentation III 1 March, Hankinson, 1/15 Piano Transcription MUMT611 Presentation III 1 March, 2007 Hankinson, 1/15 Outline Introduction Techniques Comb Filtering & Autocorrelation HMMs Blackboard Systems & Fuzzy Logic Neural Networks Examples

More information

Non Stationary Signals (Voice) Verification System Using Wavelet Transform

Non Stationary Signals (Voice) Verification System Using Wavelet Transform Non Stationary Signals (Voice) Verification System Using Wavelet Transform PPS Subhashini Associate Professor, Department of ECE, RVR & JC College of Engineering, Guntur. Dr.M.Satya Sairam Professor &

More information

Pattern Recognition in Music

Pattern Recognition in Music Pattern Recognition in Music SAMBA/07/02 Line Eikvil Ragnar Bang Huseby February 2002 Copyright Norsk Regnesentral NR-notat/NR Note Tittel/Title: Pattern Recognition in Music Dato/Date: February År/Year:

More information

A System for Automatic Chord Transcription from Audio Using Genre-Specific Hidden Markov Models

A System for Automatic Chord Transcription from Audio Using Genre-Specific Hidden Markov Models A System for Automatic Chord Transcription from Audio Using Genre-Specific Hidden Markov Models Kyogu Lee Center for Computer Research in Music and Acoustics Stanford University, Stanford CA 94305, USA

More information

Study of White Gaussian Noise with Varying Signal to Noise Ratio in Speech Signal using Wavelet

Study of White Gaussian Noise with Varying Signal to Noise Ratio in Speech Signal using Wavelet American International Journal of Research in Science, Technology, Engineering & Mathematics Available online at http://www.iasir.net ISSN (Print): 2328-3491, ISSN (Online): 2328-3580, ISSN (CD-ROM): 2328-3629

More information

Musical Creativity. Jukka Toivanen Introduction to Computational Creativity Dept. of Computer Science University of Helsinki

Musical Creativity. Jukka Toivanen Introduction to Computational Creativity Dept. of Computer Science University of Helsinki Musical Creativity Jukka Toivanen Introduction to Computational Creativity Dept. of Computer Science University of Helsinki Basic Terminology Melody = linear succession of musical tones that the listener

More information

Query By Humming: Finding Songs in a Polyphonic Database

Query By Humming: Finding Songs in a Polyphonic Database Query By Humming: Finding Songs in a Polyphonic Database John Duchi Computer Science Department Stanford University jduchi@stanford.edu Benjamin Phipps Computer Science Department Stanford University bphipps@stanford.edu

More information

CSC475 Music Information Retrieval

CSC475 Music Information Retrieval CSC475 Music Information Retrieval Monophonic pitch extraction George Tzanetakis University of Victoria 2014 G. Tzanetakis 1 / 32 Table of Contents I 1 Motivation and Terminology 2 Psychacoustics 3 F0

More information

158 ACTION AND PERCEPTION

158 ACTION AND PERCEPTION Organization of Hierarchical Perceptual Sounds : Music Scene Analysis with Autonomous Processing Modules and a Quantitative Information Integration Mechanism Kunio Kashino*, Kazuhiro Nakadai, Tomoyoshi

More information

Music Genre Classification and Variance Comparison on Number of Genres

Music Genre Classification and Variance Comparison on Number of Genres Music Genre Classification and Variance Comparison on Number of Genres Miguel Francisco, miguelf@stanford.edu Dong Myung Kim, dmk8265@stanford.edu 1 Abstract In this project we apply machine learning techniques

More information

A Bayesian Network for Real-Time Musical Accompaniment

A Bayesian Network for Real-Time Musical Accompaniment A Bayesian Network for Real-Time Musical Accompaniment Christopher Raphael Department of Mathematics and Statistics, University of Massachusetts at Amherst, Amherst, MA 01003-4515, raphael~math.umass.edu

More information

Outline. Why do we classify? Audio Classification

Outline. Why do we classify? Audio Classification Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification Implementation Future Work Why do we classify

More information

Automatic Piano Music Transcription

Automatic Piano Music Transcription Automatic Piano Music Transcription Jianyu Fan Qiuhan Wang Xin Li Jianyu.Fan.Gr@dartmouth.edu Qiuhan.Wang.Gr@dartmouth.edu Xi.Li.Gr@dartmouth.edu 1. Introduction Writing down the score while listening

More information

Introductions to Music Information Retrieval

Introductions to Music Information Retrieval Introductions to Music Information Retrieval ECE 272/472 Audio Signal Processing Bochen Li University of Rochester Wish List For music learners/performers While I play the piano, turn the page for me Tell

More information

Audio Feature Extraction for Corpus Analysis

Audio Feature Extraction for Corpus Analysis Audio Feature Extraction for Corpus Analysis Anja Volk Sound and Music Technology 5 Dec 2017 1 Corpus analysis What is corpus analysis study a large corpus of music for gaining insights on general trends

More information

A probabilistic approach to determining bass voice leading in melodic harmonisation

A probabilistic approach to determining bass voice leading in melodic harmonisation A probabilistic approach to determining bass voice leading in melodic harmonisation Dimos Makris a, Maximos Kaliakatsos-Papakostas b, and Emilios Cambouropoulos b a Department of Informatics, Ionian University,

More information

Narrative Theme Navigation for Sitcoms Supported by Fan-generated Scripts

Narrative Theme Navigation for Sitcoms Supported by Fan-generated Scripts Narrative Theme Navigation for Sitcoms Supported by Fan-generated Scripts Gerald Friedland, Luke Gottlieb, Adam Janin International Computer Science Institute (ICSI) Presented by: Katya Gonina What? Novel

More information

MUSI-6201 Computational Music Analysis

MUSI-6201 Computational Music Analysis MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)

More information

Music Database Retrieval Based on Spectral Similarity

Music Database Retrieval Based on Spectral Similarity Music Database Retrieval Based on Spectral Similarity Cheng Yang Department of Computer Science Stanford University yangc@cs.stanford.edu Abstract We present an efficient algorithm to retrieve similar

More information

Improving Polyphonic and Poly-Instrumental Music to Score Alignment

Improving Polyphonic and Poly-Instrumental Music to Score Alignment Improving Polyphonic and Poly-Instrumental Music to Score Alignment Ferréol Soulez IRCAM Centre Pompidou 1, place Igor Stravinsky, 7500 Paris, France soulez@ircamfr Xavier Rodet IRCAM Centre Pompidou 1,

More information

A Survey of Audio-Based Music Classification and Annotation

A Survey of Audio-Based Music Classification and Annotation A Survey of Audio-Based Music Classification and Annotation Zhouyu Fu, Guojun Lu, Kai Ming Ting, and Dengsheng Zhang IEEE Trans. on Multimedia, vol. 13, no. 2, April 2011 presenter: Yin-Tzu Lin ( 阿孜孜 ^.^)

More information

Speech Recognition and Signal Processing for Broadcast News Transcription

Speech Recognition and Signal Processing for Broadcast News Transcription 2.2.1 Speech Recognition and Signal Processing for Broadcast News Transcription Continued research and development of a broadcast news speech transcription system has been promoted. Universities and researchers

More information

Proceedings of Meetings on Acoustics

Proceedings of Meetings on Acoustics Proceedings of Meetings on Acoustics Volume 19, 2013 http://acousticalsociety.org/ ICA 2013 Montreal Montreal, Canada 2-7 June 2013 Musical Acoustics Session 3pMU: Perception and Orchestration Practice

More information

... A Pseudo-Statistical Approach to Commercial Boundary Detection. Prasanna V Rangarajan Dept of Electrical Engineering Columbia University

... A Pseudo-Statistical Approach to Commercial Boundary Detection. Prasanna V Rangarajan Dept of Electrical Engineering Columbia University A Pseudo-Statistical Approach to Commercial Boundary Detection........ Prasanna V Rangarajan Dept of Electrical Engineering Columbia University pvr2001@columbia.edu 1. Introduction Searching and browsing

More information

Recognising Cello Performers using Timbre Models

Recognising Cello Performers using Timbre Models Recognising Cello Performers using Timbre Models Chudy, Magdalena; Dixon, Simon For additional information about this publication click this link. http://qmro.qmul.ac.uk/jspui/handle/123456789/5013 Information

More information

Reducing False Positives in Video Shot Detection

Reducing False Positives in Video Shot Detection Reducing False Positives in Video Shot Detection Nithya Manickam Computer Science & Engineering Department Indian Institute of Technology, Bombay Powai, India - 400076 mnitya@cse.iitb.ac.in Sharat Chandran

More information

Luwei Yang. Mobile: (+86) luweiyang.com

Luwei Yang. Mobile: (+86) luweiyang.com Luwei Yang Mobile: (+86) 17502530917 luwei.yang.qm@gmail.com luweiyang.com Personal Statement A machine learning researcher obtained PhD degree from Queen Mary University of London. Looking to secure the

More information

Retrieval of textual song lyrics from sung inputs

Retrieval of textual song lyrics from sung inputs INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Retrieval of textual song lyrics from sung inputs Anna M. Kruspe Fraunhofer IDMT, Ilmenau, Germany kpe@idmt.fraunhofer.de Abstract Retrieving the

More information

Automatic Classification of Instrumental Music & Human Voice Using Formant Analysis

Automatic Classification of Instrumental Music & Human Voice Using Formant Analysis Automatic Classification of Instrumental Music & Human Voice Using Formant Analysis I Diksha Raina, II Sangita Chakraborty, III M.R Velankar I,II Dept. of Information Technology, Cummins College of Engineering,

More information

Singing voice synthesis based on deep neural networks

Singing voice synthesis based on deep neural networks INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Singing voice synthesis based on deep neural networks Masanari Nishimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

More information

Singing Pitch Extraction and Singing Voice Separation

Singing Pitch Extraction and Singing Voice Separation Singing Pitch Extraction and Singing Voice Separation Advisor: Jyh-Shing Roger Jang Presenter: Chao-Ling Hsu Multimedia Information Retrieval Lab (MIR) Department of Computer Science National Tsing Hua

More information

Effects of acoustic degradations on cover song recognition

Effects of acoustic degradations on cover song recognition Signal Processing in Acoustics: Paper 68 Effects of acoustic degradations on cover song recognition Julien Osmalskyj (a), Jean-Jacques Embrechts (b) (a) University of Liège, Belgium, josmalsky@ulg.ac.be

More information

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance Methodologies for Expressiveness Modeling of and for Music Performance by Giovanni De Poli Center of Computational Sonology, Department of Information Engineering, University of Padova, Padova, Italy About

More information

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION ULAŞ BAĞCI AND ENGIN ERZIN arxiv:0907.3220v1 [cs.sd] 18 Jul 2009 ABSTRACT. Music genre classification is an essential tool for

More information

Interacting with a Virtual Conductor

Interacting with a Virtual Conductor Interacting with a Virtual Conductor Pieter Bos, Dennis Reidsma, Zsófia Ruttkay, Anton Nijholt HMI, Dept. of CS, University of Twente, PO Box 217, 7500AE Enschede, The Netherlands anijholt@ewi.utwente.nl

More information

POLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING

POLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING POLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING Luis Gustavo Martins Telecommunications and Multimedia Unit INESC Porto Porto, Portugal lmartins@inescporto.pt Juan José Burred Communication

More information

Arts, Computers and Artificial Intelligence

Arts, Computers and Artificial Intelligence Arts, Computers and Artificial Intelligence Sol Neeman School of Technology Johnson and Wales University Providence, RI 02903 Abstract Science and art seem to belong to different cultures. Science and

More information

Browsing News and Talk Video on a Consumer Electronics Platform Using Face Detection

Browsing News and Talk Video on a Consumer Electronics Platform Using Face Detection Browsing News and Talk Video on a Consumer Electronics Platform Using Face Detection Kadir A. Peker, Ajay Divakaran, Tom Lanning Mitsubishi Electric Research Laboratories, Cambridge, MA, USA {peker,ajayd,}@merl.com

More information

QUALITY OF COMPUTER MUSIC USING MIDI LANGUAGE FOR DIGITAL MUSIC ARRANGEMENT

QUALITY OF COMPUTER MUSIC USING MIDI LANGUAGE FOR DIGITAL MUSIC ARRANGEMENT QUALITY OF COMPUTER MUSIC USING MIDI LANGUAGE FOR DIGITAL MUSIC ARRANGEMENT Pandan Pareanom Purwacandra 1, Ferry Wahyu Wibowo 2 Informatics Engineering, STMIK AMIKOM Yogyakarta 1 pandanharmony@gmail.com,

More information

AUTOMATIC ACCOMPANIMENT OF VOCAL MELODIES IN THE CONTEXT OF POPULAR MUSIC

AUTOMATIC ACCOMPANIMENT OF VOCAL MELODIES IN THE CONTEXT OF POPULAR MUSIC AUTOMATIC ACCOMPANIMENT OF VOCAL MELODIES IN THE CONTEXT OF POPULAR MUSIC A Thesis Presented to The Academic Faculty by Xiang Cao In Partial Fulfillment of the Requirements for the Degree Master of Science

More information

Automatic Rhythmic Notation from Single Voice Audio Sources

Automatic Rhythmic Notation from Single Voice Audio Sources Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung

More information

Music in Practice SAS 2015

Music in Practice SAS 2015 Sample unit of work Contemporary music The sample unit of work provides teaching strategies and learning experiences that facilitate students demonstration of the dimensions and objectives of Music in

More information

Vuzik: Music Visualization and Creation on an Interactive Surface

Vuzik: Music Visualization and Creation on an Interactive Surface Vuzik: Music Visualization and Creation on an Interactive Surface Aura Pon aapon@ucalgary.ca Junko Ichino Graduate School of Information Systems University of Electrocommunications Tokyo, Japan ichino@is.uec.ac.jp

More information

Efficient Computer-Aided Pitch Track and Note Estimation for Scientific Applications. Matthias Mauch Chris Cannam György Fazekas

Efficient Computer-Aided Pitch Track and Note Estimation for Scientific Applications. Matthias Mauch Chris Cannam György Fazekas Efficient Computer-Aided Pitch Track and Note Estimation for Scientific Applications Matthias Mauch Chris Cannam György Fazekas! 1 Matthias Mauch, Chris Cannam, George Fazekas Problem Intonation in Unaccompanied

More information

Music Segmentation Using Markov Chain Methods

Music Segmentation Using Markov Chain Methods Music Segmentation Using Markov Chain Methods Paul Finkelstein March 8, 2011 Abstract This paper will present just how far the use of Markov Chains has spread in the 21 st century. We will explain some

More information