Normalized Cumulative Spectral Distribution in Music

Similar documents
INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting

GCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES

Audio classification from time-frequency texture

MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES

Audio-Based Video Editing with Two-Channel Microphone

Speech and Speaker Recognition for the Command of an Industrial Robot

Music Information Retrieval with Temporal Features and Timbre

Music Recommendation from Song Sets

Singer Identification

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

Appendix A Types of Recorded Chords

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon

POLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING

Automatic Laughter Detection

Chord Classification of an Audio Signal using Artificial Neural Network

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

Features for Audio and Music Classification

HIDDEN MARKOV MODELS FOR SPECTRAL SIMILARITY OF SONGS. Arthur Flexer, Elias Pampalk, Gerhard Widmer

Violin Timbre Space Features

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Automatic Identification of Instrument Type in Music Signal using Wavelet and MFCC

Proposal for Application of Speech Techniques to Music Analysis

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

A prototype system for rule-based expressive modifications of audio recordings

Efficient Vocal Melody Extraction from Polyphonic Music Signals

AMusical Instrument Sample Database of Isolated Notes

CS229 Project Report Polyphonic Piano Transcription

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

Acoustic Scene Classification

Automatic Rhythmic Notation from Single Voice Audio Sources

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors

Interactive Classification of Sound Objects for Polyphonic Electro-Acoustic Music Annotation

Music Genre Classification and Variance Comparison on Number of Genres

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

Topics in Computer Music Instrument Identification. Ioanna Karydi

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

Color Image Compression Using Colorization Based On Coding Technique

Music Radar: A Web-based Query by Humming System

Automatic Laughter Detection

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)

MUSICAL INSTRUMENT RECOGNITION USING BIOLOGICALLY INSPIRED FILTERING OF TEMPORAL DICTIONARY ATOMS

Efficient Implementation of Neural Network Deinterlacing

TERRESTRIAL broadcasting of digital television (DTV)

638 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 3, MARCH 2010

Study of White Gaussian Noise with Varying Signal to Noise Ratio in Speech Signal using Wavelet

MIRAI: Multi-hierarchical, FS-tree based Music Information Retrieval System

ON THE USE OF PERCEPTUAL PROPERTIES FOR MELODY ESTIMATION

Robert Alexandru Dobre, Cristian Negrescu

A Music Retrieval System Using Melody and Lyric

Available online at ScienceDirect. Procedia Computer Science 46 (2015 )

Music Database Retrieval Based on Spectral Similarity

LOUDNESS EFFECT OF THE DIFFERENT TONES ON THE TIMBRE SUBJECTIVE PERCEPTION EXPERIMENT OF ERHU

Line-Adaptive Color Transforms for Lossless Frame Memory Compression

Multi-label classification of emotions in music

Application Of Missing Feature Theory To The Recognition Of Musical Instruments In Polyphonic Audio

On the Characterization of Distributed Virtual Environment Systems

Wipe Scene Change Detection in Video Sequences

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE

Modified Spectral Modeling Synthesis Algorithm for Digital Piri

A Luminance Adjusting Algorithm for High Resolution and High Image Quality AMOLED Displays of Mobile Phone Applications

Phone-based Plosive Detection

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

Supervised Learning in Genre Classification

Data-Driven Solo Voice Enhancement for Jazz Music Retrieval

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

Singer Traits Identification using Deep Neural Network

The song remains the same: identifying versions of the same piece using tonal descriptors

LUT OPTIMIZATION USING COMBINED APC-OMS TECHNIQUE

Feature-based Characterization of Violin Timbre

A METHOD OF MORPHING SPECTRAL ENVELOPES OF THE SINGING VOICE FOR USE WITH BACKING VOCALS

Automatic Piano Music Transcription

Acoustic scene and events recognition: how similar is it to speech recognition and music genre/instrument recognition?

Classification of Timbre Similarity

A New Overlap-Scan Circuit for High Speed and Low Data Voltage in Plasma-TV

AN ON-THE-FLY MANDARIN SINGING VOICE SYNTHESIS SYSTEM

MELODY EXTRACTION BASED ON HARMONIC CODED STRUCTURE

ISSN ICIRET-2014

A probabilistic framework for audio-based tonal key and chord recognition

Automatic Classification of Instrumental Music & Human Voice Using Formant Analysis

Non Stationary Signals (Voice) Verification System Using Wavelet Transform

Improving Frame Based Automatic Laughter Detection

Temporal summation of loudness as a function of frequency and temporal pattern

Recognising Cello Performers using Timbre Models

Fast thumbnail generation for MPEG video by using a multiple-symbol lookup table

MEL-FREQUENCY cepstral coefficients (MFCCs)

Robust Radio Broadcast Monitoring Using a Multi-Band Spectral Entropy Signature

Tempo and Beat Tracking

A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION

Polyphonic Audio Matching for Score Following and Intelligent Audio Editors

Effects of acoustic degradations on cover song recognition

Subjective Similarity of Music: Data Collection for Individuality Analysis

Musical instrument identification in continuous recordings

MELODY EXTRACTION FROM POLYPHONIC AUDIO OF WESTERN OPERA: A METHOD BASED ON DETECTION OF THE SINGER S FORMANT

Content-based Music Structure Analysis with Applications to Music Semantics Understanding

Error Resilience for Compressed Sensing with Multiple-Channel Transmission

A Study of Predict Sales Based on Random Forest Classification

Semi-supervised Musical Instrument Recognition

Automatic morphological description of sounds

Transcription:

Normalized Cumulative Spectral Distribution in Music Young-Hwan Song, Hyung-Jun Kwon, and Myung-Jin Bae Abstract As the remedy used music becomes active and meditation effect through the music is verified, people tae a growing interest about psychological balance or remedy given by music. From traditional studies, it is verified that the music of which spectral envelop varies approximately as 1/f (f is frequency) down to a frequency of low frequency bandwidth gives psychological balance. In this paper, we researched signal properties of music which gives psychological balance. In order to find this, we derived the property from voice. Music composed by voice shows large value in NCSD. We confirmed the degree of deference between music by curvature of normalized cumulative spectral distribution. In the music that gives psychological balance, the curvature shows high value, otherwise, the curvature shows low value. Keywords Cognitive Psychology, Normalized Cumulative Spectral Distribution, Curvature. I. INTRODUCTION N automatic recognition and artificial intelligence, the I importance of emotional recognition is growing up because mechanism and algorithms are operated lie human with emotional thining. The methods to extract emotional information from signal and transfer through signal have been presented and developed. In speech signal, emotional recognition is performed by the difference of factors which is used for speech recognition through emotional changes. The methods to visualize these emotional properties of signal have also been developed and used. For music recognition, it is classified into genre classification [1], instrument classification [2], music search [3] etc and it is performed by 90 factors approximately. Many of these factors are from speech recognition and audio description of MPEG. It is medically proved that music effects on the psychological and mental health. So music is recently used for therapy and mediation [4]. Even music is developed and produced for these purposes [5]. Analysis of music which is used for therapy and mediation is performed by brain wave and psychological changes from the effects on the brain and body of human [6]. The music which is used for therapy or medical purpose is produced by using the similarity from the result of medical Manuscript received June 1, 2008. Y.H. Song, H.J. Kwon, and M.J. Bae are with Information and Telecommunication Engineering Department, Soongsil University, 1-1 Sangdo 5 dong DongJa-Ku, Seoul, 156-743, Republic of Korea (phone: +82-02-824-0906; fax: +82-02-824-0906; e-mail: song0one@yahoo.co.r, mswg1@naver.com, mjbae@ssu.ac.r). experiments [6]. Even though the effects of emotional factor of music on life and psychological health are proved, it is not enough to express and classify the music for emotional recognition by signal own properties since the factors which can represent emotional parts are not enough. In this paper, we researched how can express emotional elements in music objectively. So we proposed curvature of normalized cumulative spectral distribution as new factor that can represent signal property of music which gives psychological balance and improve emotional recognition rate of music. II. MUSIC COMPOSED BY VOICE Music is growing up with musical instruments. Some inds of music have only played with instruments, and it also played by voice with instruments or not. Richard F. Voss and john Clare researched that the music which is played for a long time is characterized as 1/f on low frequency. And this is revealed is speech [7, 8]. Namely, popular music has similar property to speech in low frequency. For music, Fourier analysis has used before. And the properties from music in frequency are applied into content-based music classification and recognition. Now we analyze the music only composed by voice. It is not just speech. After that it is compared with other music to now which music has the similar property. A. Normalized Cumulative Spectral Distribution In this paper, we propose Normalized Cumulative Spectral Distribution (NCSD) as factor which can express spectral distribution of music to compare with music only composed by voice. NCSD is defined by (1). log( S ( r)) min(log( S ( r)) S ( f) = dr f (1) 0 fs /2 log( S ( )) min(log( ( )) 0 S d With NCSD, we can confirm the relative spectral distribution. It also expresses cumulative energy with increasing frequency. In speech generation model, excitation convolved with vocal trac which means generation system. So spectral envelope followed around formant envelop in frequency. In formant of speech, generally energy is concentrated in low frequency. And it is decreased with increasing frequency. On the base of this, NCSD of speech is rapidly increased on low frequency, and an increase is 115

decreased with increasing frequency. The following figure shows spectrum and NCSD of music composed by voice. In this paper, we would lie to now property of popular music which maes people comfortable for a long time. So we used sacred song for this property. III. MUSIC COMPOSED WITH INSTRUMENTS To confirm property for music composed by instruments, we classified music into classic by many instruments, classic by one instrument, classic by electronic instrument, new age, ballad, jazz, dance, old pop, pop, roc and techno. Fig. 3 Fig. 13 show the spectrum and NCSD Fig. 1 Spectrum and NCSD of Sacred Song B. NCSD Curvature In order to represent the NCSD by one factor, we defined NCDS curvature as (2). Fig. 3 Spectrum and NCSD (a) Classic by All Instruments L _ envelop = (2) L straight Where L _ envelop is the curve length of NCSD. And Lstraight is length of solid line from a to b in Fig. 2. Fig. 2 shows the NCSD and effective bandwidth. Fig. 4 Spectrum and NCSD (b) Classic by One Instrument Fig. 2 Spectrum and Normalized Cumulative Spectral Distribution and Effective Bandwidth Effective bandwidth is decided when S ( f) is lower than threshold. The threshold is from experimental result. After that NCSD Curvature is extracted on effective bandwidth. is close to 1 when the NCSD curve is similar to solid line. If the NCSD line is curved a lot, will be large. So it can represent the NCSD and relative spectral distribution. Fig. 5 Spectrum and NCSD (c) Classic by Elec Instruments 116

Fig. 6 Spectrum and NCSD (d) New Age Fig. 9 Spectrum and NCSD (g) Jazz Fig. 7 Spectrum and NCSD (e) Ballad Fig. 10 Spectrum and NCSD (h) Old Pop Fig. 8 Spectrum and NCSD (f) Dance Music Fig. 11 Spectrum and NCSD (i) Pop 117

psychological balance to people. And it is generally qualified to many people for a long time or nowadays. Fig. 12 Spectrum and NCSD (j) Roc Fig. 13 Spectrum and NCSD () Techno Table I shows for all inds of music that is classified in this paper. On the table 1, of sacred song that is just made of voice is large. And Classic and New age have large point lie sacred song. They did not use voice but they show similar point with sacred song. But ballad, dance, jazz, old pop, pop, roc and techno have small point. Classic and New age give IV. CONCLUSION In music perception of cognitive psychological, the curvature of normalized cumulative spectral distribution is high for music which gives psychological balance. That means consonance of sounds in spectrum is well. Commercial music such as pop or techno is composed stimulatively and played by lots of musical instruments. In this time, each musical instrument played with each characteristic on the spectrum since each instrument has main part of spectrum different from the others. So it figures lie a horizontal line on the spectrum. It maes gradient of normalized cumulative spectral distribution to constant nearly. We express our emotion through voice. Let assume that voice is one of instrument. In this time, the music which is only composed by voice can effect a change to psychologies. This ind of music is characterized by the output of vocal organ. In this paper we focused about that music which is only composed by voice gives psychological balance. In this ind of music, energy is mainly distributed on the low frequency. Energy is decreased rapidly by the increasing frequency. And the spectrum is distributed for all auditory frequency bandwidth even though it is only composed by voice. In the music which gives psychological balance, it is established by each sound of instrument for entire harmony, so the energy is mainly distributed on the low frequency. Even though it is composed entirely of a lot of instruments, all instruments played lie one instrument without characterizing individually. So magnitude is decreased by increasing frequency. This property is the same as of that music which is only composed of vocal organ of human gives psychological balance. It is shown that the spectral property of music made of voice is revealed at music gives psychological balance regardless of musical composition of instrument. In this paper, we confirmed the property of music gives psychological balance by the normalized cumulative spectral distribution, and the curvature represented this property on the normalized cumulative spectral distribution. Music Genre TABLE I CURVATURE Sacred Song 53.59 Classic by all instruments 71.04 Classic by one instrument 123.87 Classic by one elec instrument 74.40 New Age 73.29 Ballad 7.02 Dance 13.50 Jazz 22.52 Old Pop 8.43 Pop 6.97 Roc 14.37 Techno 31.48 =NCSD Curvature ( ) REFERENCES [1] B. Logan and A. Salomon, "A Music Similarity Function based on Signal Analysis," Proceedings of the IEEE International Conference on Multimedia and Expo (ICME2001), pp.745-748, 2001. [2] Beth Logan, "Mel Frequency Cepstral Coefficients for music modeling," Proceedings of the First International Symposium on Music Information Retrieval (ISMIR), 2000. [3] L. Rabiner and B. H. Juang, Fundamentals of speech recognition, Englewood Cliffs, NJ: Prentice Hall, 1993. [4] G. Peeters, "A Large Set of Audio Features for Sound Description (Similarity and Classification) in the Cuidado Project," IRCAM, Tech. Rep., 2004. [5] S. Essid, G. Richard, B. David, "Instrument recognition in polyphonic music based on automatic taxonomies," IEEE Trans. on Aidio, Speech and Languge processing, Vol. 14, No. 1, pp. 68-80, Jan. 2006 [6] S. Essid, G. Richard, B. David, "musical instrument recognition based on class pairwise feature selection," in Proc. 5th Int. Conf. Music Information Retrieval (ISMIR), Barcelona, Spain, Oct. 2004. 118

[7] R. F. Voss and J. Clare, 1/f noise in music and speech, Nature, Vol. 258, Issue 5533, pp. 317-318, 1975. [8] R. F. Voss and J. Clare, 1/f noise in music: Music from 1/f noise, J. Acoust. Soc. Am., Vol. 63, pp. 258-263, Jan. 1978. Y.H. Song (M 07) became a Member (M) of ASK in 2007. He is from Republic of Korea and was born in 1981. Next, He received the B.S. degree in Electronic Engineering from Soongsil University in 2007. He is currently the under M.S. degree at Soongsil University in Seoul, Korea. He received the military service from Jun, 2001 to Aug, 2003. He researched The Identification of Sound Source for Infrasound with the department of national defense. His research interests include speech signal processing, speech synthesis, speech recognition, speech coding, and audio coding. Mr. Song became a Member of ASK in 2007. 119