A method of subject extension pitch extraction for humming and singing signals

Similar documents
OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

Topic 4. Single Pitch Detection

2. AN INTROSPECTION OF THE MORPHING PROCESS

AN ALGORITHM FOR LOCATING FUNDAMENTAL FREQUENCY (F0) MARKERS IN SPEECH

Detection and demodulation of non-cooperative burst signal Feng Yue 1, Wu Guangzhi 1, Tao Min 1

THE importance of music content analysis for musical

Music Radar: A Web-based Query by Humming System

Query By Humming: Finding Songs in a Polyphonic Database

Study of White Gaussian Noise with Varying Signal to Noise Ratio in Speech Signal using Wavelet

CSC475 Music Information Retrieval

Voice & Music Pattern Extraction: A Review

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

Music Source Separation

Improving Frame Based Automatic Laughter Detection

Measurement of overtone frequencies of a toy piano and perception of its pitch

Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng

Department of Electrical & Electronic Engineering Imperial College of Science, Technology and Medicine. Project: Real-Time Speech Enhancement

Type-2 Fuzzy Logic Sensor Fusion for Fire Detection Robots

Efficient Computer-Aided Pitch Track and Note Estimation for Scientific Applications. Matthias Mauch Chris Cannam György Fazekas

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

Pitch-Synchronous Spectrogram: Principles and Applications

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

A NEW LOOK AT FREQUENCY RESOLUTION IN POWER SPECTRAL DENSITY ESTIMATION. Sudeshna Pal, Soosan Beheshti

MELODY EXTRACTION FROM POLYPHONIC AUDIO OF WESTERN OPERA: A METHOD BASED ON DETECTION OF THE SINGER S FORMANT

A Parametric Autoregressive Model for the Extraction of Electric Network Frequency Fluctuations in Audio Forensic Authentication

Phone-based Plosive Detection

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis

A Music Retrieval System Using Melody and Lyric

Speech and Speaker Recognition for the Command of an Industrial Robot

Pitch. The perceptual correlate of frequency: the perceptual dimension along which sounds can be ordered from low to high.

DELTA MODULATION AND DPCM CODING OF COLOR SIGNALS

Topic 10. Multi-pitch Analysis

MELODY EXTRACTION BASED ON HARMONIC CODED STRUCTURE

EXPLORING THE USE OF ENF FOR MULTIMEDIA SYNCHRONIZATION

Efficient Vocal Melody Extraction from Polyphonic Music Signals

Similarity Measurement of Biological Signals Using Dynamic Time Warping Algorithm

An Efficient Low Bit-Rate Video-Coding Algorithm Focusing on Moving Regions

Melody transcription for interactive applications

Broken Wires Diagnosis Method Numerical Simulation Based on Smart Cable Structure

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

Audio-Based Video Editing with Two-Channel Microphone

Automatic Rhythmic Notation from Single Voice Audio Sources

Getting Started. Connect green audio output of SpikerBox/SpikerShield using green cable to your headphones input on iphone/ipad.

HUMMING METHOD FOR CONTENT-BASED MUSIC INFORMATION RETRIEVAL

Tempo and Beat Analysis

UNIVERSAL SPATIAL UP-SCALER WITH NONLINEAR EDGE ENHANCEMENT

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

A Parametric Autoregressive Model for the Extraction of Electric Network Frequency Fluctuations in Audio Forensic Authentication

Comparing Pitch Detection Algorithms for Voice Applications

Spectrum Analyser Basics

1 Introduction to PSQM

hit), and assume that longer incidental sounds (forest noise, water, wind noise) resemble a Gaussian noise distribution.

Comparison Parameters and Speaker Similarity Coincidence Criteria:

NON-UNIFORM KERNEL SAMPLING IN AUDIO SIGNAL RESAMPLER

Pattern Recognition in Music

On Human Capability and Acoustic Cues for Discriminating Singing and Speaking Voices

International Journal of Engineering Research-Online A Peer Reviewed International Journal

Investigation of Digital Signal Processing of High-speed DACs Signals for Settling Time Testing

Reducing False Positives in Video Shot Detection

Single Channel Speech Enhancement Using Spectral Subtraction Based on Minimum Statistics

AN ON-THE-FLY MANDARIN SINGING VOICE SYNTHESIS SYSTEM

TERRESTRIAL broadcasting of digital television (DTV)

White Noise Suppression in the Time Domain Part II

Singer Recognition and Modeling Singer Error

DETECTION OF PITCHED/UNPITCHED SOUND USING PITCH STRENGTH CLUSTERING

Pitch correction on the human voice

Adaptive decoding of convolutional codes

Analysis, Synthesis, and Perception of Musical Sounds

Video coding standards

REPORT DOCUMENTATION PAGE

ECG SIGNAL COMPRESSION BASED ON FRACTALS AND RLE

Removal of Decaying DC Component in Current Signal Using a ovel Estimation Algorithm

SINGING PITCH EXTRACTION BY VOICE VIBRATO/TREMOLO ESTIMATION AND INSTRUMENT PARTIAL DELETION

Music Perception with Combined Stimulation

Singing voice synthesis based on deep neural networks

EE-217 Final Project The Hunt for Noise (and All Things Audible)

Figure 1: Feature Vector Sequence Generator block diagram.

EE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function

N T I. Introduction. II. Proposed Adaptive CTI Algorithm. III. Experimental Results. IV. Conclusion. Seo Jeong-Hoon

Calibrate, Characterize and Emulate Systems Using RFXpress in AWG Series

REpeating Pattern Extraction Technique (REPET): A Simple Method for Music/Voice Separation

Various Applications of Digital Signal Processing (DSP)

Subjective Similarity of Music: Data Collection for Individuality Analysis

LOUDNESS EFFECT OF THE DIFFERENT TONES ON THE TIMBRE SUBJECTIVE PERCEPTION EXPERIMENT OF ERHU

Hidden melody in music playing motion: Music recording using optical motion tracking system

Speech Enhancement Through an Optimized Subspace Division Technique

Guidance For Scrambling Data Signals For EMC Compliance

Music Segmentation Using Markov Chain Methods

BER MEASUREMENT IN THE NOISY CHANNEL

The Effect of Time-Domain Interpolation on Response Spectral Calculations. David M. Boore

A METHOD OF MORPHING SPECTRAL ENVELOPES OF THE SINGING VOICE FOR USE WITH BACKING VOCALS

CERIAS Tech Report Preprocessing and Postprocessing Techniques for Encoding Predictive Error Frames in Rate Scalable Video Codecs by E

Binning based algorithm for Pitch Detection in Hindustani Classical Music

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

TIME-DOMAIN TWO-DIMENSIONAL PITCH DETECTION. Gerard Benbassat TECHNICAL REPORT NO December 30, 1975

Design of Speech Signal Analysis and Processing System. Based on Matlab Gateway

Interface Practices Subcommittee SCTE STANDARD SCTE Measurement Procedure for Noise Power Ratio

Vocoder Reference Test TELECOMMUNICATIONS INDUSTRY ASSOCIATION

Available online at ScienceDirect. Procedia Computer Science 46 (2015 )

Transcription:

International Conference on Computer Science and Electronic Technology (CSET 2016) A method of subject extension pitch extraction for humming and singing signals Zhang Jinghui, Yang Shen, Wu Huahua School of Information Science and Engineering, Wuhan University of Science and Technology Abstract Pitch extraction is a key task of ahumming query The sound of humming or singing are both producedfrom system.the purpose of this paper isto find a method to extract the vibration of vocal cords, but the two kinds of production pitch accuratelyfortwo input modes, humming and sing- are not the same. The designed pitch extraction algorithm in ing.according to the characteristics of Chinese pronunciation, the paper can apply to different kinds of input signal by ana- this paper presentsa new method, namely the Pitch Extraction of lyzing their different features. The following will be described Subject Extension.According to the differencesbetween humming in detail. and singing signals, this method respectively chooses energy II. threshold and ratio of energy and entropy to detect endpoint of THE METHOD OF PITCH EXTRACTION notes.the candidate method and the shortest distance method The ideas of this paper comes from pitch extraction in are used to determine the pitch periods of the voiced segments, Chinese speech languages. Based on this, the paper proposes and subject extension method is used to determine the pitch pe- the Subject Extension approaches based on a single note of riods of the voiced/unvoiced mixed segments. Finally this algo- humming or singing signal. The main steps of the algorithm rithm is implemented in a small database, andis compared wi- are shown in Fig.1. thother similar algorithms. Experiments show our algorithm is Noise reduction and pretreatment more accurate and robust. Keywords humming query; pitch extraction; The endpoint detection Pitch extraction Post treatment endpoint detection; subject extension. Fig. 1. the main steps of the algorithm. I. INTRODUCTION A. Noise reduction and pretreatment Humming query is a typical application of content-based music retrieval and it usually has two kinds of input me- In the practical environment, the performance of recording thods[1]: humming and singing. Humming is a limited input, is usually affected by the background noise[6].so, the first requiring users to hum a tune by dada [2]. However, singing thing is reducingthe noise of the sampling frequency of 8000 is a more friendly and more casual input mode, users can sing Hz s input signal by the spectral subtraction. Then a lyrics directly. The pitch detection for the input signal is an pre-emphasiz filter is meant to compensate the high frequency important step in thehumming query system. components attenuation[7]. A prefilter using elliptic bandpass filter, which the boundary frequency is 60Hz and 3400Hz, is In recent years, with the development of wavelet transform, designed to remove the 50Hz power frequency noise and the the pitch extraction method aiming at humming and singing interference source frequency over half of the sampling fre- has made great progress. Upadhyay et al has proposed a me- quency[8]. Finally the sampling frequency is added window thod based on the variational mode decomposition and Hilbert and frame partition by hamming window, whose length is set transform for the instantaneous pitch frequency extraction[3]. 40ms and frame shift is set 10ms[9]. OnurBabacan et al has compared the recent mainstream pitch B. Endpoint detection extraction method[4]. The literature[5]is introduced the most widely used ACF method which is also used in this paper, and The composition of humming or singing can be divided in- the feature of ACF is simple and with higher precision. 2016. The authors - Published by Atlantis Press to unvoiced and voiced parts, and only the voiced part can be 304

extractedan accurate pitch frequency. The humming signal usually requires to be a plosive pronunciation, so the difference between adjacent time-domain waveform amplitude is very large, it can directly use the energy threshold to find the starting position of each note. But for the singing signal, because the singer is maybe non-professional or the lips radiation effects, the record of singing signal always takes noise and other uncertain influence. The simple use of energy for endpoint is hard to meet the requirements of actual note segmentation. This would requires the Energy entropy ratio which can still keep good performance in low signal-to-noise ratio conditions in detecting the voiced / unvoiced ingredients. The process of detection becomes more complex, andbetween the voiced / unvoiced ingredients also must be subdividedto a transition section. The specific division is shown in Fig.2. Humming A single note segmentation of humming Singing A single note segmentation of singing For the humming signal, after the endpoint detection, The pitch detection directly using ACF on audio segment, then use the shortest distance criterion and interpolation for elimination outliers. The pitch extraction of singing voiced segment is similar to the humming, but the extension of voiced interval is more complex. The main process of pitch extraction of singing as shown in Fig.3. 1) The candidates of pitch values Because the division of the voiced segment is not very strict and music signals are very complex, there will be outliers even in voiced parts to extract the pitch. When calculated the data of each frame in voiced parts, we would find three peaks and their corresponding positions in the range of pp mmmmmm ~pp mmmmmm (pp mmmmmm andpp mmax represent maximum and minimum estimated period respectively). The values of these three corresponding positions will be treated as pitch candidates and saved in the arraypp tttt, PP tttt retained only the positions and its sequences in accordance with the peak amplitude. The candidate values of pitch period as shown in Fig.4. Noise or unvoiced segments C. Pitch extraction The forward transition section Voiced segments main part of pitch extraction The backward transition section Fig.2. the specific division for humming and singing pitch period 150 100 50 0 first second third 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 Singing signal the sample points Input Fig.4. candidate values of pitch period The voiced by EER? YES Calculation three candidate by ACF Find the best candidate by minimum distance criterion Eliminating outliers Pitch period NO The pitch extraction in the extend range of voiced segment Calculation three candidate by ACF The Minimum distance criterion in the extension interval Whether the outliers is the tail YES value NO Eliminating outliers by interpolation method Modified the endpoint information Set the estimate of pitch period to zero After obtained the candidates, we will set a reasonable confidence interval. We set the position of the maximum ACF peak in the iiiihframe as KKKKKK = PP tttt (1, ii), calculate its means mmmmmmmmmm1 and standard deviations ssssssssss1, then the confidence interval is [mmmmmmmmmm1 ttheeeeeeee1, mmmmmmmmmm1 + ttheeeeeeee1].in most case, the standard deviations is very small, we don t need to deal with them further, but sometimes the standard deviations will be very large, this indicates that the KKKKKK values are oscillation. The value of KKKKKK in confidence interval is assigned to a parameterpppppp, and others are set to zero. Fig.3. the main process of pitch extraction of singing 2) The shortest distance criterion of pitch extraction In thepppppp, the nonzero partis marked aspppppp_nnnnnn, and we need to make the second confidence interval division for this part by calculating the mean mmmmmmmmmm2 and standard 305

tionsttheeeeeeee2ofpppppp(pppppp_nnnnnn) again. Now the second confidence interval is within [mmmmmmmmmm2 ttheeeeeeee2, mmmmmmmmmm2 + ttheeeeeeee2], then PPPPPPPPPPPP is copied frompppppp. If values in PPPPPP are out in the secondconfidence interval, the PPPPPPPPPPPP values are set to 0. Thus some samples of large deviationscould be abandoned. For the points in PPPPPPPPPPPP whose value is zero, we will search the best value between the front and back voiced segments by the shortest distance criterion. We assume that the iitth frame inppppppppppppis nonzero, and the ii + 1tth(orii 1tth) frame is zero. The pitch candidate value of the ii + 1tth column ispppppp(, ii + 1). In thepppppp(, ii + 1), we search an element which distance from PPPPPPPPPPPP(ii) is minimized. Assumed that the position is mark asmmmm, and we need to take PPPPPP(mmmm, ii + 1) for further judgment: Requested the pitch difference between two adjacent frames is no more than cc1 sampling periods (cc1 is a threshold, ranging between 10 and 15). Formula (1) shows this constraint for judgment conditions. PPPPPP mmmm,ii + 1 PPPPPPPPPPPP(ii) cc1 (1) This method can correct some pitch periods. However, because the voiced segment division isn t very strict, it s possible to produce the transition interval of singing, or the voiced segments is inherently unstable. If the formula (1) isn t satisfied, the value of PPPPPPPPPPPP(ii + 1) is find out only by meant of stacking or interpolation. D. The pitch extraction in the extend range of voiced segment 1) Calculation of the extended range and length Extension interval belongs to the singing section, but doesn t belong to the voiced segment. It s the overlapping part of the voiced and unvoiced. We has been obtained the information about endpoint and voiced segments by endpoint detection, but in the division of the unvoiced segments (voicing aliasing) there still exists unstable pitch information. When we have obtained the information about voiced segments, we begin to (before and after) extend the transition interval in the music signal, and the extending length is determined by the singing section and the voiced segments together.in the relationship between the singing section and the voiced segments we define as follows: in a singing section of song, the first voiced segments are both forward and backward extended, the rest of the voiced segments only extends posteriorly. 2) Autocorrelation function calculation For each frame of extended interval we still extract three pitch candidates among them by the ACF. And the pitch values corresponding to the magnitude of the ACF peak were sorted in descending order and saved inpp tttt. But the difference is that the periodic in the extended interval of music signals is not as strong as the voiced s, So the value of pitch extraction by ACF is very uncertain.the resulting in an extended interval value often meets the requirements off and on, and there are still mess data. 3) Minimum distance criterion Because the periodicity performance in the extension interval becomes worse and the region of variable composition becomes more complicated, the most value of pitch extraction in the extended interval can t reflect the actual pitch value. When we are doing the pitch detection, the first thing is to find a suitable pitch period by the minimum distance criterion. If the value is not existed, it s required that the threshold between the different of two frames is no more thancc1, or the different of more than two frame is no more thancc2. PP(ii) PP(ii + 1) cc1 (2) PP(ii) PP(ii + jj) cc2, jj = ±2, ±3, (3) Sometimes because it s impossible to seek the adjacent frame that are less than cc1 in the process of pitch extraction between the transition intervals, it only can look for by a frame or frames. 4) Post treatment The value of pitch extraction in the extended interval by the shortest distance from the pitch detection, in the corresponding position of the three formant, can t meet the conditions of the numerical.so we can set the estimate of pitch period PPPPPPPPPP(ii) is zero, and the further processing has to wait until the end of calculation in the extended interval.there are three kinds of situations for PPPPPPPPPP(ii) in the extended interval: the head, middle and tail. For the head and the middle part of the extended interval, the post treatment are used in the interpolation method. And for the tail of the extended interval. It s considered that the inaccurate endpoint detection cased such results and unvoiced segment is divided into the voiced segment incorrect. So we will modified the starting point in the music signal. 306

III. EXPERIMENTAL RESULTS AND ANALYSIS A. Experimental results The experimental data was recorded in a quiet laboratory. 15 boys and girls were invited to be recorded a song through two methods of humming and singing. There are four record versions of this song: humming by a male, singing by a male, humming by a female, and singing by a female. In this paper the volunteers hummed or sang the lyrics 我一直都在流浪 from the song "Cruel Moonlight". The effect is shown in Fig.5 and Fig.6. Fig.9. the pitch extraction result of the singing signals The vertical coordinate of Fig.9 and Fig.10 is the sampling points, the sampling interval is 1/8ms. From Figure 9, it can be seen that the algorithm is able to extract the pitch very well from humming or singing signal. The result of Fig.10 (b) is come from the pitch extraction by this paper s algorithm based on the Fig.10(a). It s fortunate that the result of the Subject Extension of Pitch Extraction don t appear the errors of most half frequency and double. Fig.5. the endpoints of a humming signal by a female Fig.6. the endpoints of a singing signal by a female Fig.10. the pitch detection results ofthe singing signal. (a) The voiced fragment pitch periods (b) the pitch periods after the subject extension Fig.7: the musical notation of lyrics " 我一直都在流浪 " The solid line shows the start of the voiced segments, and the dotted line indicating the end of the voiced segments. From Fig.5 and Fig.6, it can be seen that the results of humming signal are much better than the singing one. We have the singing section further to be divided into the voiced segments. The result is shown in Fig.8. B. Compared with several method In order to test the performance of the proposed algorithm in this paper, it will be compared with three methods:1 three-level cross-correlation method; 2the combination method of ACF and AMDF[10]; 3linear prediction and cepstrum method. Fig.11 and Fig.12 show the comparison result.it can be seen that for humming and singing signals, all of the method1,2,3 have some errors of multiple or half frequency. The smoothing filter can remove most of outliers, but there are still some unreasonable pitch periods. It s obvious that the algorithm proposed in this paper have the best performance. Fig.8. the voiced part segmentation for the singing signal Bycomparing Fig.7 withfig.8, it can be seen that Pitch detection results of this algorithm is able to detect the pitch of the voiced segment very well. 307

Fig.13. the results of four pitch extraction algorithms under different SNR. Fig.11. results of four pitch extraction algorithms for the humming signal by a female. Among them (a) method1 (b) method2 (c) method3 (d) the proposed method Fig.14. the results of four pitch extraction algorithms under different SNR. Fig.12. results of four pitch extraction algorithms for the singing signal by a female. Among them (a) method1 (b) method2 (c) method3 (d) proposed method In order to verify the robustness of the proposed algorithm, we respectively add Gaussian white noises of SNR=-5dB, 5dB, 10dB and 15dB to original humming and singing signals. Fig.13 and 14 show the experiment results. For the humming signal Fig.13 show that the proposed method extracts pitch periods accurately under the conditions of SNR=5dB, 10dB and 15dB, and there is no obvious outliers. However all of the other methods are appeared outliers in some extent. It also can be seen that the detect effects of this four methods are decreased quickly, but the proposed algorithm still can basically ensure the accuracy of pitch detection. For singing signal Fig.14 show that, under the condition of SNR=10dB and 15dB, the proposed algorithm is substantially unaffected by the noise. When SNR=5dB, it still can ensure the accuracy of pitch extraction, but the voiced interval is extended excess. Because the voiced/unvoiced segments can t be estimated correctly under the condition of SNR=-5dB, the process of pitch extraction is influenced. IV. CONCLUSION In this paper, we have present a method of subject extension pitch extraction for humming and singing signals. This method respectively choose energy threshold and ratio of energy and entropy to divide into the voiced and unvoiced segment. Then ACF method is used to extract the pitch periods of the voiced segments. The candidate method and the shortest distance method are used to improve the accuracy of pitch detection. Finally this algorithm is compared with other similar algorithms. Experiments showed our algorithm hadhigher accuracy and better anti-noise performance. 308

V. REFERENCES [1] Li, Zhou Mingquan, Xia Xiaoliang etc.. Improved pitch detection method and in music retrieval application [J]. Computer engineering and applications, 2011, 47 (6): 127-130. [2] Chhayani, N.H, Patil, H. development of corpora for person recognition using humming, singing and speech[c] Oriental Cocosda held jointly with 2013 Con-ference on Asian spoken language research and Evaluation. IEEE 2013:1-6. [3] Upadhyay A, Pachori R B. A new method for determination of instantaneous pitch frequency from speech signals[c]// Signal Processing and Signal Processing Education Workshop. IEEE, 2015. [4] Babacan O, Drugman T, D'Alessandro N, et al. A comparative study of pitch extraction algorithms on a large variety of singing sounds[c]// Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on. IEEE, 2014:7815-7819. [5] Rabiner L. On the use of autocorrelation analysis for pitch detection[j]. Acoustics Speech & Signal Processing IEEE Transactions on, 2012, 25(1):24-33. [6] Yuan Gang, Liu Zhikun, Tang Xiaoming, et al. Study on the mechanism of speech enhancement algorithms [C].2008 Annual Conference on communication theory and signal processing.2008:402-407. [7] Xing Weili. Research and implementation of content based audio retrieval technology [D]. Northwestern University, 2004 [8] Li Xuelong. Design and implementation of music retrieval system based on melody matching [D]. Beijing University of Technology, 2010 [9] Zi Lin. Humming music retrieval system research and design based on [D]. University of Electronic Science and technology, 2011 [10] Li Zhijun, Yin Xia. An improved algorithm [J]. audio technology, pitch detection based on AMDF and ACF 2011, 35 (1): 50-52. 309