Singing Voice Conversion Using Posted Waveform Data on Music Social Media

Similar documents
Singing voice synthesis based on deep neural networks

Promises and challenges of electronic journals 169. Heting Chu Palmer School of Library & Information Science, Long Island University, NY, USA

Perceptual Quantiser (PQ) to Hybrid Log-Gamma (HLG) Transcoding

A DISPLAY INDEPENDENT HIGH DYNAMIC RANGE TELEVISION SYSTEM

Automatic Laughter Detection

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

Automatic Rhythmic Notation from Single Voice Audio Sources

DXR.1 Digital Audio Codec

Singer Traits Identification using Deep Neural Network

Audio-Based Video Editing with Two-Channel Microphone

Singer Identification

Chord Classification of an Audio Signal using Artificial Neural Network

A Comparative Study of Spectral Transformation Techniques for Singing Voice Synthesis

Automatic Laughter Detection

Life Science Journal 2014;11(6)

Outline. Introduction to number systems: sign/magnitude, ones complement, twos complement Review of latches, flip flops, counters

Subjective Similarity of Music: Data Collection for Individuality Analysis

Advanced Signal Processing 2

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Lab 3 : CMOS Sequential Logic Gates

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION

On Human Capability and Acoustic Cues for Discriminating Singing and Speaking Voices

1. Introduction NCMMSC2009

JAMIA. Information Information for Authors

A Music Retrieval System Using Melody and Lyric

Towards Complexity Studies of Indonesian Songs

Efficient Computer-Aided Pitch Track and Note Estimation for Scientific Applications. Matthias Mauch Chris Cannam György Fazekas

THE importance of music content analysis for musical

Speech and Speaker Recognition for the Command of an Industrial Robot

Music Genre Classification and Variance Comparison on Number of Genres

Improving Frame Based Automatic Laughter Detection

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

An Efficient Low Bit-Rate Video-Coding Algorithm Focusing on Moving Regions

ISSN ICIRET-2014

SINGING COMPANION LESSON BOOK

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon

Expressive Singing Synthesis based on Unit Selection for the Singing Synthesis Challenge 2016

EXPLORING THE USE OF ENF FOR MULTIMEDIA SYNCHRONIZATION

Christine Baldwin Project Manager, SuperJournal. David Pullinger Project Director, SuperJournal

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

Lecture Notes 12: Digital Cellular Communications

2. AN INTROSPECTION OF THE MORPHING PROCESS

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

Acoustic Scene Classification

Experiments on musical instrument separation using multiplecause

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

Music Information Retrieval Community

Robert Alexandru Dobre, Cristian Negrescu

AUTOMATIC TIMBRE CLASSIFICATION OF ETHNOMUSICOLOGICAL AUDIO RECORDINGS

Gaussian Mixture Model for Singing Voice Separation from Stereophonic Music

Keywords Separation of sound, percussive instruments, non-percussive instruments, flexible audio source separation toolbox

SYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors

Automatic Piano Music Transcription

MODELS of music begin with a representation of the

Music Source Separation

Retrieval of textual song lyrics from sung inputs

Topics in Computer Music Instrument Identification. Ioanna Karydi

Computer Organization

hit), and assume that longer incidental sounds (forest noise, water, wind noise) resemble a Gaussian noise distribution.

An Efficient Test Pattern Generator -Mersenne Twister-

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

AUDIO KEY LINKS: PLAYBACK DEVICES IMPROVEMENT IST PRESTO Preservation Technologies for European Broadcast Archives

On human capability and acoustic cues for discriminating singing and speaking voices

FULL-AUTOMATIC DJ MIXING SYSTEM WITH OPTIMAL TEMPO ADJUSTMENT BASED ON MEASUREMENT FUNCTION OF USER DISCOMFORT

Voice & Music Pattern Extraction: A Review

Music Recommendation from Song Sets

The Ukulele Circle of Fifths - Song Structure Lesson

MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES

AUTOMATIC IDENTIFICATION FOR SINGING STYLE BASED ON SUNG MELODIC CONTOUR CHARACTERIZED IN PHASE PLANE

By Jon R. Davids, MD, Daniel M. Weigl, MD, Joye P. Edmonds, MLIS, AHIP, and Dawn W. Blackhurst, DrPH

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis

CS229 Project Report Polyphonic Piano Transcription

Research Article. ISSN (Print) *Corresponding author Shireen Fathima

A System for Automatic Chord Transcription from Audio Using Genre-Specific Hidden Markov Models

Tempo and Beat Analysis

GYROPHONE RECOGNIZING SPEECH FROM GYROSCOPE SIGNALS. Yan Michalevsky (1), Gabi Nakibly (2) and Dan Boneh (1)

Singer Recognition and Modeling Singer Error

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES

Application Of Missing Feature Theory To The Recognition Of Musical Instruments In Polyphonic Audio

Statistical Modeling and Retrieval of Polyphonic Music

A Bootstrap Method for Training an Accurate Audio Segmenter

Recognising Cello Performers Using Timbre Models

Supervised Learning in Genre Classification

GRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM

Predicting Time-Varying Musical Emotion Distributions from Multi-Track Audio

Music Segmentation Using Markov Chain Methods

ON FINDING MELODIC LINES IN AUDIO RECORDINGS. Matija Marolt

A repetition-based framework for lyric alignment in popular songs

Figure 1: Feature Vector Sequence Generator block diagram.

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES

Release Year Prediction for Songs

UNIVERSAL SPATIAL UP-SCALER WITH NONLINEAR EDGE ENHANCEMENT

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University

REpeating Pattern Extraction Technique (REPET): A Simple Method for Music/Voice Separation

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

Analysis of Packet Loss for Compressed Video: Does Burst-Length Matter?

Transcription:

Singing Voice Conversion Using Poste Waveform Data on Music Social Meia Koki Sena, Yukiya Hono, Kei Sawaa, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku an Keiichi Tokua Department of Computer Science an Engineering, Nagoya Institute of Technology, Nagoya, Japan E-mail: {kksn924, hono, swkei, bonanza, uratec, nankaku, tokua}@sp.nitech.ac.jp Tel: +81-52-735-5479 Abstract This paper proposes a metho of selecting training ata for many-to-one singing voice conversion (VC) from ata on the social meia music app nana. On this social meia app, users can share souns such as speaking, singing, an instrumental music recore by their smartphones. The number of hours of accumulate ata has exceee one million, an it is regare as big ata. It is wiely known that big ata can create huge values by avance eep learning technology. A lot of post ata of multiple users having sung the same song is containe in nana s atabase. This ata is consiere suitable training ata for VC. This is because VC frameworks base on statistical approaches often require parallel ata sets that consist of pairs of ata of source an target singers who sing the same phrases. The propose metho can compose parallel ata sets that can be use for many-to-one statistical VCs from nanas atabase by extracting frames that have small ifferences in the timing of utterances, base on the results of ynamic programming (DP) matching. Experimental results inicate that a system that uses training ata compose by our metho can convert acoustic more accurately than a system that oes not use the metho. I. INTRODUCTION Social meia has mae it possible for people all over the worl to transmit their information. There are many kins of social meia websites an apps, such as YouTube, Facebook, an Instagram, an a large amount of ata transmitte by users has been accumulating. This big ata has increasing potential in creating values for every fiel[1]. Recently, machine learning methos of ealing with big ata have been wiely researche in many institutes an laboratories. The Multi-Genre Broacast (MGB) Challenge an official challenge of the IEEE Automatic Speech Recognition an Unerstaning Workshop (ASRU) is one of the international workshops evaluating big ata technologies relate to the speech fiel[3]. The challenge at ASRU 2015 was an evaluation of transcription[6], [7], speaker iarization[4], [5], ialect etection, an lightly supervise alignment using approximately 1,600 hours of British Broacasting Corporation (BBC) television program ata that ha been recore. Commercialize proucts using eep learning technology alreay exist, such as the speech recognition system employe in Google Home, which was traine using tens of thousans of hours of speech ata[2]. The social meia app nana [8] stores big ata of music ata. This social meia app is esigne to allow users to share singing an instrumental souns easily using their smartphone. Up to more than one million hours of ata have been uploae, an the amount continues to increase. In nana, users can collaborate with other users uploae posts by overubbing other user s souns with their own souns. In particular, accompaniment posts of popular songs are collaborate on by many users. The relationship between collaborating an collaborate posts is represente by a tree structure. Because each tree generally consists of one song, the same song is sung in almost all singing posts of each tree. A atabase that contains a large number of posts in which the same songs are sung by multiple users has large potential. This ata can be use for a parallel ata set, the training ata of voice conversion (VC). VC is a metho of converting a speaker s voice into another kin of voice, especially another speaker s voice, while maintaining linguistic information. Statistical approaches have been wiely researche[9], [10]. The conventional statistical VC is often base on a Gaussian mixture moel (GMM) [11]} More recently, a VC framework base on eep neural networks (DNNs) has been propose[12], [13]. These statistical VC typically train statistical moels by using a parallel ata set that consists of pairs of speech ata from source an target speakers uttering the same sentences. Not every ata can be use when the nana s ata is use for VC training ata. For example, the atabase contains some partly sung ata. We propose a metho base on ynamic programming (DP) matching for composing training ata from the atabase. The target ata an the other ata in the same collaboration tree are compare by DP matching, then a parallel ata set is extracte. The rest of this paper is organize as follows. Section 2 an 3 escribe the social meia music app nana an voice conversion using nana s ata, respectively. Section 4 escribes the experimental conitions an experimental results. Section 5 presents concluing remarks an future work. II. SOCIAL MEDIA MUSIC APP NANA The social meia music app nana[8] was evelope by nana music, Inc. as a social music platform. The users can recor an uploa souns such as speaking, singing, an instrumental music to nana with their smartphones. Through the app, users worlwie can communicate with each other through music. As of April 2018, there are six million users in 113 countries. 978-988-14768-5-2 2018 APSIPA 1913 APSIPA-ASC 2018

Proceeings, APSIPA Annual Summit an Conference 2018 Guitar Collab Pf. Bass Ba. Dr. Drums Piano Vocal Fig. 4. A tree structure of an example collaboration relationship. Fig. 1. The recoring process. Fig. 2. Playing the soun of a post. Fig. 3. Showing all the collaborators. Pf. Accompaniment Collab More than 61 million posts have accumulate in the atabase. Furthermore, the number of posts is still increasing. The users uploa souns to nana accoring to the following proceure: 1) Recor souns. 2) A information about recore souns, such as title, artist s name, an explanation. 3) Choose soun effects, such as echo, to arrange souns. Fig. 1 shows the screen of the recoring process. Users can listen to uploae posts, as shown in Fig. 2, an the users get feeback such as Comment an Applause (function equivalent to Like ) from other users. In aition to these general functions, users can collaborate on posts. This characteristic function is calle Collab. The users can post their souns, which are overubbe on another user s post. The function has the following two main effects. First, multiple users can create one soun together. For example, multiple users singing is overubbe to create choruses, an instruments are overubbe to create ban performances Fig. 3 shows the relationship between collaborators on a post. On the screen shown in the figure, you can see all the posts in the Collab series of each post. Secon, users can easily post accompanie singing voices because they can use an accompaniment post of another user. They o not have to prepare an accompaniment soun source by themselves; all they nee to o is sing. Because many users have poste their singing in this way, the atabase hols many songs sung by multiple users. Popular songs are typically sung by tens of thousans of users. The posts using Collab have two types of ata. The first type is mixe soun source ata that all the posts in the Collab series are overubbe by. The users are able to listen to only this type of soun source. The other type is single source ata that consists of just a soun recore when posting. In most cases, each ata represents only one singing voice or one instrumental soun, although some of this type of ata has multiple souns; for example, singing with an instrument, such as guitar or piano, at the same time. All the posts uploae using Collab are relate to the collaborate post. This relationship is represente by a tree structure representing each post as a noe. When a post A Target Post Parallel Data Fig. 5. The posts that are regare as the same song because they have the same root noe post. exists an a post B collaborates with A, the post A becomes the parent noe an the post B becomes the chil noe. Fig. 4 shows an example of the tree structure. The post that collaborate first becomes the root noe an is accompaniment in most cases. In Fig. 4, the guitar post is the root noe. Generally, every song tree is compose of singing voices an instrumental souns relate to one song. In almost all singing voice posts of each tree, the same song is sung because they have been sung with the same accompaniment. Fig. 5 shows an example of singing voice posts regare as the same song in a tree. We focus on this tree structure to extract such singing voice posts sung by many users for many-to-one singing voice conversion. III. VOICE CONVERSION USING SINGING POST DATA Voice conversion (VC) is a metho of converting an input speaker s voice into various types of voices while keeping linguistic information unchange. This is mainly use for speaker conversion. A typical VC framework uses a statistical approach[9], [10]. In statistical VC, parallel ata sets, which consist of pairs of speech ata from source an target speakers uttering the same sentences, are use for training moels. One conventional statistical VC is base on a Gaussian mixture moel (GMM) [11] GMM-base VC represents the relationship between acoustic of a source an of a target speaker using linear combine multiple Gaussian istributions. A new approach base on eep neural networks (DNNs) has been propose[12], [13]. This can convert acoustic at a higher egree of precision than the GMM-base one. VC approaches are also istinguishe by the number of source 1914

Many-to-One VC Training Data Singer B Singer C Converter Singer Z (Statistical Moel) Source Target Fig. 6. Overview of many-to-one singing VC Training Source DTW Neural Network training Target Conversion Conversion by Neural Network Fig. 8. [The training ata extraction metho that we propose. Input Fig. 7. Overview of our VC system Converte an target speakers. In aition, other approaches exist, such as singing VC. Examples also inclue conversion of sexuality, age, etc. We employ DNN-base many-to-one singing VC in our system, which converts an arbitrary singer s voice into a particular singer s voice. In many-to-one singing VC, the input singing of an arbitrary singer (source singer) is converte into the singing of a particular singer (target singer), as shown in Fig. 6. Therefore, the parallel ata set has to consist of multiple source singers voices an one target singer s voice. Fig. 7 shows an overview of our VC system. In the training step, first, the acoustic are extracte from source an target ata. Then, time alignment between these feature sequences are obtaine by ynamic time wrapping (DTW)[14]. Finally, the neural network conversion moel is traine using time-aligne acoustic feature sequences. In the conversion step, acoustic extracte from input ata are converte by the traine moel frame-by-frame. Then, the output singing voice is synthesize using a vocoer from converte. We extracte the singing ata set of many users singing the same song from the nana s atabase an applie it to this VC system because it is suitable for a parallel ata set. Although it is generally ifficult to get intene ata from big ata, such singing voice ata is easily extracte from the atabase using tree structure representing collaboration relationships (Fig. 5). However, all of the extracte ata is not necessarily the same phrases, because users can recor an post arbitrary content. For instance, there are some posts in which a singer is harmonizing with another singer an others sing only the hook of a song. Hence, a metho is neee to remove unsuitable ata an create appropriate parallel ata sets. An approach employing ynamic programming (DP) match- Max istance Fig. 9. Max istance of matching pass from iagonal. Segments (not to be use) Segment (to be use) Fig. 10. Selection of training ata consiering segment length. ing to an extract parallel ata set has been propose on the assumption that any two ata of users singing the same song has a higher similarity than two other ata of users singing ifferent songs[15]. DP matching is a classical elastic matching metho, wiely applie to pattern recognition tasks such as speech recognition[14] an character recognition[16]. It can ynamically match each vector of two vector series that have ifferent lengths, an the result is calle a matching path. Then, the accumulation of Eucliean istances between matche vectors is calculate simultaneously at the en of matching. It inicates similarity between vector series. Therefore, it is possible to compare the similarities between two ranomly selecte post ata in the atabase that have ifferent lengths. In the conventional metho, first, a target post is ecie. Then, all the singing voice posts in the same tree are compare with the target post by DP matching, an the posts that have a small value of the accumulation of istances are extracte as source ata of a parallel ata set. However, in this metho, most singers of the selecte source ata woul be similar to the target singer because the accumulation of istances epens on the similarity between singers voices as well as what song was sung. In many-to-one singing VC, various types of voice ata shoul be use for source training 1915

ata to convert arbitrary singers voices. Our metho uses matching paths instea of the accumulation of istances. In our metho, the ifferences in utterance timing between matche frames is calculate from matching paths, an the pairs of frames that have a smaller value of calculation results than threshol are extracte to be use for training base on the hypothesis that the posts that were sung with the same accompaniment have small ifferences in the utterance timing of each phrase. When satisfying the conitions of this metho, the matching path is close to iagonal, as shown in Fig. 8. We call every part of the matching path extracte with this metho a segment. We expect to use our metho to remove unsuitable ata an create parallel ata sets that consists of various types of voices. In this metho, two parameters have to be set. The first is the maximum value of istance between matching paths an iagonal. We call this parameter the max istance (Fig. 9). Increasing this value increases the amount of ata while egraing the quality of the ata. The secon one is the minimum value of segment length. We call this parameter the min seg-size. Fig. 10 shows an example of selection base on segment length. Increasing this value reuces the amount of ata while improving the quality of the ata. IV. EXPERIMENTS Two experiments were conucte to evaluate the propose metho. A. Experimental conitions In this section, we show common experimental conitions. We use ata from 9 trees of songs A, B,..., an I in nana s atabase. The target post ata was the full-length main meloy sung by one female singer. The source post ata were ranomly selecte from each tree, incluing posts singing backing chorus or partly singing. Singing voice signals were sample at 32 khz, an acoustic were extracte with a 5-ms shift. As an acoustic feature, 0 th through 43 th mel-cepstral coefficients were extracte from the smoothe spectrum analyze by STRAIGHT[17]. The DNN use in this system was traine from mel-cepstral coefficients. The architecture of the DNN was a 3-hienlayer fee-forwar neural network with 1024 units per hien layer. were normalize by the mean being zero an the variance being one. The mel-cepstral istortion between the target an the converte mel-cepstra was use as the objective evaluation measure, which is efine as: Mel-CD = 10 D 2 (c (1) c (2) ln 10 )2, (1) =1 where c (1) an c (2) are the th coefficients of the target an the converte mel-cepstra, respectively. B. Close test for parameter consieration This experiment was carrie out to etermine the two parameters, the max istance (the maximum value of istance between matching path an iagonal) an the min seg-size (the minimum value of segment length). Combinations of the parameters were compare base on the mel-cepstral istortion. From the 9 trees, 8 trees (songs A, B,..., an H) were selecte. Then, 50 source posts an 1 target post were selecte from each tree as the training ata. The total training ata of the source singers is 400 an the target singer is 8. Table I an II show the experimental results. They escribe Mel-CD an the percentage of extracte frames to all frames. In Table I, the best Mel-CD value in every column is bolface, an the best value in the table is unerline. In Table II, the corresponing cells are bol-face an unerline. The value - means that no frames were extracte to train the moel because no matching path satisfie the conition[s?]. Although more frames were use for training in the lowerleft sie of Table II, there are cells that have smaller Mel- CD values near the iagonal in Table I. This is because that there is a trae-off between the quantity an the quality of the extracte ata. The quality of the ata improve in the higher right sie of the table where the values of max istance are smaller an the values of min seg-size are larger. These results inicate that our metho with optimum parameters improves conversion accuracy. C. Open test In this experiment, four moels were traine using the singing ata of a ifferent number of collaboration trees. They TABLE I MEL-CD [DB] Min seg-size (s) 0 0.1 1 3 5 7 0.025 5.925 5.993 - - - - 0.05 5.922 5.921 6.868 - - - 0.1 5.924 5.931 6.063 6.930 - - Max 0.2 5.949 5.947 5.946 6.049 6.188 6.354 istance 0.3 5.979 5.973 5.945 5.959 6.0145 6.061 (s) 0.4 5.991 5.990 5.972 5.951 5.965 5.981 0.5 6.005 5.998 5.991 5.955 5.964 5.961 0.6 6.013 6.010 6.003 5.963 5.964 5.968 0.7 6.023 6.016 6.010 5.981 5.959 5.968 0.8 6.035 6.026 6.014 5.993 5.975 5.963 Without the propose metho: 6.014 B TABLE II THE RATIO OF THE NUMBER OF FRAMES USED FOR TRAINING (%) Min seg-size [s] 0 0.1 1 3 5 7 0.025 10.12 2.67 - - - - 0.05 18.08 16.70 0.01 - - - 0.1 29.91 29.39 2.78 0.01 - - Max 0.2 44.57 44.37 25.58 3.96 0.74 0.16 istance 0.3 53.73 53.61 44.33 19.66 8.42 3.81 (s) 0.4 59.96 59.88 54.98 36.88 22.77 14.41 0.5 64.52 64.45 63.18 48.85 35.94 26.49 0.6 68.00 67.95 66.93 57.65 46.69 38.29 0.7 70.78 70.74 69.91 63.28 55.03 47.42 0.8 73.04 73.01 72.35 67.30 61.09 54.76 The total number of frames: 12,247,701 1916

Mel-CD [B] 6.70 6.60 6.50 6.40 6.30 6.20 6.10 6.00 5.90 1 2 4 8 The Number of Trees Fig. 11. Mel-CD [B] (open test) were compare for open ata using the best parameters of our metho in the previous experiment as follows: Max istance: 0.05 secon Min seg-size: 0.1 secon The four moels were traine using post ata in 1 (song A), 2 (songs A an B), 4 (songs A, B, C, an D), an 8 (songs A, B,..., an H) trees, respectively. Four hunre post ata sung by source singers were ranomly selecte from the tree/trees for all moels as source singer training ata. One post ata was selecte from each tree as a training ata of the target singer. The test ata set was compose of the post ata in the tree of song I, which inclue 18 source singers ata an one target singer s ata. Fig. 11 shows the results. Increasing the number of trees mostly cause a ecrease in mel-cepstral istortion because the iversity of the training ata improve. However, the moel using 4 trees outperforme the moel using 8 trees. It is assume that each tree has ifferent suitable parameters. Therefore, it is possible that parameters suitable for each tree are more largely iffer in the moel using 8trees than the moel using 4 trees. V. CONCLUSIONS Using ata poste to social meia, we propose a metho of extracting training ata that can be use for many-to-one singing voice conversions. For training, we use the pairs of matche frames that have small ifferences in utterance timing. We assume that posts that were sung with the same accompaniment woul have little ifference in utterance timing. Experimental results showe that setting two appropriate parameters (the maximum value of istance between the matching path an the iagonal an minimum value of segment length) while consiering the trae-off between the quantity an the quality of the training ata improve the objective evaluation measure. Increasing the number of trees use for training ata cause us to accurately convert songs that were not use for training. Future works inclue researching proper parameters base on various elements (such as tempo), applying ifferent parameters for each song, an subjective evaluation. VI. ACKNOWLEDGMENT This research was supporte by nana music, Inc. REFERENCES [1] J. Yin, W. Lo, an Z. Wu, From Big Data to Great Services, 2016 IEEE International Congress on Big Data (BigData Congress), pp. 165 172, 2016. [2] B. Li, T. Sainath, A. Narayanan, J. Caroselli, M. Bacchiani, A. Misra, I. Shafran, H. Sak, G. Punak, K. Chin, K-C Sim, R. Weiss, K. Wilson, E. Variani, C. Kim, O. Siohan, M. Wein-traub, E. McDermott, R. Rose, an M. Shannon, Moeling for Google Home, in INTERSPEECH-2017, Aug. 2017, pp. 399 403. [3] P. Bell, M. J. F. Gales, T. Hain, J. Kilgour, P. Lanchantin, X. Liu, A. McParlan, S. Renals, O. Saz, M. Wester, an P. C. Woolan, The MGB Challenge: Evaluating Multi-Genre Broacast meia recognition, IEEE Automatic Speech Recognition an Unerstaning Workshop, 2015. [4] P. Karanasou, M. J. F. Gales, P. Lanchantin, X. Liu, Y. Qian, L. Wang, P. C. Woolan, an C. Zhang, Speaker iarisation an longituinal linking in multi-genre broacast ata, IEEE Automatic Speech Recognition an Unerstaning Workshop, 2015. [5] J. Villalba, A. Ortega, A. Miguel, an L. Lleia, Variational Bayesian PLDA for speaker iarization in the MGB Challenge, IEEE Automatic Speech Recognition an Unerstaning Workshop, 2015. [6] P. C. Woolan, X. Liu, Y. Qian, C. Zhang, M. J. F. Gales, P. Karanasou, P. Lanchantin, an L. Wang, Cambrige University transcription systems for the Multi-Genre Broacast Challenge, IEEE Automatic Speech Recognition an Unerstaning Workshop, 2015. [7] O. Saz, M. Doulaty, S. Deena, R. Milner, R. Ng, M. Hasan, Y. Liu, an T. Hain, The 2015 Sheffiel system for transcription of multi-genre broacast meia, IEEE Automatic Speech Recognition an Unerstaning Workshop, 2015. [8] nana, https://nana-music.com/ (2018) [9] T. Toa, A. W. Black, an K. Tokua, Spectral Conversion Base on Maximum Likelihoo Estimation Consiering Global Variance of Converte Parameter, ICASSP 2005, 2005 [10] T. Toa, A. W. Black, an K. Tokua, Voice Conversion Base on Maximum-Likelihoo Estimation of Spectral Parameter Trajectory, IEEE Transactions on Auio, Speech, an Language Processing, Vol. 15, No. 8, 2007. [11] Y. Stylianou, O. Cappe, an E. Moulines, Continuous Probabilistic Transform for Voice Conversion, Proc. of IEEE Trans. Speech Auio Process., vol. 6, pp. 131 142, 1998. [12] S. Desai, E. V. Raghavenra, B. Yegnanarayana, A. W. Black an K. Prahalla, Voice conversion using artificial neural networks, Proceeings of ICASSP 2009 pp. 3893 3896, 2009. [13] N. Hosaka, K. Hashimoto, K. Oura, Y. Nankaku, an K. Tokua, Voice Conversion Base on Trajectory Moel Training of Neural Networks Consiering Global Variance, Interspeech 2016, 2016. [14] H. Sakoe an S. Chiba, Dynamic programming algorithm optimization for spoken wor recognition, IEEE Transactions on s, Speech, an Signal Processing, vol. 26, No. 1, pp. 43 49, 1978. [15] Y. Hono, K. Sawaa, K. Hashimoto, K. Oura, Y. Nankaku, K. Tokua, D. Kono, an D. Ishikawa Singing voice conversion using post ata in music SNS, Proc. of al Society of Japan Autumn Meeting, 1-8-16, pp. 209 210, 2017 (in Japanese). [16] K. Yoshia an H. Sakoe, Online Hanwritten Character Recognition for a Personal Computer System, IEEE Transactions on Consumer Electronics, vol. CE-28, No. 3, pp. 202 209, 1982. [17] H. Kawahara, I. Masua-Katsuse, an A. Cheveigne, Restructuring speech representations using a pitch-aaptive time-frequency smoothing an an instantaneous-frequency-base F0 extraction: Possible role of a repetitive structure in souns, Speech Communication, Vol. 27, No. 3 4, pp. 187 207, 1999. 1917