Automatic Construction of Synthetic Musical Instruments and Performers

Similar documents
Combining Instrument and Performance Models for High-Quality Music Synthesis

Music Understanding and the Future of Music

A Bootstrap Method for Training an Accurate Audio Segmenter

Analysis, Synthesis, and Perception of Musical Sounds

2. AN INTROSPECTION OF THE MORPHING PROCESS

A prototype system for rule-based expressive modifications of audio recordings

Tempo and Beat Analysis

Music Alignment and Applications. Introduction

Experiments on musical instrument separation using multiplecause

Expressive Singing Synthesis based on Unit Selection for the Singing Synthesis Challenge 2016

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES

SYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS

Topic 10. Multi-pitch Analysis

Music Radar: A Web-based Query by Humming System

Topics in Computer Music Instrument Identification. Ioanna Karydi

A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS

MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES

Query By Humming: Finding Songs in a Polyphonic Database

Recognising Cello Performers Using Timbre Models

Transcription An Historical Overview

Recognising Cello Performers using Timbre Models

GCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam

Automatic Piano Music Transcription

A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES

Modified Spectral Modeling Synthesis Algorithm for Digital Piri

Computer Coordination With Popular Music: A New Research Agenda 1

Week 14 Music Understanding and Classification

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

CONTENT-BASED MELODIC TRANSFORMATIONS OF AUDIO MATERIAL FOR A MUSIC PROCESSING APPLICATION

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

Usage metrics: tools for evaluating science collections

LOUDNESS EFFECT OF THE DIFFERENT TONES ON THE TIMBRE SUBJECTIVE PERCEPTION EXPERIMENT OF ERHU

Polyphonic Audio Matching for Score Following and Intelligent Audio Editors

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon

Gus (Guangyu) Xia , NYU Shanghai, Shanghai, Tel: (412) Webpage:

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

Automatic music transcription

Pitch. The perceptual correlate of frequency: the perceptual dimension along which sounds can be ordered from low to high.

Topic 4. Single Pitch Detection

Structured training for large-vocabulary chord recognition. Brian McFee* & Juan Pablo Bello

An interdisciplinary approach to audio effect classification

Pitch Perception and Grouping. HST.723 Neural Coding and Perception of Sound

PULSE-DEPENDENT ANALYSES OF PERCUSSIVE MUSIC

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

Neural Network for Music Instrument Identi cation

Automatic Laughter Detection

WE ADDRESS the development of a novel computational

TOWARDS IMPROVING ONSET DETECTION ACCURACY IN NON- PERCUSSIVE SOUNDS USING MULTIMODAL FUSION

Automatic Laughter Detection

TECHNIQUES FOR AUTOMATIC MUSIC TRANSCRIPTION. Juan Pablo Bello, Giuliano Monti and Mark Sandler

Automatic Labelling of tabla signals

Measurement of overtone frequencies of a toy piano and perception of its pitch

TERRESTRIAL broadcasting of digital television (DTV)

Multiple instrument tracking based on reconstruction error, pitch continuity and instrument activity

Singing voice synthesis based on deep neural networks

Keywords Separation of sound, percussive instruments, non-percussive instruments, flexible audio source separation toolbox

POLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng

A System for Acoustic Chord Transcription and Key Extraction from Audio Using Hidden Markov models Trained on Synthesized Audio

GUIDELINES FOR THE PREPARATION AND SUBMISSION OF YOUR THESIS OR DISSERTATION

Real-Time Spectrogram (RTS tm )

REAL-TIME PITCH TRAINING SYSTEM FOR VIOLIN LEARNERS

MELODY EXTRACTION BASED ON HARMONIC CODED STRUCTURE

A Composition for Clarinet and Real-Time Signal Processing: Using Max on the IRCAM Signal Processing Workstation

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

hit), and assume that longer incidental sounds (forest noise, water, wind noise) resemble a Gaussian noise distribution.

THE importance of music content analysis for musical

Music 209 Advanced Topics in Computer Music Lecture 4 Time Warping

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

MultiSpec Tutorial: Visualizing Growing Degree Day (GDD) Images. In this tutorial, the MultiSpec image processing software will be used to:

Improving Frame Based Automatic Laughter Detection

NON-LINEAR EFFECTS MODELING FOR POLYPHONIC PIANO TRANSCRIPTION

ADSR AMP. ENVELOPE. Moog Music s Guide To Analog Synthesized Percussion. The First Step COMMON VOLUME ENVELOPES

Lecture 9 Source Separation

White Noise Suppression in the Time Domain Part II

Detecting Musical Key with Supervised Learning

Onset Detection and Music Transcription for the Irish Tin Whistle

Multichannel Satellite Image Resolution Enhancement Using Dual-Tree Complex Wavelet Transform and NLM Filtering

Acoustic Instrument Message Specification

Laboratory Assignment 3. Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB

Automatic Music Transcription: The Use of a. Fourier Transform to Analyze Waveform Data. Jake Shankman. Computer Systems Research TJHSST. Dr.

International Journal of Computer Architecture and Mobility (ISSN ) Volume 1-Issue 7, May 2013

ELEC 484 Project Pitch Synchronous Overlap-Add

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University

Part I Of An Exclusive Interview With The Father Of Digital FM Synthesis. By Tom Darter.

EXPLORING THE USE OF ENF FOR MULTIMEDIA SYNCHRONIZATION

Paulo V. K. Borges. Flat 1, 50A, Cephas Av. London, UK, E1 4AR (+44) PRESENTATION

Predicting Performance of PESQ in Case of Single Frame Losses

RELIABILITY REASON FOR A COMMERCIAL INADVERTENT-INTERCHANGE SETTLEMENT STANDARD.

RF (Wireless) Fundamentals 1- Day Seminar

Speech and Speaker Recognition for the Command of an Industrial Robot

MUSI-6201 Computational Music Analysis

Timbre perception

A Matlab toolbox for. Characterisation Of Recorded Underwater Sound (CHORUS) USER S GUIDE

Transcription:

Ph.D. Thesis Proposal Automatic Construction of Synthetic Musical Instruments and Performers Ning Hu Carnegie Mellon University

Thesis Committee Roger B. Dannenberg, Chair Michael S. Lewicki Richard M. Stern David Wessel (UC Berkeley)

Roadmap Introduction System Structure Main Modules Audio Alignment and Segmentation Instrument Model Performance Model Schedule Conclusion

Introduction Define thesis topic Thesis statement Start with modeling the trumpet for classical music Contributions of the thesis Criteria for success

Title Topic Definition Automatic Construction of Synthetic Musical Instruments and Performers Meaning Framework that builds music synthesis

Title Topic Definition Automatic Construction of Synthetic Musical Instruments and Performers Meaning Framework that builds music synthesis Automatic construction

Title Topic Definition Automatic Construction of Synthetic Musical Instruments and Performers Meaning Framework that builds music synthesis Automatic construction Instrument model

Title Topic Definition Automatic Construction of Synthetic Musical Instruments and Performers Meaning Framework that builds music synthesis Automatic construction Instrument model Performance model

Title Topic Definition Automatic Construction of Synthetic Musical Instruments and Performers Meaning Framework that builds music synthesis High quality High performance Capable of modeling different instruments Capable of modeling different music styles

Title Topic Definition Automatic Construction of Synthetic Musical Instruments and Performers Meaning Automatic construction Use machine learning techniques Learn from performance examples Constructs instrument model and performance model

Title Topic Definition Automatic Construction of Synthetic Musical Instruments and Performers Meaning Instrument model Similar to traditional concept of synthesis Input: control signals Output: synthesized sound samples

Title Topic Definition Automatic Construction of Synthetic Musical Instruments and Performers Meaning Performance model Generating appropriate control signals from music context is crucial Drives instrument model Input: digital score (music notation) Output: control signals

Title Topic Definition Automatic Construction of Synthetic Musical Instruments and Performers Meaning Framework that builds music synthesis Automatic construction Instrument model Performance model

Thesis Statement To create a system framework that can automatically create high-quality musical instrument synthesis by using machine-learning techniques to construct the instrument model and the performance model by learning from the performance examples (acoustic recordings and their corresponding scores).

Start with Modeling Musical Instrument: Trumpet Music Style: Classical Music

Reasons for Modeling the Trumpet (1) Most wind instrument synthesizers do not sound realistic Conflict between: Working mechanisms of wind instruments Driven by continuous energy exerted by player Continuous control drives sound production Basic structure of synthesizers Mostly sampling-based Based on single, isolated notes Do not offer a wide range of control

Reasons for Modeling the Trumpet (2) Previous research By Dannenberg and Derenyi (1998) Similar scheme Produces convincing trumpet sound

Reasons for Modeling Classical Music Characteristics of classical music Purer playing style More faithful to the score Fewer articulation effects Characteristics of non-classical music Significant inharmonic & transient sounds Need to model noise with a residual model

Contributions of the thesis Use machine learning techniques Automatically create high-quality synthesis The problem of control Problems of note-oriented synthesis Problems of physical models This approach simplifies the problem

Criteria for success Minimum requirement Design, implement & test basic framework Being able to synthesize realistic trumpet performance for classical music Automated modeling process Tested on one or two wind instruments. Extra tasks Extend system framework Model different musical instruments Model different music styles Future work

Roadmap Introduction System Structure Main Modules Audio Alignment and Segmentation Instrument Model Performance Model Schedule Conclusion

System Structure Synthesis process Pre-processing training data Training process for instrument model Training process for performance model

Synthesis Process Performance Model Control Signals Instrument Model

Pre-processing training data Segmented Audio Automatic Alignment & Segmentation Segmented Score

Training the Performance Model Segmented Segmented Audio Score Correspond Parameter Extraction Control Signals Performance Model Compare Control Signals Error

Training the Instrument Model Segmented Segmented Audio Score Correspond Parameter Extraction Control Signals Spectral Analysis Spectra Supervised Learning Instrument Model

Roadmap Introduction System Structure Main Modules Audio Alignment and Segmentation Instrument Model Performance Model Schedule Conclusion

Audio Alignment Find recording Polyphonic audio alignment score Extract feature sequences from Acoustic recording Score Find optimal alignment Dynamic programming (DP) or Hidden Markov Model (HMM) Satisfactory results Correspondence

Audio Segmentation (1) Dannenberg, et. al., (1999) early work Define rules & thresholds Use features: power; #peaks/period; #zerocrossings/period Not reliable and accurate enough Precise alignment = reliable segmentation Require higher accuracy Need further modification

Audio Segmentation (2) Kapanci & Pfeffer s (2004) work Segmentation problem Classification problem Hierarchical machine-learning framework Detect soft onset: compare frames separated by increasingly longer distances Classification Modules

Roadmap Introduction System Structure Main Modules Audio Alignment and Segmentation Instrument Model Performance Model Schedule Conclusion

Instrument Model Control Signal Instrument Model Harmonic Model + Residual model

Harmonic Model Table 1 Amplitude Control Spectrum Generation Waveform Construction Phase Interpolation + Frequency Control Table 2 Control signals Spectra Spectra Wavetables Wavetables Sound samples

Control Signals Spectra Spectral interpolation (Dannenberg, et. al., 1998) Memory-based approach (Wessel, et. al., 1998) Neural network (Wessel, et. al., 1998)

Spectral interpolation (Dannenberg, et. al., 1998) Generate Spectral lookup table Record a set of sounds Obtain a spectrogram for each sound Retain specific spectra at thresholds 2D Interpolated Spectral Lookup Actual Spectra Frequency Synthetic Spectra Amplitude

Memory-based approach (Wessel, et. al., 1998) Index spectra in a n dim. space Interpolate among k nearest neighbors Special case Spectral interpolation technique n=2 (frequency & amplitude) k=4 Linear interpolation

Neural network (Wessel, et. al., 1998) A feed-forward neural network with multiples layers Input: frequency, amplitude Output: info of sinusoidal components Back-propagation learning method Advantage: very compact & generalize well

Harmonic Model Table 1 Amplitude Control Spectrum Generation Waveform Construction Phase Interpolation + Frequency Control Table 2 Control signals Spectra Spectra Wavetables Wavetables Sound samples

Residual Model Modeling attacks Use recorded attacks Phase matching For other instruments

Roadmap Introduction System Structure Main Modules Audio Alignment and Segmentation Instrument Model Performance Model Schedule Conclusion

Performance Model Performance Model Control Signal Control signals Amplitude envelope Frequency envelope Error metrics Envelope representation Mapping scheme

Amplitude Envelope A tongued note A slurred note *Figures borrowed from (Dannenberg, et. al., 1998)

Error Metrics Acoustic and Synthetic Envelopes Typical metric: RMS error Recent work by (Horner et. al., 2004) Measure perceptual difference Useful reference for error metrics

Envelope Representation Envelope representation Characteristics of amplitude envelope Specific duration Specific shape Specific properties of each part

Ways to Represent Envelopes Collection of general parameters Candidates: center of mass, global/local maximum/minimum, etc. Manual vs. automatic selection Wavelets Hierarchical decomposing functions Very powerful and popular

Mapping Scheme Non-linear regression Find music context Examples: Neural network actual envelopes Kalman filters X k = AX k-1 + W k-1 Function approximation Pattern clustering Classify envelopes into clusters For each input data point: Use corresponding representative envelope Stretch, scale & interpolate accordingly Considered as a form of case-based reasoning

Roadmap Introduction System Structure Main Modules Audio Alignment and Segmentation Instrument Model Performance Model Schedule Conclusion

Oct 2004 ~ Nov 2004 (past) Thesis Proposal Dec 2004 ~ May 2005 Dec 2004 ~ Jan 2005 Jan 2005 Feb 2005 Mar 2005 ~ Apr 2005 May 2005 Schedule (1) Collect performance examples for initial experiments Get familiar with SNDAN package Read the original code by Dannenberg & Derenyi Propose thesis topic System Development Implement Audio Alignment & Segmentation module Incorporate SNDAN package to training data pre-processing stage Develop and compare instrument models Design and implement performance model System integration and testing

Schedule (2) Jun ~ Aug 2005 System Evaluation & Tuning Jun 2005 Model other instruments Jul 2005 Synthesize other types of music Aug 2005 System and model evaluation and fine tuning Sep ~ Nov 2005 Writing Thesis Sep 2005 ~ Thesis write-up and revisions Nov 2005 Nov 2005 Thesis defense

Conclusion Propose a scheme: automatically construct high-quality instrument synthesis by learning from performance examples Machine learning a crucial role. Future work Modeling different instruments Modeling different music styles Make it work in real-time

Roadmap Introduction System Structure Main Modules Audio Alignment and Segmentation Instrument Model Performance Model Schedule Conclusion

Thank You Acknowledgements Roger B. Dannenberg Istvan Derenyi James Beauchamp (SNDAN) Contact Ning Hu (ninghu@cs.cmu.edu)