Automatic music transcription
|
|
- Herbert Ford
- 5 years ago
- Views:
Transcription
1 Music transcription 1 Music transcription 2 Automatic music transcription Sources: * Klapuri, Introduction to music transcription, * Klapuri, Eronen, Astola: Analysis of the Meter of Acoustic Musical Signals, IEEE TASLP * Klapuri, Multiple fundamental frequency estimation by summing harmonic amplitudes, ISMIR * Ryynänen, Klapuri, Automatic transcription of melody, bass line, and chords in polyphonic music, Computer Music Journal, Introduction to music transcription Contents: Introduction to music transcription Rhythm analysis Multiple-F0 analysis Acoustic and musicological models Vocals separation and lyrics Application to music retrieval Music transcription 3 / klap Music transcription 3 Music transcription Music transcription 4 Excerpt from Song #034 in the RWC popular music database: Figures top-down: 1. time-domain signal 2. spectrogram 3. musical notation 4. piano roll Complete vs. partial transcription complete transcription is sometimes impossible or irrelevant partial transcription: for example melody / bass line / percussions / chords etc. Applications and related areas music retrieval structured audio coding intelligent processing / effects stage lighting, automatic accompaniment etc. equipment computer games music perception Anything missing?
2 Perspectives on music transcription Music transcription 5 Perspectives on music transcription Acoustic and musicological models Music transcription 6 Music transcription is a wide topic It is useful to structure the problem by decomposing it into smaller and more tractable subproblems Speech recognition systems depend on language models e.g. probabilities of different word sequences (N gram models) Musicological information is equally important for transcription e.g. probabilities of tone sequences or combinations instrument models P( ) P( ) Acoustic signal Internal models Analysis Result Music transcription 7 Time structure analysis Music transcription 8 2 Onset detection and meter analysis Onset detection = Detection of the beginnings of sounds in an acoustic signal for example tapping foot to music (beat tracking) may include several time scales detect moments of musical stress in an audio signal and discover underlying periodicities in them Applications beat-synchronous feature extraction temporal framework for audio editing synchronization of audio/audio or audio/video
3 Music transcription 9 Music transcription 10 Measuring degree of change in music Characterizes the temporal regularity of the moments of stress Basic idea is to analyse the periodicity of the change signal Figure: Musical meter is hierarchical structure pulse sensations at different time scales tactus level is the most prominent ( foot tapping rate ) tatum: time quantum (fastest pulse) measure pulse: related to harmonic change rate Moments of change are important for onset detection and meter analysis change in the intensity, pitch or timbre of a sound moments of musical stress (accents) are caused by the beginnings of sound events, sudden changes in loudness or timbre, harmonic changes Perceptual change should be estimated to detect what humans detect and to ignore what humans ignore musically meaningful rhythmic parsing Measuring degree of change in music Music transcription 11 Measuring degree of change in music Music transcription 12 Time-domain signal some data reduction is needed But: the power envelope of a signal is not sufficient Frequency selectivity of hearing: audibility of a change at each critical band is only affected by the spectral components within the same band components within a single critical band may mask each other but this does not happen if the frequency separation is sufficiently large Measure change independently at critical bands, and then combine the results Scheirer: perceived rhythmic content of many music types remains the same if only the power envelopes of a few subbands are preserved and then used to modulate a white noise signal one band is not enough applies to music with strong beat
4 Music transcription 13 Measuring degree of change in music: In practice Measuring degree of change in music Degree of change at each band Music transcription 14 music signal Filterbank... Perceived change at subband d / dt xb ( n) x ( n) b d ln xb ( n) dt Filterbank: Fourier transforms in successive ~ 20ms time frames (50% overlap) in each frame n, measure the power x b (n) within b=1,2,...,36 triangular-response bandpass filters that are uniformly distributed on Mel-frequency scale (50Hz 20kHz) f Mel... Combine results output 2595log10(1 f Hz ) 700 Denote by x b (n) the power at critical band b=1,...,36 as a function of time (frame index) n How to measure the degree of change at subbands? Differential? For humans, the smallest detectable change in intensity, I, is approximately proportional to the intensity I of the signal, the same amount of increase being more prominent in a quiet signal. Audible ratio I / I is approximately constant Thus it is reasonable to normalize the differential of power with power: d / dt xb ( n) x ( n) b d ln dt Figure (piano onset): dashed line: (d/dt) x b (n) solid line: (d/dt) ln[x b (n)] x ( n) b Measuring degree of change in music Degree of change at each band A numerically robust way of calculating the logarithm is the µ-law compression, y b n ln 1 ln 1 x b n Music transcription 15 Measuring degree of change in music Summary music signal Filterbank x b (n)... Perceived change at subband... Music transcription 16 Combine results output v(n) constant determines the degree of compression for x b (n) ( = / x ) Differentiate, and retain only positive changes (HWR(x)=max(x, 0)): y b (n) = HWR{y b (n) y b (n 1)} power envelope -law compress d / dt, rectify Finally: sum across channels to estimate overall change 36 ( ) = '( ) =1
5 Measured change signals Music transcription 17 Degree of change ( accent ) Music transcription 18 v(n) v(n) v(n) signal level adaptation would be needed Accent signals (degree of change) Degree of accent as a function of time As described above Pulse strengths ( saliences ) Music transcription 19 Bank of comb filters Music transcription 20 Use bank of comb filters for periodicity analysis We used a = 0.5 where T is half-time in samples (3s) x(n) 1-a a z -k y(n) Metrical pulse saliences Strengths of different metrical pulses at time n (resonator energies) Use comb filters for period analysis Magnitude response: a = 0.9 k = 7 Impulse response:
6 Bank of comb filters Music transcription 21 Higher-level modeling Music transcription 22 Time-varying energies of each comb filter in the filterbank r(,n), input impulse train r(,n), input white noise (t, )= 1 t = -t +1 2 Ø º (t, ) ø ß Figure: r(,n), 1,2,...,100 for an impulse train (period 24 samples) and for white noise r(,n) can be further normalized to get rid of the trend (details are beyond the scope of this course) Meter tatum, tactus, measure Higher-level modeling Music transcription 23 Demonstrations Music transcription 24 Observed: (normalized) comb filter energies r(,n) Prior probabilities (typical tempo values): log-normal distribution Temporal continuity constraints: p(next tempo / prev tempo)
7 Music transcription 25 Introduction Music transcription 26 3 Polyphonic pitch analysis Pitch information is an essential part of almost all Western music Extracting pitch information from recorded audio is hard spectrogram can be calculated straightforwardly piano-roll... more tricky Multiple F0 estimation = F0 estimation in polyphonic signals music variety of sources, wide pitch range, presence of drums A number of completely different approaches have been proposed in the literature Musical sounds Most Western instruments produce harmonic sounds Figure: trumpet sound (260Hz) in time and frequency domains period in time-domain: 1/F0 period in frequency-domain: F0 Music transcription 27 1/ F0 Music transcription 28 How about just autocorrelation function (ACF)? Autocorrelation function (ACF) based algorithms are among the most frequently used single-pitch estimators Usually the maximum value in ACF is taken as 1/F0 period Short-time ACF r( ) for a discrete time domain signal x(n): r( ) 1 N N n 1 n 0 x( n) x( n ) F0 Signal x(n): (vowel [ae]) Frequency (Hz) ACF:
8 Autocorrelation function Music transcription 29 Autocorrelation function Music transcription 30 Short-time ACF within a time frame of length N : N 1 r( ) xnxn ( ) ( ) n 0 Short-time ACF for real-valued signals can be computed via FFT as K / k 2 r( ) IDFT X k cos X k K K where IDFT is inverse Fourier transform and X(k) is DFT of x(n) (padding zeros so that FFT length is twice the length of x) The latter identity is true only for real-valued (audio) signals k 0 From the frequency-domain interpretation, we see at least three properties of ACF that make it non-robust for the period analysis of polyphonic audio the entire spectrum is used (weighting with values btw -1 and 1) all integer multiples of f s / are given the same (unity) weight squaring the spectrum emphasizes timbral properties (formants etc.) In the following, we propose a method which makes three basic modifications to ACF to enhance its robustness 1. sharper peaks (cf. comb filter); 2. weight harmonics ( 1 g(,m) More reliable method* Music transcription 31 Proposed method Summing harmonic amplitudes Music transcription 32 Starting point is conceptually very simple 1. Input signal is first spectrally flattened ( whitened ) to suppress timbral information 2. The salience (strength) of a F0 candidate is calculated as a weighted sum of the amplitudes of its harmonic partials ( t )= t, =1 ( ) ( t, ) where f,m = mf s / is the frequency of the m:th harmonic partial of a F0 candidate f s / f s is the sampling rate, and function g(,m) defines the weight of partial m of period in the sum Y(f) is the short-time Fourier transform of the whitened time-domain signal The basic idea of harmonic summation is intuitively appealing: pitch perception is closely related to time-domain periodicity of sounds Fourier theorem states that a periodic signal can be represented with spectral components at integer multiples of the inverse of the period Question of an optimal mapping of the Fourier spectrum to pitch spectrum (or, a piano roll) is closely related to these methods here, function g(,m) is learned by brute-force optimization ( 300Hz): * Klapuri, A., Multiple fundamental frequency estimation by summing harmonic amplitudes," 7th International Conference on Music Information Retrieval, Victoria, Canada, Oct M s g, m Y f m 1, m g, m g1 g2 f, m fs / mf / s 1 m
9 Proposed method Spectral whitening Music transcription 33 Proposed method Calculation of the F0 salience function Music transcription 34 One of the big challenges in F0 estimation is to make systems robust for different sound sources A way to achieve this is to try to suppress timbral information prior to the actual F0 estimation Whitening 1. Calculate DFT X(k) of the input signal x(n) 2. Calculate standard deviations b (= sqrt(power)) within subbands in the frequency domain (square and sum frequency bins within bands, then sqrt) 3. Calculate bandwise compression coefficients b = b / b, where = 0.3 is a parameter determining the amount of spectral whitening 4. Whitened spectrum Y(k) is obtained by weighting each subband with its compression coefficent and then recombining the subbands Calculated as M s g, m max Y k m 1 k where the set,m defines a range of frequency bins in the vicinity of the m:th overtone of the F0 candidate f s / : where denotes rounding and denotes spacing between fundamental period candidates ( = 1 or 0.5) Weight function was found by optimisation (, m k t, = / t+dt /2 g, m g g f 1 2, m ( ),, /( t-dt /2) fs / mf / s 300Hz): Proposed method Predominant F0 estimation Music transcription 35 Maximum of the salience function s( ) is a quite robust indicator of one of the correct F0s in a polyphonic audio signal predominant F0 estimation: find one (any) of the correct F0s But the second or third-highest peak is often due to the same sound and located at that is half or twice the position of the highest peak Multiple-F0 estimation accuracy can be improved by an iterative estimation and cancellation scheme where each detected sound is cancelled from the mixture and s( ) is updated accordingly before deciding the next F0 Iterative estimation and cancellation Music transcription 36 Step 1: Residual spectrum Y R (k) is initialized to Y(k). A spectrum of detected sounds, Y D (k), is initialized to zero. Step 2: Fundamental period 0 is estimated using Y R (k) to compute s( ). The maximum of s( ) determines 0 Step 3: Harmonic partials of 0 are located at bins mk / 0 m=1,2,...m. Spectrum of the time-domain window function is translated to those frequencies, weighted by g(,m) and added to Y D (k). Step 4: The residual spectrum is updated as Y R (k) max(0, Y R (k) d Y D (k)) where d = 0.2 is a free parameter. Step 5: Return to Step 2. Y R (k)
10 Iterative estimation and cancellation Music transcription 37 F0 gram : piano-roll with confidence levels Music transcription 38 first,... second,... third,... fourth iteration: Music transcription 39 F0 gram : piano-roll with salience (RWC-P #25) Music transcription 40 F0 gram : piano-roll with salience (RWC-P #95)
11 Remarks Music transcription 41 Music transcription 42 The principle of summing harmonic amplitude is very simple, yet it suffices for predominant-f0 estimation in polyphonic signals, provided that the weight g(,m) are appropriate Iterative detection and cancellation helps to remove harmonics and subharmonics of already detected sounds and to reveal remaining sounds behind the most prominent ones Reasonably accurate for a wide range of instruments and F0s 4 Acoustic and musicological modeling Music transcription 43 Music transcription 44 Why acoustic modeling of notes? Acoustic modeling of notes Frame-wise F0 strengths must be processed to get discrete notes (MIDI, score) pitch quantization, onsets, offsets clean up frame-wise errors 1. Extract frame-wise F0 salience (strength) and its differential (here not doing peak-picking or iterative cancellation) Examples in the following Ryynänen, M. and Klapuri, A., Automatic transcription of melody, bass line, and chords in polyphonic music, Computer Music Journal, 32(3), Fall Ryynänen, Klapuri, WASPAA Use training data (RWC Popular Music database) to learn acoustic models for note events (100 pieces with audio + time-aligned MIDI)
12 Music transcription 45 Music transcription 46 Music transcription system Music transcription system Figure: Acoustic model Musicological model: musical key estimation N-gram models for note sequences Combination of an acoustic model and a musicological model (HMMs) Music transcription 47 Music transcription 48 Transcription examples Case study: Singing transcription Complete polyphonic transcriptionhttp:// Ryynänen, Klapuri, Modeling of note events for singing transcription, SAPA Workshop, Transcription of melody, bass, and chords: acoustic signal Feature extraction pitch voicing, accent, meter Probabilistic models discrete note sequence Estimated pitch track has to be post-processed to get notes
13 Case study: Singing transcription Music transcription 49 Brother can you spare me a dime Pieni tytön tylleröinen
Tempo and Beat Analysis
Advanced Course Computer Science Music Processing Summer Term 2010 Meinard Müller, Peter Grosche Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Tempo and Beat Analysis Musical Properties:
More informationTranscription of the Singing Melody in Polyphonic Music
Transcription of the Singing Melody in Polyphonic Music Matti Ryynänen and Anssi Klapuri Institute of Signal Processing, Tampere University Of Technology P.O.Box 553, FI-33101 Tampere, Finland {matti.ryynanen,
More informationDrum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods
Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods Kazuyoshi Yoshii, Masataka Goto and Hiroshi G. Okuno Department of Intelligence Science and Technology National
More informationTempo and Beat Tracking
Tutorial Automatisierte Methoden der Musikverarbeitung 47. Jahrestagung der Gesellschaft für Informatik Tempo and Beat Tracking Meinard Müller, Christof Weiss, Stefan Balke International Audio Laboratories
More informationTHE importance of music content analysis for musical
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 15, NO. 1, JANUARY 2007 333 Drum Sound Recognition for Polyphonic Audio Signals by Adaptation and Matching of Spectrogram Templates With
More informationTopic 10. Multi-pitch Analysis
Topic 10 Multi-pitch Analysis What is pitch? Common elements of music are pitch, rhythm, dynamics, and the sonic qualities of timbre and texture. An auditory perceptual attribute in terms of which sounds
More informationRobert Alexandru Dobre, Cristian Negrescu
ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Automatic Music Transcription Software Based on Constant Q
More informationOBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES
OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES Vishweshwara Rao and Preeti Rao Digital Audio Processing Lab, Electrical Engineering Department, IIT-Bombay, Powai,
More informationCSC475 Music Information Retrieval
CSC475 Music Information Retrieval Monophonic pitch extraction George Tzanetakis University of Victoria 2014 G. Tzanetakis 1 / 32 Table of Contents I 1 Motivation and Terminology 2 Psychacoustics 3 F0
More informationA MID-LEVEL REPRESENTATION FOR CAPTURING DOMINANT TEMPO AND PULSE INFORMATION IN MUSIC RECORDINGS
th International Society for Music Information Retrieval Conference (ISMIR 9) A MID-LEVEL REPRESENTATION FOR CAPTURING DOMINANT TEMPO AND PULSE INFORMATION IN MUSIC RECORDINGS Peter Grosche and Meinard
More informationPOST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS
POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS Andrew N. Robertson, Mark D. Plumbley Centre for Digital Music
More informationQuery By Humming: Finding Songs in a Polyphonic Database
Query By Humming: Finding Songs in a Polyphonic Database John Duchi Computer Science Department Stanford University jduchi@stanford.edu Benjamin Phipps Computer Science Department Stanford University bphipps@stanford.edu
More informationInstrument Recognition in Polyphonic Mixtures Using Spectral Envelopes
Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu
More informationSINGING PITCH EXTRACTION BY VOICE VIBRATO/TREMOLO ESTIMATION AND INSTRUMENT PARTIAL DELETION
th International Society for Music Information Retrieval Conference (ISMIR ) SINGING PITCH EXTRACTION BY VOICE VIBRATO/TREMOLO ESTIMATION AND INSTRUMENT PARTIAL DELETION Chao-Ling Hsu Jyh-Shing Roger Jang
More informationMUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES
MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES Jun Wu, Yu Kitano, Stanislaw Andrzej Raczynski, Shigeki Miyabe, Takuya Nishimoto, Nobutaka Ono and Shigeki Sagayama The Graduate
More informationMusic Information Retrieval
Music Information Retrieval When Music Meets Computer Science Meinard Müller International Audio Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de Berlin MIR Meetup 20.03.2017 Meinard Müller
More informationDAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes
DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring 2009 Week 6 Class Notes Pitch Perception Introduction Pitch may be described as that attribute of auditory sensation in terms
More informationDigital Signal. Continuous. Continuous. amplitude. amplitude. Discrete-time Signal. Analog Signal. Discrete. Continuous. time. time.
Discrete amplitude Continuous amplitude Continuous amplitude Digital Signal Analog Signal Discrete-time Signal Continuous time Discrete time Digital Signal Discrete time 1 Digital Signal contd. Analog
More informationTOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC
TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu
More informationMUSIC is a ubiquitous and vital part of the lives of billions
1088 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 5, NO. 6, OCTOBER 2011 Signal Processing for Music Analysis Meinard Müller, Member, IEEE, Daniel P. W. Ellis, Senior Member, IEEE, Anssi
More informationBeethoven, Bach, and Billions of Bytes
Lecture Music Processing Beethoven, Bach, and Billions of Bytes New Alliances between Music and Computer Science Meinard Müller International Audio Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de
More informationMeasurement of overtone frequencies of a toy piano and perception of its pitch
Measurement of overtone frequencies of a toy piano and perception of its pitch PACS: 43.75.Mn ABSTRACT Akira Nishimura Department of Media and Cultural Studies, Tokyo University of Information Sciences,
More informationHUMANS have a remarkable ability to recognize objects
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 21, NO. 9, SEPTEMBER 2013 1805 Musical Instrument Recognition in Polyphonic Audio Using Missing Feature Approach Dimitrios Giannoulis,
More informationWeek 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University
Week 14 Query-by-Humming and Music Fingerprinting Roger B. Dannenberg Professor of Computer Science, Art and Music Overview n Melody-Based Retrieval n Audio-Score Alignment n Music Fingerprinting 2 Metadata-based
More informationMusic Complexity Descriptors. Matt Stabile June 6 th, 2008
Music Complexity Descriptors Matt Stabile June 6 th, 2008 Musical Complexity as a Semantic Descriptor Modern digital audio collections need new criteria for categorization and searching. Applicable to:
More informationMusic Radar: A Web-based Query by Humming System
Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,
More informationHUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH
Proc. of the th Int. Conference on Digital Audio Effects (DAFx-), Hamburg, Germany, September -8, HUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH George Tzanetakis, Georg Essl Computer
More informationVoice & Music Pattern Extraction: A Review
Voice & Music Pattern Extraction: A Review 1 Pooja Gautam 1 and B S Kaushik 2 Electronics & Telecommunication Department RCET, Bhilai, Bhilai (C.G.) India pooja0309pari@gmail.com 2 Electrical & Instrumentation
More informationMODELING RHYTHM SIMILARITY FOR ELECTRONIC DANCE MUSIC
MODELING RHYTHM SIMILARITY FOR ELECTRONIC DANCE MUSIC Maria Panteli University of Amsterdam, Amsterdam, Netherlands m.x.panteli@gmail.com Niels Bogaards Elephantcandy, Amsterdam, Netherlands niels@elephantcandy.com
More informationAutomatic Rhythmic Notation from Single Voice Audio Sources
Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung
More informationMusic Source Separation
Music Source Separation Hao-Wei Tseng Electrical and Engineering System University of Michigan Ann Arbor, Michigan Email: blakesen@umich.edu Abstract In popular music, a cover version or cover song, or
More informationInteracting with a Virtual Conductor
Interacting with a Virtual Conductor Pieter Bos, Dennis Reidsma, Zsófia Ruttkay, Anton Nijholt HMI, Dept. of CS, University of Twente, PO Box 217, 7500AE Enschede, The Netherlands anijholt@ewi.utwente.nl
More informationAnalysis, Synthesis, and Perception of Musical Sounds
Analysis, Synthesis, and Perception of Musical Sounds The Sound of Music James W. Beauchamp Editor University of Illinois at Urbana, USA 4y Springer Contents Preface Acknowledgments vii xv 1. Analysis
More informationExperiments on musical instrument separation using multiplecause
Experiments on musical instrument separation using multiplecause models J Klingseisen and M D Plumbley* Department of Electronic Engineering King's College London * - Corresponding Author - mark.plumbley@kcl.ac.uk
More informationMultiple instrument tracking based on reconstruction error, pitch continuity and instrument activity
Multiple instrument tracking based on reconstruction error, pitch continuity and instrument activity Holger Kirchhoff 1, Simon Dixon 1, and Anssi Klapuri 2 1 Centre for Digital Music, Queen Mary University
More informationPitch Perception and Grouping. HST.723 Neural Coding and Perception of Sound
Pitch Perception and Grouping HST.723 Neural Coding and Perception of Sound Pitch Perception. I. Pure Tones The pitch of a pure tone is strongly related to the tone s frequency, although there are small
More informationEE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function
EE391 Special Report (Spring 25) Automatic Chord Recognition Using A Summary Autocorrelation Function Advisor: Professor Julius Smith Kyogu Lee Center for Computer Research in Music and Acoustics (CCRMA)
More informationTopic 4. Single Pitch Detection
Topic 4 Single Pitch Detection What is pitch? A perceptual attribute, so subjective Only defined for (quasi) harmonic sounds Harmonic sounds are periodic, and the period is 1/F0. Can be reliably matched
More informationON FINDING MELODIC LINES IN AUDIO RECORDINGS. Matija Marolt
ON FINDING MELODIC LINES IN AUDIO RECORDINGS Matija Marolt Faculty of Computer and Information Science University of Ljubljana, Slovenia matija.marolt@fri.uni-lj.si ABSTRACT The paper presents our approach
More informationON THE USE OF PERCEPTUAL PROPERTIES FOR MELODY ESTIMATION
Proc. of the 4 th Int. Conference on Digital Audio Effects (DAFx-), Paris, France, September 9-23, 2 Proc. of the 4th International Conference on Digital Audio Effects (DAFx-), Paris, France, September
More informationPULSE-DEPENDENT ANALYSES OF PERCUSSIVE MUSIC
PULSE-DEPENDENT ANALYSES OF PERCUSSIVE MUSIC FABIEN GOUYON, PERFECTO HERRERA, PEDRO CANO IUA-Music Technology Group, Universitat Pompeu Fabra, Barcelona, Spain fgouyon@iua.upf.es, pherrera@iua.upf.es,
More informationA prototype system for rule-based expressive modifications of audio recordings
International Symposium on Performance Science ISBN 0-00-000000-0 / 000-0-00-000000-0 The Author 2007, Published by the AEC All rights reserved A prototype system for rule-based expressive modifications
More informationAutomatic Construction of Synthetic Musical Instruments and Performers
Ph.D. Thesis Proposal Automatic Construction of Synthetic Musical Instruments and Performers Ning Hu Carnegie Mellon University Thesis Committee Roger B. Dannenberg, Chair Michael S. Lewicki Richard M.
More informationComputational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)
Computational Models of Music Similarity 1 Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST) Abstract The perceived similarity of two pieces of music is multi-dimensional,
More informationMusic Segmentation Using Markov Chain Methods
Music Segmentation Using Markov Chain Methods Paul Finkelstein March 8, 2011 Abstract This paper will present just how far the use of Markov Chains has spread in the 21 st century. We will explain some
More informationAcoustic Scene Classification
Acoustic Scene Classification Marc-Christoph Gerasch Seminar Topics in Computer Music - Acoustic Scene Classification 6/24/2015 1 Outline Acoustic Scene Classification - definition History and state of
More informationMUSICAL meter is a hierarchical structure, which consists
50 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 1, JANUARY 2010 Music Tempo Estimation With k-nn Regression Antti J. Eronen and Anssi P. Klapuri, Member, IEEE Abstract An approach
More informationAudio. Meinard Müller. Beethoven, Bach, and Billions of Bytes. International Audio Laboratories Erlangen. International Audio Laboratories Erlangen
Meinard Müller Beethoven, Bach, and Billions of Bytes When Music meets Computer Science Meinard Müller International Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de School of Mathematics University
More informationSemi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis
Semi-automated extraction of expressive performance information from acoustic recordings of piano music Andrew Earis Outline Parameters of expressive piano performance Scientific techniques: Fourier transform
More information2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t
MPEG-7 FOR CONTENT-BASED MUSIC PROCESSING Λ Emilia GÓMEZ, Fabien GOUYON, Perfecto HERRERA and Xavier AMATRIAIN Music Technology Group, Universitat Pompeu Fabra, Barcelona, SPAIN http://www.iua.upf.es/mtg
More informationIntroductions to Music Information Retrieval
Introductions to Music Information Retrieval ECE 272/472 Audio Signal Processing Bochen Li University of Rochester Wish List For music learners/performers While I play the piano, turn the page for me Tell
More informationMusic Representations
Lecture Music Processing Music Representations Meinard Müller International Audio Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de Book: Fundamentals of Music Processing Meinard Müller Fundamentals
More information2. AN INTROSPECTION OF THE MORPHING PROCESS
1. INTRODUCTION Voice morphing means the transition of one speech signal into another. Like image morphing, speech morphing aims to preserve the shared characteristics of the starting and final signals,
More informationHST 725 Music Perception & Cognition Assignment #1 =================================================================
HST.725 Music Perception and Cognition, Spring 2009 Harvard-MIT Division of Health Sciences and Technology Course Director: Dr. Peter Cariani HST 725 Music Perception & Cognition Assignment #1 =================================================================
More informationGCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam
GCT535- Sound Technology for Multimedia Timbre Analysis Graduate School of Culture Technology KAIST Juhan Nam 1 Outlines Timbre Analysis Definition of Timbre Timbre Features Zero-crossing rate Spectral
More informationA CHROMA-BASED SALIENCE FUNCTION FOR MELODY AND BASS LINE ESTIMATION FROM MUSIC AUDIO SIGNALS
A CHROMA-BASED SALIENCE FUNCTION FOR MELODY AND BASS LINE ESTIMATION FROM MUSIC AUDIO SIGNALS Justin Salamon Music Technology Group Universitat Pompeu Fabra, Barcelona, Spain justin.salamon@upf.edu Emilia
More informationLecture 10 Harmonic/Percussive Separation
10420CS 573100 音樂資訊檢索 Music Information Retrieval Lecture 10 Harmonic/Percussive Separation Yi-Hsuan Yang Ph.D. http://www.citi.sinica.edu.tw/pages/yang/ yang@citi.sinica.edu.tw Music & Audio Computing
More informationMusic Representations. Beethoven, Bach, and Billions of Bytes. Music. Research Goals. Piano Roll Representation. Player Piano (1900)
Music Representations Lecture Music Processing Sheet Music (Image) CD / MP3 (Audio) MusicXML (Text) Beethoven, Bach, and Billions of Bytes New Alliances between Music and Computer Science Dance / Motion
More informationTopics in Computer Music Instrument Identification. Ioanna Karydi
Topics in Computer Music Instrument Identification Ioanna Karydi Presentation overview What is instrument identification? Sound attributes & Timbre Human performance The ideal algorithm Selected approaches
More informationEffects of acoustic degradations on cover song recognition
Signal Processing in Acoustics: Paper 68 Effects of acoustic degradations on cover song recognition Julien Osmalskyj (a), Jean-Jacques Embrechts (b) (a) University of Liège, Belgium, josmalsky@ulg.ac.be
More informationAN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY
AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY Eugene Mikyung Kim Department of Music Technology, Korea National University of Arts eugene@u.northwestern.edu ABSTRACT
More informationLaboratory Assignment 3. Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB
Laboratory Assignment 3 Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB PURPOSE In this laboratory assignment, you will use MATLAB to synthesize the audio tones that make up a well-known
More informationAuthor Index. Absolu, Brandt 165. Montecchio, Nicola 187 Mukherjee, Bhaswati 285 Müllensiefen, Daniel 365. Bay, Mert 93
Author Index Absolu, Brandt 165 Bay, Mert 93 Datta, Ashoke Kumar 285 Dey, Nityananda 285 Doraisamy, Shyamala 391 Downie, J. Stephen 93 Ehmann, Andreas F. 93 Esposito, Roberto 143 Gerhard, David 119 Golzari,
More informationAUTOMATIC CHARACTERIZATION OF DIGITAL MUSIC FOR RHYTHMIC AUDITORY STIMULATION
11th International Society for Music Information Retrieval Conference (ISMIR 21) AUTOMATIC CHARACTERIZATION OF DIGITAL MUSIC FOR RHYTHMIC AUDITORY STIMULATION Eric Humphrey Music Engineering Technology
More informationCS229 Project Report Polyphonic Piano Transcription
CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project
More informationUsing the new psychoacoustic tonality analyses Tonality (Hearing Model) 1
02/18 Using the new psychoacoustic tonality analyses 1 As of ArtemiS SUITE 9.2, a very important new fully psychoacoustic approach to the measurement of tonalities is now available., based on the Hearing
More informationInvestigation of Digital Signal Processing of High-speed DACs Signals for Settling Time Testing
Universal Journal of Electrical and Electronic Engineering 4(2): 67-72, 2016 DOI: 10.13189/ujeee.2016.040204 http://www.hrpub.org Investigation of Digital Signal Processing of High-speed DACs Signals for
More informationA Beat Tracking System for Audio Signals
A Beat Tracking System for Audio Signals Simon Dixon Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria. simon@ai.univie.ac.at April 7, 2000 Abstract We present
More informationKrzysztof Rychlicki-Kicior, Bartlomiej Stasiak and Mykhaylo Yatsymirskyy Lodz University of Technology
Krzysztof Rychlicki-Kicior, Bartlomiej Stasiak and Mykhaylo Yatsymirskyy Lodz University of Technology 26.01.2015 Multipitch estimation obtains frequencies of sounds from a polyphonic audio signal Number
More informationFurther Topics in MIR
Tutorial Automatisierte Methoden der Musikverarbeitung 47. Jahrestagung der Gesellschaft für Informatik Further Topics in MIR Meinard Müller, Christof Weiss, Stefan Balke International Audio Laboratories
More informationStudy of White Gaussian Noise with Varying Signal to Noise Ratio in Speech Signal using Wavelet
American International Journal of Research in Science, Technology, Engineering & Mathematics Available online at http://www.iasir.net ISSN (Print): 2328-3491, ISSN (Online): 2328-3580, ISSN (CD-ROM): 2328-3629
More informationLecture 9 Source Separation
10420CS 573100 音樂資訊檢索 Music Information Retrieval Lecture 9 Source Separation Yi-Hsuan Yang Ph.D. http://www.citi.sinica.edu.tw/pages/yang/ yang@citi.sinica.edu.tw Music & Audio Computing Lab, Research
More informationSubjective Similarity of Music: Data Collection for Individuality Analysis
Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp
More informationMusic Representations
Advanced Course Computer Science Music Processing Summer Term 00 Music Representations Meinard Müller Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Music Representations Music Representations
More informationSemi-supervised Musical Instrument Recognition
Semi-supervised Musical Instrument Recognition Master s Thesis Presentation Aleksandr Diment 1 1 Tampere niversity of Technology, Finland Supervisors: Adj.Prof. Tuomas Virtanen, MSc Toni Heittola 17 May
More informationMUSI-6201 Computational Music Analysis
MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)
More informationDrum Source Separation using Percussive Feature Detection and Spectral Modulation
ISSC 25, Dublin, September 1-2 Drum Source Separation using Percussive Feature Detection and Spectral Modulation Dan Barry φ, Derry Fitzgerald^, Eugene Coyle φ and Bob Lawlor* φ Digital Audio Research
More informationPOLYPHONIC TRANSCRIPTION BASED ON TEMPORAL EVOLUTION OF SPECTRAL SIMILARITY OF GAUSSIAN MIXTURE MODELS
17th European Signal Processing Conference (EUSIPCO 29) Glasgow, Scotland, August 24-28, 29 POLYPHOIC TRASCRIPTIO BASED O TEMPORAL EVOLUTIO OF SPECTRAL SIMILARITY OF GAUSSIA MIXTURE MODELS F.J. Cañadas-Quesada,
More informationAPPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC
APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC Vishweshwara Rao, Sachin Pant, Madhumita Bhaskar and Preeti Rao Department of Electrical Engineering, IIT Bombay {vishu, sachinp,
More informationOnset Detection and Music Transcription for the Irish Tin Whistle
ISSC 24, Belfast, June 3 - July 2 Onset Detection and Music Transcription for the Irish Tin Whistle Mikel Gainza φ, Bob Lawlor*, Eugene Coyle φ and Aileen Kelleher φ φ Digital Media Centre Dublin Institute
More informationTempo Estimation and Manipulation
Hanchel Cheng Sevy Harris I. Introduction Tempo Estimation and Manipulation This project was inspired by the idea of a smart conducting baton which could change the sound of audio in real time using gestures,
More informationTiming In Expressive Performance
Timing In Expressive Performance 1 Timing In Expressive Performance Craig A. Hanson Stanford University / CCRMA MUS 151 Final Project Timing In Expressive Performance Timing In Expressive Performance 2
More informationEfficient Vocal Melody Extraction from Polyphonic Music Signals
http://dx.doi.org/1.5755/j1.eee.19.6.4575 ELEKTRONIKA IR ELEKTROTECHNIKA, ISSN 1392-1215, VOL. 19, NO. 6, 213 Efficient Vocal Melody Extraction from Polyphonic Music Signals G. Yao 1,2, Y. Zheng 1,2, L.
More informationRhythm related MIR tasks
Rhythm related MIR tasks Ajay Srinivasamurthy 1, André Holzapfel 1 1 MTG, Universitat Pompeu Fabra, Barcelona, Spain 10 July, 2012 Srinivasamurthy et al. (UPF) MIR tasks 10 July, 2012 1 / 23 1 Rhythm 2
More informationMusic Information Retrieval with Temporal Features and Timbre
Music Information Retrieval with Temporal Features and Timbre Angelina A. Tzacheva and Keith J. Bell University of South Carolina Upstate, Department of Informatics 800 University Way, Spartanburg, SC
More informationCONTENT-BASED MELODIC TRANSFORMATIONS OF AUDIO MATERIAL FOR A MUSIC PROCESSING APPLICATION
CONTENT-BASED MELODIC TRANSFORMATIONS OF AUDIO MATERIAL FOR A MUSIC PROCESSING APPLICATION Emilia Gómez, Gilles Peterschmitt, Xavier Amatriain, Perfecto Herrera Music Technology Group Universitat Pompeu
More informationViolin Timbre Space Features
Violin Timbre Space Features J. A. Charles φ, D. Fitzgerald*, E. Coyle φ φ School of Control Systems and Electrical Engineering, Dublin Institute of Technology, IRELAND E-mail: φ jane.charles@dit.ie Eugene.Coyle@dit.ie
More informationAn Empirical Comparison of Tempo Trackers
An Empirical Comparison of Tempo Trackers Simon Dixon Austrian Research Institute for Artificial Intelligence Schottengasse 3, A-1010 Vienna, Austria simon@oefai.at An Empirical Comparison of Tempo Trackers
More informationMusic Synchronization. Music Synchronization. Music Data. Music Data. General Goals. Music Information Retrieval (MIR)
Advanced Course Computer Science Music Processing Summer Term 2010 Music ata Meinard Müller Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Music Synchronization Music ata Various interpretations
More informationSimple Harmonic Motion: What is a Sound Spectrum?
Simple Harmonic Motion: What is a Sound Spectrum? A sound spectrum displays the different frequencies present in a sound. Most sounds are made up of a complicated mixture of vibrations. (There is an introduction
More informationNOTE-LEVEL MUSIC TRANSCRIPTION BY MAXIMUM LIKELIHOOD SAMPLING
NOTE-LEVEL MUSIC TRANSCRIPTION BY MAXIMUM LIKELIHOOD SAMPLING Zhiyao Duan University of Rochester Dept. Electrical and Computer Engineering zhiyao.duan@rochester.edu David Temperley University of Rochester
More informationMusic Tempo Estimation with k-nn Regression
SUBMITTED TO IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, 2008 1 Music Tempo Estimation with k-nn Regression *Antti Eronen and Anssi Klapuri Abstract An approach for tempo estimation from
More informationCHAPTER 4 SEGMENTATION AND FEATURE EXTRACTION
69 CHAPTER 4 SEGMENTATION AND FEATURE EXTRACTION According to the overall architecture of the system discussed in Chapter 3, we need to carry out pre-processing, segmentation and feature extraction. This
More informationImproving Beat Tracking in the presence of highly predominant vocals using source separation techniques: Preliminary study
Improving Beat Tracking in the presence of highly predominant vocals using source separation techniques: Preliminary study José R. Zapata and Emilia Gómez Music Technology Group Universitat Pompeu Fabra
More informationPOLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING
POLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING Luis Gustavo Martins Telecommunications and Multimedia Unit INESC Porto Porto, Portugal lmartins@inescporto.pt Juan José Burred Communication
More informationWE ADDRESS the development of a novel computational
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 18, NO. 3, MARCH 2010 663 Dynamic Spectral Envelope Modeling for Timbre Analysis of Musical Instrument Sounds Juan José Burred, Member,
More informationSoundprism: An Online System for Score-Informed Source Separation of Music Audio Zhiyao Duan, Student Member, IEEE, and Bryan Pardo, Member, IEEE
IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 5, NO. 6, OCTOBER 2011 1205 Soundprism: An Online System for Score-Informed Source Separation of Music Audio Zhiyao Duan, Student Member, IEEE,
More informationAUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION
AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION Halfdan Rump, Shigeki Miyabe, Emiru Tsunoo, Nobukata Ono, Shigeki Sagama The University of Tokyo, Graduate
More informationTopic 11. Score-Informed Source Separation. (chroma slides adapted from Meinard Mueller)
Topic 11 Score-Informed Source Separation (chroma slides adapted from Meinard Mueller) Why Score-informed Source Separation? Audio source separation is useful Music transcription, remixing, search Non-satisfying
More informationTranscription An Historical Overview
Transcription An Historical Overview By Daniel McEnnis 1/20 Overview of the Overview In the Beginning: early transcription systems Piszczalski, Moorer Note Detection Piszczalski, Foster, Chafe, Katayose,
More information