Experimenting with Musically Motivated Convolutional Neural Networks
|
|
- Damon Cannon
- 5 years ago
- Views:
Transcription
1 Experimenting with Musically Motivated Convolutional Neural Networks Jordi Pons 1, Thomas Lidy 2 and Xavier Serra 1 1 Music Technology Group, Universitat Pompeu Fabra, Barcelona 2 Institute of Software Technology and Interactive Systems, TU Wien January 23, 2017 Download the paper CLICKING HERE! J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 1 / 18
2 Outline Motivation Motivation J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 2 / 18
3 Input representation: why log-mel spectrograms for CNN? Input spectrograms for CNN: interpretable filters in time and frequency! N= seg. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 3 / 18
4 Squared/rectangular filters: inertia from computer vision. Are these efficiently representing the relevant local stationaries in music data? J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 4 / 18
5 Squared/rectangular filters: inertia from computer vision. Are these efficiently representing the relevant local stationaries in music data? J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 5 / 18
6 CNNs modeling music data? Three filter shapes are discussed in the following: 1 Squared/rectangular filters 2 Temporal filters 3 Frequency filters Which musical concepts can these filters model? J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 6 / 18
7 Temporal filters (1-by-n): Setting m = 1, NO frequency features but temporal cues. Filters can learn musical concepts at different time-scales depending how n is set, i.e.: Onsets, attack-sustain-release: n N. BPM and rhythm patterns: n N. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 7 / 18
8 Frequency filters (m-by-1): Setting n = 1, frequency features but NO temporal cues. Filters can learn different aspects depending how m is set, i.e.: Timbre + note: m = M. Similar to NMF!! Timbre: m < M. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 8 / 18
9 Squared/rectangular filters (m-by-n filters) learning time and frequency features at the same time. Filters can learn different aspects depending how m/n are set, i.e.: Bass or kick modeling: m M and n N. Represented by a sub-band for a short-time. Cymbals or snare drums modeling: m = M and n N. Broad in frequency with a fixed decay time. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 9 / 18
10 Architectures J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 10 / 18
11 Architectures J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 11 / 18
12 Architectures Joint architecture: Time-Frequency J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 12 / 18
13 Black-box and time architectures Filter shape Accuracy: mean ± std Architecture (m,n) # param. 10 cross-fold validation Black-box (12,8) ± 3.39 % Time (1,60) ± 4.72 % Frequency (32,1) ± 5.82 % Frequency (36,1) ± 5.38 % Frequency (40,1) ± 5.63 % Time-Frequency (1,60)-(32,1) ± 4.29 % Time-FrequencyInit (1,60)-(32,1) ± 4.44 % 1 Genre classification task with the Ballroom dataset % (Marchand et al.) using time and frequency cues 82.3% (Gouyon et al.) using only time cues 15.9% predicting most probable class 2 Black-box and Time-Frequency architecture achieve inferior results than the state-of-the-art. 3 Time architecture achieve equivalent results as its baseline. 4 Frequency architecture outperforms the random baseline: frequency features are more relevant than expected. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 13 / 18
14 Pitch invariance experiment Filter shape Accuracy: mean ± std Architecture (m,n) # param. 10 cross-fold validation Black-box (12,8) ± 3.39 % Time (1,60) ± 4.72 % Frequency (32,1) ± 5.82 % Frequency (36,1) ± 5.38 % Frequency (40,1) ± 5.63 % Time-Frequency (1,60)-(32,1) ± 4.29 % Time-FrequencyInit (1,60)-(32,1) ± 4.44 % 1 Designing the filters such that they can convolve in frequency (m < M), helps predicting the Ballroom classes. because is pitch invariant? because the network is more expressive? J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 14 / 18
15 Time-Frequency experiment Filter shape Accuracy: mean ± std Architecture (m,n) # param. 10 cross-fold validation Black-box (12,8) ± 3.39 % Time (1,60) ± 4.72 % Frequency (32,1) ± 5.82 % Frequency (36,1) ± 5.38 % Frequency (40,1) ± 5.63 % Time-Frequency (1,60)-(32,1) ± 4.29 % Time-FrequencyInit (1,60)-(32,1) ± Pre-initializing the weights is beneficial. 2 With a much less expressive network, Time-FrequencyInit, we obtain similar accuracy results than Black-box. we propose efficient way of representing music data. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 15 / 18
16 We have discussed how several CNNs filter shapes can model musical aspects. We have proposed some musically motivated deep learning architectures. We have shown that these can achieve competitive results on predicting the Ballroom dataset classes. understand what the architectures are learning efficient way of representing musical concepts J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 16 / 18
17 Thanks! Reproduce our research, the code is available: github.com/jordipons/cbmi2016/ Experimenting with Musically Motivated Convolutional Neural Networks Jordi Pons 1, Thomas Lidy 2 and Xavier Serra 1 1 Music Technology Group, Universitat Pompeu Fabra, Barcelona 2 Institute of Software Technology and Interactive Systems, TU Wien J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 17 / 18
18 Ballroom dataset 698 songs. 30 seconds long 8 music genres: cha-cha-cha, jive, quickstep, rumba, samba, tango, viennese-waltz and slow-waltz. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 18 / 18
Deep learning for music data processing
Deep learning for music data processing A personal (re)view of the state-of-the-art Jordi Pons www.jordipons.me Music Technology Group, DTIC, Universitat Pompeu Fabra, Barcelona. 31st January 2017 Jordi
More informationTimbre Analysis of Music Audio Signals with Convolutional Neural Networks
Timbre Analysis of Music Audio Signals with Convolutional Neural Networks Jordi Pons, Olga Slizovskaia, Rong Gong, Emilia Gómez and Xavier Serra Music Technology Group, Universitat Pompeu Fabra, Barcelona.
More informationIn this project you will learn how to code a live music performance, that you can add to and edit without having to stop the music!
Live DJ Introduction: In this project you will learn how to code a live music performance, that you can add to and edit without having to stop the music! Step 1: Drums Let s start by creating a simple
More informationTOWARDS CHARACTERISATION OF MUSIC VIA RHYTHMIC PATTERNS
TOWARDS CHARACTERISATION OF MUSIC VIA RHYTHMIC PATTERNS Simon Dixon Austrian Research Institute for AI Vienna, Austria Fabien Gouyon Universitat Pompeu Fabra Barcelona, Spain Gerhard Widmer Medical University
More informationBAYESIAN METER TRACKING ON LEARNED SIGNAL REPRESENTATIONS
BAYESIAN METER TRACKING ON LEARNED SIGNAL REPRESENTATIONS Andre Holzapfel, Thomas Grill Austrian Research Institute for Artificial Intelligence (OFAI) andre@rhythmos.org, thomas.grill@ofai.at ABSTRACT
More informationClassification of Dance Music by Periodicity Patterns
Classification of Dance Music by Periodicity Patterns Simon Dixon Austrian Research Institute for AI Freyung 6/6, Vienna 1010, Austria simon@oefai.at Elias Pampalk Austrian Research Institute for AI Freyung
More informationMODAL ANALYSIS AND TRANSCRIPTION OF STROKES OF THE MRIDANGAM USING NON-NEGATIVE MATRIX FACTORIZATION
MODAL ANALYSIS AND TRANSCRIPTION OF STROKES OF THE MRIDANGAM USING NON-NEGATIVE MATRIX FACTORIZATION Akshay Anantapadmanabhan 1, Ashwin Bellur 2 and Hema A Murthy 1 1 Department of Computer Science and
More informationADSR AMP. ENVELOPE. Moog Music s Guide To Analog Synthesized Percussion. The First Step COMMON VOLUME ENVELOPES
Moog Music s Guide To Analog Synthesized Percussion Creating tones for reproducing the family of instruments in which sound arises from the striking of materials with sticks, hammers, or the hands. The
More informationRHYTHMIC PATTERN MODELING FOR BEAT AND DOWNBEAT TRACKING IN MUSICAL AUDIO
RHYTHMIC PATTERN MODELING FOR BEAT AND DOWNBEAT TRACKING IN MUSICAL AUDIO Florian Krebs, Sebastian Böck, and Gerhard Widmer Department of Computational Perception Johannes Kepler University, Linz, Austria
More informationMUSICAL INSTRUMENT RECOGNITION USING BIOLOGICALLY INSPIRED FILTERING OF TEMPORAL DICTIONARY ATOMS
MUSICAL INSTRUMENT RECOGNITION USING BIOLOGICALLY INSPIRED FILTERING OF TEMPORAL DICTIONARY ATOMS Steven K. Tjoa and K. J. Ray Liu Signals and Information Group, Department of Electrical and Computer Engineering
More informationMusicological perspective. Martin Clayton
Musicological perspective Martin Clayton Agenda Introductory presentations (Xavier, Martin, Baris) [30 min.] Musicological perspective (Martin) [30 min.] Corpus-based research (Xavier, Baris) [30 min.]
More informationEVALUATION OF FEATURE EXTRACTORS AND PSYCHO-ACOUSTIC TRANSFORMATIONS FOR MUSIC GENRE CLASSIFICATION
EVALUATION OF FEATURE EXTRACTORS AND PSYCHO-ACOUSTIC TRANSFORMATIONS FOR MUSIC GENRE CLASSIFICATION Thomas Lidy Andreas Rauber Vienna University of Technology Department of Software Technology and Interactive
More informationChord Classification of an Audio Signal using Artificial Neural Network
Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------
More informationMusic BCI ( )
Music BCI (006-2015) Matthias Treder, Benjamin Blankertz Technische Universität Berlin, Berlin, Germany September 5, 2016 1 Introduction We investigated the suitability of musical stimuli for use in a
More informationA Music Information Retrieval Approach Based on Power Laws
A Music Information Retrieval Approach Based on Power Laws Patrick Roos and Bill Manaris Computer Science Department, College of Charleston, 66 George Street, Charleston, SC 29424, USA {patrick.roos, manaris}@cs.cofc.edu
More informationgresearch Focus Cognitive Sciences
Learning about Music Cognition by Asking MIR Questions Sebastian Stober August 12, 2016 CogMIR, New York City sstober@uni-potsdam.de http://www.uni-potsdam.de/mlcog/ MLC g Machine Learning in Cognitive
More informationAUTOMATIC MUSIC TRANSCRIPTION WITH CONVOLUTIONAL NEURAL NETWORKS USING INTUITIVE FILTER SHAPES. A Thesis. presented to
AUTOMATIC MUSIC TRANSCRIPTION WITH CONVOLUTIONAL NEURAL NETWORKS USING INTUITIVE FILTER SHAPES A Thesis presented to the Faculty of California Polytechnic State University, San Luis Obispo In Partial Fulfillment
More informationSinger Traits Identification using Deep Neural Network
Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic
More informationAUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION
AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION Halfdan Rump, Shigeki Miyabe, Emiru Tsunoo, Nobukata Ono, Shigeki Sagama The University of Tokyo, Graduate
More informationMusic Genre Classification
Music Genre Classification chunya25 Fall 2017 1 Introduction A genre is defined as a category of artistic composition, characterized by similarities in form, style, or subject matter. [1] Some researchers
More informationPULSE-DEPENDENT ANALYSES OF PERCUSSIVE MUSIC
PULSE-DEPENDENT ANALYSES OF PERCUSSIVE MUSIC FABIEN GOUYON, PERFECTO HERRERA, PEDRO CANO IUA-Music Technology Group, Universitat Pompeu Fabra, Barcelona, Spain fgouyon@iua.upf.es, pherrera@iua.upf.es,
More informationRepresentations of Sound in Deep Learning of Audio Features from Music
Representations of Sound in Deep Learning of Audio Features from Music Sergey Shuvaev, Hamza Giaffar, and Alexei A. Koulakov Cold Spring Harbor Laboratory, Cold Spring Harbor, NY Abstract The work of a
More informationAutomatic characterization of ornamentation from bassoon recordings for expressive synthesis
Automatic characterization of ornamentation from bassoon recordings for expressive synthesis Montserrat Puiggròs, Emilia Gómez, Rafael Ramírez, Xavier Serra Music technology Group Universitat Pompeu Fabra
More information2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t
MPEG-7 FOR CONTENT-BASED MUSIC PROCESSING Λ Emilia GÓMEZ, Fabien GOUYON, Perfecto HERRERA and Xavier AMATRIAIN Music Technology Group, Universitat Pompeu Fabra, Barcelona, SPAIN http://www.iua.upf.es/mtg
More informationTOWARD UNDERSTANDING EXPRESSIVE PERCUSSION THROUGH CONTENT BASED ANALYSIS
TOWARD UNDERSTANDING EXPRESSIVE PERCUSSION THROUGH CONTENT BASED ANALYSIS Matthew Prockup, Erik M. Schmidt, Jeffrey Scott, and Youngmoo E. Kim Music and Entertainment Technology Laboratory (MET-lab) Electrical
More informationON RHYTHM AND GENERAL MUSIC SIMILARITY
10th International Society for Music Information Retrieval Conference (ISMIR 2009) ON RHYTHM AND GENERAL MUSIC SIMILARITY Tim Pohle 1, Dominik Schnitzer 1,2, Markus Schedl 1, Peter Knees 1 and Gerhard
More informationy POWER USER MUSIC PRODUCTION and PERFORMANCE With the MOTIF ES Mastering the Sample SLICE function
y POWER USER MUSIC PRODUCTION and PERFORMANCE With the MOTIF ES Mastering the Sample SLICE function Phil Clendeninn Senior Product Specialist Technology Products Yamaha Corporation of America Working with
More informationA Survey of Audio-Based Music Classification and Annotation
A Survey of Audio-Based Music Classification and Annotation Zhouyu Fu, Guojun Lu, Kai Ming Ting, and Dengsheng Zhang IEEE Trans. on Multimedia, vol. 13, no. 2, April 2011 presenter: Yin-Tzu Lin ( 阿孜孜 ^.^)
More informationMusic Similarity and Cover Song Identification: The Case of Jazz
Music Similarity and Cover Song Identification: The Case of Jazz Simon Dixon and Peter Foster s.e.dixon@qmul.ac.uk Centre for Digital Music School of Electronic Engineering and Computer Science Queen Mary
More informationAn AI Approach to Automatic Natural Music Transcription
An AI Approach to Automatic Natural Music Transcription Michael Bereket Stanford University Stanford, CA mbereket@stanford.edu Karey Shi Stanford Univeristy Stanford, CA kareyshi@stanford.edu Abstract
More informationMusical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons
Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Róisín Loughran roisin.loughran@ul.ie Jacqueline Walker jacqueline.walker@ul.ie Michael O Neill University
More informationMUSI-6201 Computational Music Analysis
MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)
More informationLecture 9 Source Separation
10420CS 573100 音樂資訊檢索 Music Information Retrieval Lecture 9 Source Separation Yi-Hsuan Yang Ph.D. http://www.citi.sinica.edu.tw/pages/yang/ yang@citi.sinica.edu.tw Music & Audio Computing Lab, Research
More informationStructured training for large-vocabulary chord recognition. Brian McFee* & Juan Pablo Bello
Structured training for large-vocabulary chord recognition Brian McFee* & Juan Pablo Bello Small chord vocabularies Typically a supervised learning problem N C:maj C:min C#:maj C#:min D:maj D:min......
More informationTHE importance of music content analysis for musical
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 15, NO. 1, JANUARY 2007 333 Drum Sound Recognition for Polyphonic Audio Signals by Adaptation and Matching of Spectrogram Templates With
More informationBEAT HISTOGRAM FEATURES FROM NMF-BASED NOVELTY FUNCTIONS FOR MUSIC CLASSIFICATION
BEAT HISTOGRAM FEATURES FROM NMF-BASED NOVELTY FUNCTIONS FOR MUSIC CLASSIFICATION Athanasios Lykartsis Technische Universität Berlin Audio Communication Group alykartsis@mail.tu-berlin.de Chih-Wei Wu Georgia
More informationFrom Low-level to High-level: Comparative Study of Music Similarity Measures
From Low-level to High-level: Comparative Study of Music Similarity Measures Dmitry Bogdanov, Joan Serrà, Nicolas Wack, and Perfecto Herrera Music Technology Group Universitat Pompeu Fabra Roc Boronat,
More informationMODELING MUSICAL RHYTHM AT SCALE WITH THE MUSIC GENOME PROJECT Chestnut St Webster Street Philadelphia, PA Oakland, CA 94612
MODELING MUSICAL RHYTHM AT SCALE WITH THE MUSIC GENOME PROJECT Matthew Prockup +, Andreas F. Ehmann, Fabien Gouyon, Erik M. Schmidt, Youngmoo E. Kim + {mprockup, ykim}@drexel.edu, {fgouyon, aehmann, eschmidt}@pandora.com
More informationComing Soon! New Latin Styles. by Marc Dicciani
Coming Soon! New Latin Styles by Marc Dicciani A brand new book and CD of more than 60 pages containing both traditional and contemporary drumset patterns of select Afro-Cuban and Brazilian styles Featured
More informationLecture 10 Harmonic/Percussive Separation
10420CS 573100 音樂資訊檢索 Music Information Retrieval Lecture 10 Harmonic/Percussive Separation Yi-Hsuan Yang Ph.D. http://www.citi.sinica.edu.tw/pages/yang/ yang@citi.sinica.edu.tw Music & Audio Computing
More informationJOINT BEAT AND DOWNBEAT TRACKING WITH RECURRENT NEURAL NETWORKS
JOINT BEAT AND DOWNBEAT TRACKING WITH RECURRENT NEURAL NETWORKS Sebastian Böck, Florian Krebs, and Gerhard Widmer Department of Computational Perception Johannes Kepler University Linz, Austria sebastian.boeck@jku.at
More informationGCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam
GCT535- Sound Technology for Multimedia Timbre Analysis Graduate School of Culture Technology KAIST Juhan Nam 1 Outlines Timbre Analysis Definition of Timbre Timbre Features Zero-crossing rate Spectral
More informationLesson 1 name: Style Studies: Drum Set
Lesson 1 name: Style Studies: Drum Set Beginner Level 1. Overview: In this lesson, the goal and purpose is to attain an independence skill from hands to feet on a drum set and better understanding of styles
More informationAnalysing Musical Pieces Using harmony-analyser.org Tools
Analysing Musical Pieces Using harmony-analyser.org Tools Ladislav Maršík Dept. of Software Engineering, Faculty of Mathematics and Physics Charles University, Malostranské nám. 25, 118 00 Prague 1, Czech
More informationMusic Information Retrieval
Music Information Retrieval When Music Meets Computer Science Meinard Müller International Audio Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de Berlin MIR Meetup 20.03.2017 Meinard Müller
More informationNeural Network for Music Instrument Identi cation
Neural Network for Music Instrument Identi cation Zhiwen Zhang(MSE), Hanze Tu(CCRMA), Yuan Li(CCRMA) SUN ID: zhiwen, hanze, yuanli92 Abstract - In the context of music, instrument identi cation would contribute
More informationMusic Genre Classification and Variance Comparison on Number of Genres
Music Genre Classification and Variance Comparison on Number of Genres Miguel Francisco, miguelf@stanford.edu Dong Myung Kim, dmk8265@stanford.edu 1 Abstract In this project we apply machine learning techniques
More informationUSING VOICE SUPPRESSION ALGORITHMS TO IMPROVE BEAT TRACKING IN THE PRESENCE OF HIGHLY PREDOMINANT VOCALS. Jose R. Zapata and Emilia Gomez
USING VOICE SUPPRESSION ALGORITHMS TO IMPROVE BEAT TRACKING IN THE PRESENCE OF HIGHLY PREDOMINANT VOCALS Jose R. Zapata and Emilia Gomez Music Technology Group Universitat Pompeu Fabra, Barcelona, Spain
More informationGOOD-SOUNDS.ORG: A FRAMEWORK TO EXPLORE GOODNESS IN INSTRUMENTAL SOUNDS
GOOD-SOUNDS.ORG: A FRAMEWORK TO EXPLORE GOODNESS IN INSTRUMENTAL SOUNDS Giuseppe Bandiera 1 Oriol Romani Picas 1 Hiroshi Tokuda 2 Wataru Hariya 2 Koji Oishi 2 Xavier Serra 1 1 Music Technology Group, Universitat
More informationMusic 209 Advanced Topics in Computer Music Lecture 4 Time Warping
Music 209 Advanced Topics in Computer Music Lecture 4 Time Warping 2006-2-9 Professor David Wessel (with John Lazzaro) (cnmat.berkeley.edu/~wessel, www.cs.berkeley.edu/~lazzaro) www.cs.berkeley.edu/~lazzaro/class/music209
More informationMusic Mood Classification - an SVM based approach. Sebastian Napiorkowski
Music Mood Classification - an SVM based approach Sebastian Napiorkowski Topics on Computer Music (Seminar Report) HPAC - RWTH - SS2015 Contents 1. Motivation 2. Quantification and Definition of Mood 3.
More informationIMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM
IMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM Thomas Lidy, Andreas Rauber Vienna University of Technology, Austria Department of Software
More informationRhythm related MIR tasks
Rhythm related MIR tasks Ajay Srinivasamurthy 1, André Holzapfel 1 1 MTG, Universitat Pompeu Fabra, Barcelona, Spain 10 July, 2012 Srinivasamurthy et al. (UPF) MIR tasks 10 July, 2012 1 / 23 1 Rhythm 2
More informationA COMPARISON OF MELODY EXTRACTION METHODS BASED ON SOURCE-FILTER MODELLING
A COMPARISON OF MELODY EXTRACTION METHODS BASED ON SOURCE-FILTER MODELLING Juan J. Bosch 1 Rachel M. Bittner 2 Justin Salamon 2 Emilia Gómez 1 1 Music Technology Group, Universitat Pompeu Fabra, Spain
More informationA CLASSIFICATION-BASED POLYPHONIC PIANO TRANSCRIPTION APPROACH USING LEARNED FEATURE REPRESENTATIONS
12th International Society for Music Information Retrieval Conference (ISMIR 2011) A CLASSIFICATION-BASED POLYPHONIC PIANO TRANSCRIPTION APPROACH USING LEARNED FEATURE REPRESENTATIONS Juhan Nam Stanford
More informationLSTM Neural Style Transfer in Music Using Computational Musicology
LSTM Neural Style Transfer in Music Using Computational Musicology Jett Oristaglio Dartmouth College, June 4 2017 1. Introduction In the 2016 paper A Neural Algorithm of Artistic Style, Gatys et al. discovered
More informationAutomatic Piano Music Transcription
Automatic Piano Music Transcription Jianyu Fan Qiuhan Wang Xin Li Jianyu.Fan.Gr@dartmouth.edu Qiuhan.Wang.Gr@dartmouth.edu Xi.Li.Gr@dartmouth.edu 1. Introduction Writing down the score while listening
More informationSTRUCTURAL CHANGE ON MULTIPLE TIME SCALES AS A CORRELATE OF MUSICAL COMPLEXITY
STRUCTURAL CHANGE ON MULTIPLE TIME SCALES AS A CORRELATE OF MUSICAL COMPLEXITY Matthias Mauch Mark Levy Last.fm, Karen House, 1 11 Bache s Street, London, N1 6DL. United Kingdom. matthias@last.fm mark@last.fm
More informationSchool District of Marshfield Course Syllabus
School District of Marshfield Course Syllabus Course Name: Soundscapes Length of course: Semester Credits: ½ Credit Course Description: The purpose of Soundscape is understanding the expression of the
More informationLets go through the chart together step by step looking at each bit and understanding what the Chart is asking us to do.
Lesson Twenty Lesson 20 IDS PAS2 Performing a Song- The Buzz Lesson Objectives Developing our ability to play a piece of music. Strengthen our understanding chart reading. Apply many of the skills learned
More informationExpressive Singing Synthesis based on Unit Selection for the Singing Synthesis Challenge 2016
Expressive Singing Synthesis based on Unit Selection for the Singing Synthesis Challenge 2016 Jordi Bonada, Martí Umbert, Merlijn Blaauw Music Technology Group, Universitat Pompeu Fabra, Spain jordi.bonada@upf.edu,
More informationTopic 11. Score-Informed Source Separation. (chroma slides adapted from Meinard Mueller)
Topic 11 Score-Informed Source Separation (chroma slides adapted from Meinard Mueller) Why Score-informed Source Separation? Audio source separation is useful Music transcription, remixing, search Non-satisfying
More informationMusic Representations. Beethoven, Bach, and Billions of Bytes. Music. Research Goals. Piano Roll Representation. Player Piano (1900)
Music Representations Lecture Music Processing Sheet Music (Image) CD / MP3 (Audio) MusicXML (Text) Beethoven, Bach, and Billions of Bytes New Alliances between Music and Computer Science Dance / Motion
More informationComposer Identification of Digital Audio Modeling Content Specific Features Through Markov Models
Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Aric Bartle (abartle@stanford.edu) December 14, 2012 1 Background The field of composer recognition has
More informationPreview Only. A Holiday Encore for Band. Arranged by ROBERT W. SMITH (ASCAP) and MICHAEL STORY (ASCAP)
Band Expressions Series and to All a Good Night A Holiday Encore for Band Arranged by ROBERT W. SMITH (ASCAP) and MICHAEL STORY (ASCAP) 1 Conductor 8 C Flute 2 Oboe 4 1st B% Clarinet 4 2nd B% Clarinet
More informationadvanced spectral processing
advanced spectral processing Jordi Janer Music Technology Group Universitat Pompeu Fabra, Barcelona jordi.janer @ upf.edu CDSIM UPF May 2014 hkp://mtg.upf.edu/ Outline 1. IntroducNon to spectral processing
More informationAllegretto we are the leaders in culture an eisteddfod with a difference
Allegretto Provincial Eisteddfod Syllabus Carpe Diem Allegretto we are the leaders in culture an eisteddfod with a difference The Name and Mission of this Organization: The Eisteddfod is known as Allegretto
More informationCS 7643: Deep Learning
CS 7643: Deep Learning Topics: Stride, padding Pooling layers Fully-connected layers as convolutions Backprop in conv layers Dhruv Batra Georgia Tech Invited Talks Sumit Chopra on CNNs for Pixel Labeling
More informationSub Kick This particular miking trick is one that can be used to bring great low-end presence to the kick drum.
Kick Drum As the heartbeat of the contemporary drum kit, the kick drum sound we ve grown accustomed to hearing is both boomy and round on the bottom and has a nice, bright click in the high mid range.
More informationThe Million Song Dataset
The Million Song Dataset AUDIO FEATURES The Million Song Dataset There is no data like more data Bob Mercer of IBM (1985). T. Bertin-Mahieux, D.P.W. Ellis, B. Whitman, P. Lamere, The Million Song Dataset,
More informationUnifying Low-level and High-level Music. Similarity Measures
Unifying Low-level and High-level Music 1 Similarity Measures Dmitry Bogdanov, Joan Serrà, Nicolas Wack, Perfecto Herrera, and Xavier Serra Abstract Measuring music similarity is essential for multimedia
More informationAudio. Meinard Müller. Beethoven, Bach, and Billions of Bytes. International Audio Laboratories Erlangen. International Audio Laboratories Erlangen
Meinard Müller Beethoven, Bach, and Billions of Bytes When Music meets Computer Science Meinard Müller International Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de School of Mathematics University
More informationAUSTRALIAN PIPE BAND COLLEGE PRELIMINARY DRUMMING SYLLABUS Written by Greg Bassani B.Ed., Dip.T., B.Tech., APBA Principal of Drumming, 2004.
AUSTRALIAN PIPE BAND COLLEGE PRELIMINARY DRUMMING SYLLABUS 2004 Written by Greg Bassani B.Ed., Dip.T., B.Tech., APBA Principal of Drumming, 2004. Version 1.0/2004 G:\APBC DRUMMING SYLLABUS 2004_Preliminary.doc
More informationSTRUCTURAL ANALYSIS AND SEGMENTATION OF MUSIC SIGNALS
STRUCTURAL ANALYSIS AND SEGMENTATION OF MUSIC SIGNALS A DISSERTATION SUBMITTED TO THE DEPARTMENT OF TECHNOLOGY OF THE UNIVERSITAT POMPEU FABRA FOR THE PROGRAM IN COMPUTER SCIENCE AND DIGITAL COMMUNICATION
More informationHST 725 Music Perception & Cognition Assignment #1 =================================================================
HST.725 Music Perception and Cognition, Spring 2009 Harvard-MIT Division of Health Sciences and Technology Course Director: Dr. Peter Cariani HST 725 Music Perception & Cognition Assignment #1 =================================================================
More informationPredicting the immediate future with Recurrent Neural Networks: Pre-training and Applications
Predicting the immediate future with Recurrent Neural Networks: Pre-training and Applications Introduction Brandon Richardson December 16, 2011 Research preformed from the last 5 years has shown that the
More informationComputational Rhythm Similarity Development and Verification Through Deep Networks and Musically Motivated Analysis
NEW YORK UNIVERSITY Computational Rhythm Similarity Development and Verification Through Deep Networks and Musically Motivated Analysis by Tlacael Esparza Submitted in partial fulfillment of the requirements
More informationUsing FR-4x Editor. Introduction. Installation. Connecting the FR-4x to the Editor
Introduction FR-4x Editor is an editor for creating sets (SET) and user programs (UPG). This document explains how to edit the parameters and how to save the set or user program. (This document explains
More informationMusic Understanding and the Future of Music
Music Understanding and the Future of Music Roger B. Dannenberg Professor of Computer Science, Art, and Music Carnegie Mellon University Why Computers and Music? Music in every human society! Computers
More informationLow-Latency Instrument Separation in Polyphonic Audio Using Timbre Models
Low-Latency Instrument Separation in Polyphonic Audio Using Timbre Models Ricard Marxer, Jordi Janer, and Jordi Bonada Universitat Pompeu Fabra, Music Technology Group, Roc Boronat 138, Barcelona {ricard.marxer,jordi.janer,jordi.bonada}@upf.edu
More informationMultidimensional analysis of interdependence in a string quartet
International Symposium on Performance Science The Author 2013 ISBN tbc All rights reserved Multidimensional analysis of interdependence in a string quartet Panos Papiotis 1, Marco Marchini 1, and Esteban
More informationUsing the MPEG-7 Standard for the Description of Musical Content
Using the MPEG-7 Standard for the Description of Musical Content EMILIA GÓMEZ, FABIEN GOUYON, PERFECTO HERRERA, XAVIER AMATRIAIN Music Technology Group, Institut Universitari de l Audiovisual Universitat
More informationDrum Source Separation using Percussive Feature Detection and Spectral Modulation
ISSC 25, Dublin, September 1-2 Drum Source Separation using Percussive Feature Detection and Spectral Modulation Dan Barry φ, Derry Fitzgerald^, Eugene Coyle φ and Bob Lawlor* φ Digital Audio Research
More informationMusic Emotion Recognition. Jaesung Lee. Chung-Ang University
Music Emotion Recognition Jaesung Lee Chung-Ang University Introduction Searching Music in Music Information Retrieval Some information about target music is available Query by Text: Title, Artist, or
More informationTempo and Beat Analysis
Advanced Course Computer Science Music Processing Summer Term 2010 Meinard Müller, Peter Grosche Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Tempo and Beat Analysis Musical Properties:
More informationLEARNING A FEATURE SPACE FOR SIMILARITY IN WORLD MUSIC
LEARNING A FEATURE SPACE FOR SIMILARITY IN WORLD MUSIC Maria Panteli, Emmanouil Benetos, Simon Dixon Centre for Digital Music, Queen Mary University of London, United Kingdom {m.panteli, emmanouil.benetos,
More informationFirst Step Towards Enhancing Word Embeddings with Pitch Accents for DNN-based Slot Filling on Recognized Text
First Step Towards Enhancing Word Embeddings with Pitch Accents for DNN-based Slot Filling on Recognized Text Sabrina Stehwien, Ngoc Thang Vu IMS, University of Stuttgart March 16, 2017 Slot Filling sequential
More informationMelody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng
Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the
More informationLEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception
LEARNING AUDIO SHEET MUSIC CORRESPONDENCES Matthias Dorfer Department of Computational Perception Short Introduction... I am a PhD Candidate in the Department of Computational Perception at Johannes Kepler
More informationLaughbot: Detecting Humor in Spoken Language with Language and Audio Cues
Laughbot: Detecting Humor in Spoken Language with Language and Audio Cues Kate Park katepark@stanford.edu Annie Hu anniehu@stanford.edu Natalie Muenster ncm000@stanford.edu Abstract We propose detecting
More informationCONTENT-BASED MELODIC TRANSFORMATIONS OF AUDIO MATERIAL FOR A MUSIC PROCESSING APPLICATION
CONTENT-BASED MELODIC TRANSFORMATIONS OF AUDIO MATERIAL FOR A MUSIC PROCESSING APPLICATION Emilia Gómez, Gilles Peterschmitt, Xavier Amatriain, Perfecto Herrera Music Technology Group Universitat Pompeu
More informationMuseGAN: Multi-track Sequential Generative Adversarial Networks for Symbolic Music Generation and Accompaniment
MuseGAN: Multi-track Sequential Generative Adversarial Networks for Symbolic Music Generation and Accompaniment Hao-Wen Dong*, Wen-Yi Hsiao*, Li-Chia Yang, Yi-Hsuan Yang Research Center of IT Innovation,
More informationDr Kelly Jakubowski Music Psychologist October 2017
Dr Kelly Jakubowski Music Psychologist October 2017 Overview Musical rhythm: Introduction Rhythm and movement Rhythm and language Rhythm and social engagement Introduction Engaging with music can teach
More informationAutomatic Identification of Samples in Hip Hop Music
Automatic Identification of Samples in Hip Hop Music Jan Van Balen 1, Martín Haro 2, and Joan Serrà 3 1 Dept of Information and Computing Sciences, Utrecht University, the Netherlands 2 Music Technology
More informationChairs: Josep Lladós (CVC, Universitat Autònoma de Barcelona)
Session 3: Optical Music Recognition Chairs: Nina Hirata (University of São Paulo) Josep Lladós (CVC, Universitat Autònoma de Barcelona) Session outline (each paper: 10 min presentation) On the Potential
More informationThis is why when you come close to dance music being played, the first thing that you hear is the boom-boom-boom of the kick drum.
Unit 02 Creating Music Learners must select and create key musical elements and organise them into a complete original musical piece in their chosen style using a DAW. The piece must use a minimum of 4
More informationClassroom. Chapter 2: Lesson 12
Classroom Chapter 2: Lesson 12 Adventus Incorporated, 2001 Chapter 2: The Grand Staff Lesson 12 For specific information on Piano Suite Premier and how to purchase, please visit one of the following links:
More informationMusic Curriculum Kindergarten
Music Curriculum Kindergarten Wisconsin Model Standards for Music A: Singing Echo short melodic patterns appropriate to grade level Sing kindergarten repertoire with appropriate posture and breathing Maintain
More informationMusic Synchronization. Music Synchronization. Music Data. Music Data. General Goals. Music Information Retrieval (MIR)
Advanced Course Computer Science Music Processing Summer Term 2010 Music ata Meinard Müller Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Music Synchronization Music ata Various interpretations
More informationCS 591 S1 Computational Audio
4/29/7 CS 59 S Computational Audio Wayne Snyder Computer Science Department Boston University Today: Comparing Musical Signals: Cross- and Autocorrelations of Spectral Data for Structure Analysis Segmentation
More information