Experimenting with Musically Motivated Convolutional Neural Networks

Size: px
Start display at page:

Download "Experimenting with Musically Motivated Convolutional Neural Networks"

Transcription

1 Experimenting with Musically Motivated Convolutional Neural Networks Jordi Pons 1, Thomas Lidy 2 and Xavier Serra 1 1 Music Technology Group, Universitat Pompeu Fabra, Barcelona 2 Institute of Software Technology and Interactive Systems, TU Wien January 23, 2017 Download the paper CLICKING HERE! J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 1 / 18

2 Outline Motivation Motivation J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 2 / 18

3 Input representation: why log-mel spectrograms for CNN? Input spectrograms for CNN: interpretable filters in time and frequency! N= seg. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 3 / 18

4 Squared/rectangular filters: inertia from computer vision. Are these efficiently representing the relevant local stationaries in music data? J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 4 / 18

5 Squared/rectangular filters: inertia from computer vision. Are these efficiently representing the relevant local stationaries in music data? J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 5 / 18

6 CNNs modeling music data? Three filter shapes are discussed in the following: 1 Squared/rectangular filters 2 Temporal filters 3 Frequency filters Which musical concepts can these filters model? J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 6 / 18

7 Temporal filters (1-by-n): Setting m = 1, NO frequency features but temporal cues. Filters can learn musical concepts at different time-scales depending how n is set, i.e.: Onsets, attack-sustain-release: n N. BPM and rhythm patterns: n N. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 7 / 18

8 Frequency filters (m-by-1): Setting n = 1, frequency features but NO temporal cues. Filters can learn different aspects depending how m is set, i.e.: Timbre + note: m = M. Similar to NMF!! Timbre: m < M. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 8 / 18

9 Squared/rectangular filters (m-by-n filters) learning time and frequency features at the same time. Filters can learn different aspects depending how m/n are set, i.e.: Bass or kick modeling: m M and n N. Represented by a sub-band for a short-time. Cymbals or snare drums modeling: m = M and n N. Broad in frequency with a fixed decay time. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 9 / 18

10 Architectures J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 10 / 18

11 Architectures J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 11 / 18

12 Architectures Joint architecture: Time-Frequency J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 12 / 18

13 Black-box and time architectures Filter shape Accuracy: mean ± std Architecture (m,n) # param. 10 cross-fold validation Black-box (12,8) ± 3.39 % Time (1,60) ± 4.72 % Frequency (32,1) ± 5.82 % Frequency (36,1) ± 5.38 % Frequency (40,1) ± 5.63 % Time-Frequency (1,60)-(32,1) ± 4.29 % Time-FrequencyInit (1,60)-(32,1) ± 4.44 % 1 Genre classification task with the Ballroom dataset % (Marchand et al.) using time and frequency cues 82.3% (Gouyon et al.) using only time cues 15.9% predicting most probable class 2 Black-box and Time-Frequency architecture achieve inferior results than the state-of-the-art. 3 Time architecture achieve equivalent results as its baseline. 4 Frequency architecture outperforms the random baseline: frequency features are more relevant than expected. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 13 / 18

14 Pitch invariance experiment Filter shape Accuracy: mean ± std Architecture (m,n) # param. 10 cross-fold validation Black-box (12,8) ± 3.39 % Time (1,60) ± 4.72 % Frequency (32,1) ± 5.82 % Frequency (36,1) ± 5.38 % Frequency (40,1) ± 5.63 % Time-Frequency (1,60)-(32,1) ± 4.29 % Time-FrequencyInit (1,60)-(32,1) ± 4.44 % 1 Designing the filters such that they can convolve in frequency (m < M), helps predicting the Ballroom classes. because is pitch invariant? because the network is more expressive? J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 14 / 18

15 Time-Frequency experiment Filter shape Accuracy: mean ± std Architecture (m,n) # param. 10 cross-fold validation Black-box (12,8) ± 3.39 % Time (1,60) ± 4.72 % Frequency (32,1) ± 5.82 % Frequency (36,1) ± 5.38 % Frequency (40,1) ± 5.63 % Time-Frequency (1,60)-(32,1) ± 4.29 % Time-FrequencyInit (1,60)-(32,1) ± Pre-initializing the weights is beneficial. 2 With a much less expressive network, Time-FrequencyInit, we obtain similar accuracy results than Black-box. we propose efficient way of representing music data. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 15 / 18

16 We have discussed how several CNNs filter shapes can model musical aspects. We have proposed some musically motivated deep learning architectures. We have shown that these can achieve competitive results on predicting the Ballroom dataset classes. understand what the architectures are learning efficient way of representing musical concepts J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 16 / 18

17 Thanks! Reproduce our research, the code is available: github.com/jordipons/cbmi2016/ Experimenting with Musically Motivated Convolutional Neural Networks Jordi Pons 1, Thomas Lidy 2 and Xavier Serra 1 1 Music Technology Group, Universitat Pompeu Fabra, Barcelona 2 Institute of Software Technology and Interactive Systems, TU Wien J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 17 / 18

18 Ballroom dataset 698 songs. 30 seconds long 8 music genres: cha-cha-cha, jive, quickstep, rumba, samba, tango, viennese-waltz and slow-waltz. J. Pons, T. Lidy and X. Serra January 23, 2017 Experimenting with Musically Motivated CNN 18 / 18

Deep learning for music data processing

Deep learning for music data processing Deep learning for music data processing A personal (re)view of the state-of-the-art Jordi Pons www.jordipons.me Music Technology Group, DTIC, Universitat Pompeu Fabra, Barcelona. 31st January 2017 Jordi

More information

Timbre Analysis of Music Audio Signals with Convolutional Neural Networks

Timbre Analysis of Music Audio Signals with Convolutional Neural Networks Timbre Analysis of Music Audio Signals with Convolutional Neural Networks Jordi Pons, Olga Slizovskaia, Rong Gong, Emilia Gómez and Xavier Serra Music Technology Group, Universitat Pompeu Fabra, Barcelona.

More information

In this project you will learn how to code a live music performance, that you can add to and edit without having to stop the music!

In this project you will learn how to code a live music performance, that you can add to and edit without having to stop the music! Live DJ Introduction: In this project you will learn how to code a live music performance, that you can add to and edit without having to stop the music! Step 1: Drums Let s start by creating a simple

More information

TOWARDS CHARACTERISATION OF MUSIC VIA RHYTHMIC PATTERNS

TOWARDS CHARACTERISATION OF MUSIC VIA RHYTHMIC PATTERNS TOWARDS CHARACTERISATION OF MUSIC VIA RHYTHMIC PATTERNS Simon Dixon Austrian Research Institute for AI Vienna, Austria Fabien Gouyon Universitat Pompeu Fabra Barcelona, Spain Gerhard Widmer Medical University

More information

BAYESIAN METER TRACKING ON LEARNED SIGNAL REPRESENTATIONS

BAYESIAN METER TRACKING ON LEARNED SIGNAL REPRESENTATIONS BAYESIAN METER TRACKING ON LEARNED SIGNAL REPRESENTATIONS Andre Holzapfel, Thomas Grill Austrian Research Institute for Artificial Intelligence (OFAI) andre@rhythmos.org, thomas.grill@ofai.at ABSTRACT

More information

Classification of Dance Music by Periodicity Patterns

Classification of Dance Music by Periodicity Patterns Classification of Dance Music by Periodicity Patterns Simon Dixon Austrian Research Institute for AI Freyung 6/6, Vienna 1010, Austria simon@oefai.at Elias Pampalk Austrian Research Institute for AI Freyung

More information

MODAL ANALYSIS AND TRANSCRIPTION OF STROKES OF THE MRIDANGAM USING NON-NEGATIVE MATRIX FACTORIZATION

MODAL ANALYSIS AND TRANSCRIPTION OF STROKES OF THE MRIDANGAM USING NON-NEGATIVE MATRIX FACTORIZATION MODAL ANALYSIS AND TRANSCRIPTION OF STROKES OF THE MRIDANGAM USING NON-NEGATIVE MATRIX FACTORIZATION Akshay Anantapadmanabhan 1, Ashwin Bellur 2 and Hema A Murthy 1 1 Department of Computer Science and

More information

ADSR AMP. ENVELOPE. Moog Music s Guide To Analog Synthesized Percussion. The First Step COMMON VOLUME ENVELOPES

ADSR AMP. ENVELOPE. Moog Music s Guide To Analog Synthesized Percussion. The First Step COMMON VOLUME ENVELOPES Moog Music s Guide To Analog Synthesized Percussion Creating tones for reproducing the family of instruments in which sound arises from the striking of materials with sticks, hammers, or the hands. The

More information

RHYTHMIC PATTERN MODELING FOR BEAT AND DOWNBEAT TRACKING IN MUSICAL AUDIO

RHYTHMIC PATTERN MODELING FOR BEAT AND DOWNBEAT TRACKING IN MUSICAL AUDIO RHYTHMIC PATTERN MODELING FOR BEAT AND DOWNBEAT TRACKING IN MUSICAL AUDIO Florian Krebs, Sebastian Böck, and Gerhard Widmer Department of Computational Perception Johannes Kepler University, Linz, Austria

More information

MUSICAL INSTRUMENT RECOGNITION USING BIOLOGICALLY INSPIRED FILTERING OF TEMPORAL DICTIONARY ATOMS

MUSICAL INSTRUMENT RECOGNITION USING BIOLOGICALLY INSPIRED FILTERING OF TEMPORAL DICTIONARY ATOMS MUSICAL INSTRUMENT RECOGNITION USING BIOLOGICALLY INSPIRED FILTERING OF TEMPORAL DICTIONARY ATOMS Steven K. Tjoa and K. J. Ray Liu Signals and Information Group, Department of Electrical and Computer Engineering

More information

Musicological perspective. Martin Clayton

Musicological perspective. Martin Clayton Musicological perspective Martin Clayton Agenda Introductory presentations (Xavier, Martin, Baris) [30 min.] Musicological perspective (Martin) [30 min.] Corpus-based research (Xavier, Baris) [30 min.]

More information

EVALUATION OF FEATURE EXTRACTORS AND PSYCHO-ACOUSTIC TRANSFORMATIONS FOR MUSIC GENRE CLASSIFICATION

EVALUATION OF FEATURE EXTRACTORS AND PSYCHO-ACOUSTIC TRANSFORMATIONS FOR MUSIC GENRE CLASSIFICATION EVALUATION OF FEATURE EXTRACTORS AND PSYCHO-ACOUSTIC TRANSFORMATIONS FOR MUSIC GENRE CLASSIFICATION Thomas Lidy Andreas Rauber Vienna University of Technology Department of Software Technology and Interactive

More information

Chord Classification of an Audio Signal using Artificial Neural Network

Chord Classification of an Audio Signal using Artificial Neural Network Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------

More information

Music BCI ( )

Music BCI ( ) Music BCI (006-2015) Matthias Treder, Benjamin Blankertz Technische Universität Berlin, Berlin, Germany September 5, 2016 1 Introduction We investigated the suitability of musical stimuli for use in a

More information

A Music Information Retrieval Approach Based on Power Laws

A Music Information Retrieval Approach Based on Power Laws A Music Information Retrieval Approach Based on Power Laws Patrick Roos and Bill Manaris Computer Science Department, College of Charleston, 66 George Street, Charleston, SC 29424, USA {patrick.roos, manaris}@cs.cofc.edu

More information

gresearch Focus Cognitive Sciences

gresearch Focus Cognitive Sciences Learning about Music Cognition by Asking MIR Questions Sebastian Stober August 12, 2016 CogMIR, New York City sstober@uni-potsdam.de http://www.uni-potsdam.de/mlcog/ MLC g Machine Learning in Cognitive

More information

AUTOMATIC MUSIC TRANSCRIPTION WITH CONVOLUTIONAL NEURAL NETWORKS USING INTUITIVE FILTER SHAPES. A Thesis. presented to

AUTOMATIC MUSIC TRANSCRIPTION WITH CONVOLUTIONAL NEURAL NETWORKS USING INTUITIVE FILTER SHAPES. A Thesis. presented to AUTOMATIC MUSIC TRANSCRIPTION WITH CONVOLUTIONAL NEURAL NETWORKS USING INTUITIVE FILTER SHAPES A Thesis presented to the Faculty of California Polytechnic State University, San Luis Obispo In Partial Fulfillment

More information

Singer Traits Identification using Deep Neural Network

Singer Traits Identification using Deep Neural Network Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic

More information

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION Halfdan Rump, Shigeki Miyabe, Emiru Tsunoo, Nobukata Ono, Shigeki Sagama The University of Tokyo, Graduate

More information

Music Genre Classification

Music Genre Classification Music Genre Classification chunya25 Fall 2017 1 Introduction A genre is defined as a category of artistic composition, characterized by similarities in form, style, or subject matter. [1] Some researchers

More information

PULSE-DEPENDENT ANALYSES OF PERCUSSIVE MUSIC

PULSE-DEPENDENT ANALYSES OF PERCUSSIVE MUSIC PULSE-DEPENDENT ANALYSES OF PERCUSSIVE MUSIC FABIEN GOUYON, PERFECTO HERRERA, PEDRO CANO IUA-Music Technology Group, Universitat Pompeu Fabra, Barcelona, Spain fgouyon@iua.upf.es, pherrera@iua.upf.es,

More information

Representations of Sound in Deep Learning of Audio Features from Music

Representations of Sound in Deep Learning of Audio Features from Music Representations of Sound in Deep Learning of Audio Features from Music Sergey Shuvaev, Hamza Giaffar, and Alexei A. Koulakov Cold Spring Harbor Laboratory, Cold Spring Harbor, NY Abstract The work of a

More information

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis

Automatic characterization of ornamentation from bassoon recordings for expressive synthesis Automatic characterization of ornamentation from bassoon recordings for expressive synthesis Montserrat Puiggròs, Emilia Gómez, Rafael Ramírez, Xavier Serra Music technology Group Universitat Pompeu Fabra

More information

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t MPEG-7 FOR CONTENT-BASED MUSIC PROCESSING Λ Emilia GÓMEZ, Fabien GOUYON, Perfecto HERRERA and Xavier AMATRIAIN Music Technology Group, Universitat Pompeu Fabra, Barcelona, SPAIN http://www.iua.upf.es/mtg

More information

TOWARD UNDERSTANDING EXPRESSIVE PERCUSSION THROUGH CONTENT BASED ANALYSIS

TOWARD UNDERSTANDING EXPRESSIVE PERCUSSION THROUGH CONTENT BASED ANALYSIS TOWARD UNDERSTANDING EXPRESSIVE PERCUSSION THROUGH CONTENT BASED ANALYSIS Matthew Prockup, Erik M. Schmidt, Jeffrey Scott, and Youngmoo E. Kim Music and Entertainment Technology Laboratory (MET-lab) Electrical

More information

ON RHYTHM AND GENERAL MUSIC SIMILARITY

ON RHYTHM AND GENERAL MUSIC SIMILARITY 10th International Society for Music Information Retrieval Conference (ISMIR 2009) ON RHYTHM AND GENERAL MUSIC SIMILARITY Tim Pohle 1, Dominik Schnitzer 1,2, Markus Schedl 1, Peter Knees 1 and Gerhard

More information

y POWER USER MUSIC PRODUCTION and PERFORMANCE With the MOTIF ES Mastering the Sample SLICE function

y POWER USER MUSIC PRODUCTION and PERFORMANCE With the MOTIF ES Mastering the Sample SLICE function y POWER USER MUSIC PRODUCTION and PERFORMANCE With the MOTIF ES Mastering the Sample SLICE function Phil Clendeninn Senior Product Specialist Technology Products Yamaha Corporation of America Working with

More information

A Survey of Audio-Based Music Classification and Annotation

A Survey of Audio-Based Music Classification and Annotation A Survey of Audio-Based Music Classification and Annotation Zhouyu Fu, Guojun Lu, Kai Ming Ting, and Dengsheng Zhang IEEE Trans. on Multimedia, vol. 13, no. 2, April 2011 presenter: Yin-Tzu Lin ( 阿孜孜 ^.^)

More information

Music Similarity and Cover Song Identification: The Case of Jazz

Music Similarity and Cover Song Identification: The Case of Jazz Music Similarity and Cover Song Identification: The Case of Jazz Simon Dixon and Peter Foster s.e.dixon@qmul.ac.uk Centre for Digital Music School of Electronic Engineering and Computer Science Queen Mary

More information

An AI Approach to Automatic Natural Music Transcription

An AI Approach to Automatic Natural Music Transcription An AI Approach to Automatic Natural Music Transcription Michael Bereket Stanford University Stanford, CA mbereket@stanford.edu Karey Shi Stanford Univeristy Stanford, CA kareyshi@stanford.edu Abstract

More information

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons Róisín Loughran roisin.loughran@ul.ie Jacqueline Walker jacqueline.walker@ul.ie Michael O Neill University

More information

MUSI-6201 Computational Music Analysis

MUSI-6201 Computational Music Analysis MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)

More information

Lecture 9 Source Separation

Lecture 9 Source Separation 10420CS 573100 音樂資訊檢索 Music Information Retrieval Lecture 9 Source Separation Yi-Hsuan Yang Ph.D. http://www.citi.sinica.edu.tw/pages/yang/ yang@citi.sinica.edu.tw Music & Audio Computing Lab, Research

More information

Structured training for large-vocabulary chord recognition. Brian McFee* & Juan Pablo Bello

Structured training for large-vocabulary chord recognition. Brian McFee* & Juan Pablo Bello Structured training for large-vocabulary chord recognition Brian McFee* & Juan Pablo Bello Small chord vocabularies Typically a supervised learning problem N C:maj C:min C#:maj C#:min D:maj D:min......

More information

THE importance of music content analysis for musical

THE importance of music content analysis for musical IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 15, NO. 1, JANUARY 2007 333 Drum Sound Recognition for Polyphonic Audio Signals by Adaptation and Matching of Spectrogram Templates With

More information

BEAT HISTOGRAM FEATURES FROM NMF-BASED NOVELTY FUNCTIONS FOR MUSIC CLASSIFICATION

BEAT HISTOGRAM FEATURES FROM NMF-BASED NOVELTY FUNCTIONS FOR MUSIC CLASSIFICATION BEAT HISTOGRAM FEATURES FROM NMF-BASED NOVELTY FUNCTIONS FOR MUSIC CLASSIFICATION Athanasios Lykartsis Technische Universität Berlin Audio Communication Group alykartsis@mail.tu-berlin.de Chih-Wei Wu Georgia

More information

From Low-level to High-level: Comparative Study of Music Similarity Measures

From Low-level to High-level: Comparative Study of Music Similarity Measures From Low-level to High-level: Comparative Study of Music Similarity Measures Dmitry Bogdanov, Joan Serrà, Nicolas Wack, and Perfecto Herrera Music Technology Group Universitat Pompeu Fabra Roc Boronat,

More information

MODELING MUSICAL RHYTHM AT SCALE WITH THE MUSIC GENOME PROJECT Chestnut St Webster Street Philadelphia, PA Oakland, CA 94612

MODELING MUSICAL RHYTHM AT SCALE WITH THE MUSIC GENOME PROJECT Chestnut St Webster Street Philadelphia, PA Oakland, CA 94612 MODELING MUSICAL RHYTHM AT SCALE WITH THE MUSIC GENOME PROJECT Matthew Prockup +, Andreas F. Ehmann, Fabien Gouyon, Erik M. Schmidt, Youngmoo E. Kim + {mprockup, ykim}@drexel.edu, {fgouyon, aehmann, eschmidt}@pandora.com

More information

Coming Soon! New Latin Styles. by Marc Dicciani

Coming Soon! New Latin Styles. by Marc Dicciani Coming Soon! New Latin Styles by Marc Dicciani A brand new book and CD of more than 60 pages containing both traditional and contemporary drumset patterns of select Afro-Cuban and Brazilian styles Featured

More information

Lecture 10 Harmonic/Percussive Separation

Lecture 10 Harmonic/Percussive Separation 10420CS 573100 音樂資訊檢索 Music Information Retrieval Lecture 10 Harmonic/Percussive Separation Yi-Hsuan Yang Ph.D. http://www.citi.sinica.edu.tw/pages/yang/ yang@citi.sinica.edu.tw Music & Audio Computing

More information

JOINT BEAT AND DOWNBEAT TRACKING WITH RECURRENT NEURAL NETWORKS

JOINT BEAT AND DOWNBEAT TRACKING WITH RECURRENT NEURAL NETWORKS JOINT BEAT AND DOWNBEAT TRACKING WITH RECURRENT NEURAL NETWORKS Sebastian Böck, Florian Krebs, and Gerhard Widmer Department of Computational Perception Johannes Kepler University Linz, Austria sebastian.boeck@jku.at

More information

GCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam

GCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam GCT535- Sound Technology for Multimedia Timbre Analysis Graduate School of Culture Technology KAIST Juhan Nam 1 Outlines Timbre Analysis Definition of Timbre Timbre Features Zero-crossing rate Spectral

More information

Lesson 1 name: Style Studies: Drum Set

Lesson 1 name: Style Studies: Drum Set Lesson 1 name: Style Studies: Drum Set Beginner Level 1. Overview: In this lesson, the goal and purpose is to attain an independence skill from hands to feet on a drum set and better understanding of styles

More information

Analysing Musical Pieces Using harmony-analyser.org Tools

Analysing Musical Pieces Using harmony-analyser.org Tools Analysing Musical Pieces Using harmony-analyser.org Tools Ladislav Maršík Dept. of Software Engineering, Faculty of Mathematics and Physics Charles University, Malostranské nám. 25, 118 00 Prague 1, Czech

More information

Music Information Retrieval

Music Information Retrieval Music Information Retrieval When Music Meets Computer Science Meinard Müller International Audio Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de Berlin MIR Meetup 20.03.2017 Meinard Müller

More information

Neural Network for Music Instrument Identi cation

Neural Network for Music Instrument Identi cation Neural Network for Music Instrument Identi cation Zhiwen Zhang(MSE), Hanze Tu(CCRMA), Yuan Li(CCRMA) SUN ID: zhiwen, hanze, yuanli92 Abstract - In the context of music, instrument identi cation would contribute

More information

Music Genre Classification and Variance Comparison on Number of Genres

Music Genre Classification and Variance Comparison on Number of Genres Music Genre Classification and Variance Comparison on Number of Genres Miguel Francisco, miguelf@stanford.edu Dong Myung Kim, dmk8265@stanford.edu 1 Abstract In this project we apply machine learning techniques

More information

USING VOICE SUPPRESSION ALGORITHMS TO IMPROVE BEAT TRACKING IN THE PRESENCE OF HIGHLY PREDOMINANT VOCALS. Jose R. Zapata and Emilia Gomez

USING VOICE SUPPRESSION ALGORITHMS TO IMPROVE BEAT TRACKING IN THE PRESENCE OF HIGHLY PREDOMINANT VOCALS. Jose R. Zapata and Emilia Gomez USING VOICE SUPPRESSION ALGORITHMS TO IMPROVE BEAT TRACKING IN THE PRESENCE OF HIGHLY PREDOMINANT VOCALS Jose R. Zapata and Emilia Gomez Music Technology Group Universitat Pompeu Fabra, Barcelona, Spain

More information

GOOD-SOUNDS.ORG: A FRAMEWORK TO EXPLORE GOODNESS IN INSTRUMENTAL SOUNDS

GOOD-SOUNDS.ORG: A FRAMEWORK TO EXPLORE GOODNESS IN INSTRUMENTAL SOUNDS GOOD-SOUNDS.ORG: A FRAMEWORK TO EXPLORE GOODNESS IN INSTRUMENTAL SOUNDS Giuseppe Bandiera 1 Oriol Romani Picas 1 Hiroshi Tokuda 2 Wataru Hariya 2 Koji Oishi 2 Xavier Serra 1 1 Music Technology Group, Universitat

More information

Music 209 Advanced Topics in Computer Music Lecture 4 Time Warping

Music 209 Advanced Topics in Computer Music Lecture 4 Time Warping Music 209 Advanced Topics in Computer Music Lecture 4 Time Warping 2006-2-9 Professor David Wessel (with John Lazzaro) (cnmat.berkeley.edu/~wessel, www.cs.berkeley.edu/~lazzaro) www.cs.berkeley.edu/~lazzaro/class/music209

More information

Music Mood Classification - an SVM based approach. Sebastian Napiorkowski

Music Mood Classification - an SVM based approach. Sebastian Napiorkowski Music Mood Classification - an SVM based approach Sebastian Napiorkowski Topics on Computer Music (Seminar Report) HPAC - RWTH - SS2015 Contents 1. Motivation 2. Quantification and Definition of Mood 3.

More information

IMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM

IMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM IMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM Thomas Lidy, Andreas Rauber Vienna University of Technology, Austria Department of Software

More information

Rhythm related MIR tasks

Rhythm related MIR tasks Rhythm related MIR tasks Ajay Srinivasamurthy 1, André Holzapfel 1 1 MTG, Universitat Pompeu Fabra, Barcelona, Spain 10 July, 2012 Srinivasamurthy et al. (UPF) MIR tasks 10 July, 2012 1 / 23 1 Rhythm 2

More information

A COMPARISON OF MELODY EXTRACTION METHODS BASED ON SOURCE-FILTER MODELLING

A COMPARISON OF MELODY EXTRACTION METHODS BASED ON SOURCE-FILTER MODELLING A COMPARISON OF MELODY EXTRACTION METHODS BASED ON SOURCE-FILTER MODELLING Juan J. Bosch 1 Rachel M. Bittner 2 Justin Salamon 2 Emilia Gómez 1 1 Music Technology Group, Universitat Pompeu Fabra, Spain

More information

A CLASSIFICATION-BASED POLYPHONIC PIANO TRANSCRIPTION APPROACH USING LEARNED FEATURE REPRESENTATIONS

A CLASSIFICATION-BASED POLYPHONIC PIANO TRANSCRIPTION APPROACH USING LEARNED FEATURE REPRESENTATIONS 12th International Society for Music Information Retrieval Conference (ISMIR 2011) A CLASSIFICATION-BASED POLYPHONIC PIANO TRANSCRIPTION APPROACH USING LEARNED FEATURE REPRESENTATIONS Juhan Nam Stanford

More information

LSTM Neural Style Transfer in Music Using Computational Musicology

LSTM Neural Style Transfer in Music Using Computational Musicology LSTM Neural Style Transfer in Music Using Computational Musicology Jett Oristaglio Dartmouth College, June 4 2017 1. Introduction In the 2016 paper A Neural Algorithm of Artistic Style, Gatys et al. discovered

More information

Automatic Piano Music Transcription

Automatic Piano Music Transcription Automatic Piano Music Transcription Jianyu Fan Qiuhan Wang Xin Li Jianyu.Fan.Gr@dartmouth.edu Qiuhan.Wang.Gr@dartmouth.edu Xi.Li.Gr@dartmouth.edu 1. Introduction Writing down the score while listening

More information

STRUCTURAL CHANGE ON MULTIPLE TIME SCALES AS A CORRELATE OF MUSICAL COMPLEXITY

STRUCTURAL CHANGE ON MULTIPLE TIME SCALES AS A CORRELATE OF MUSICAL COMPLEXITY STRUCTURAL CHANGE ON MULTIPLE TIME SCALES AS A CORRELATE OF MUSICAL COMPLEXITY Matthias Mauch Mark Levy Last.fm, Karen House, 1 11 Bache s Street, London, N1 6DL. United Kingdom. matthias@last.fm mark@last.fm

More information

School District of Marshfield Course Syllabus

School District of Marshfield Course Syllabus School District of Marshfield Course Syllabus Course Name: Soundscapes Length of course: Semester Credits: ½ Credit Course Description: The purpose of Soundscape is understanding the expression of the

More information

Lets go through the chart together step by step looking at each bit and understanding what the Chart is asking us to do.

Lets go through the chart together step by step looking at each bit and understanding what the Chart is asking us to do. Lesson Twenty Lesson 20 IDS PAS2 Performing a Song- The Buzz Lesson Objectives Developing our ability to play a piece of music. Strengthen our understanding chart reading. Apply many of the skills learned

More information

Expressive Singing Synthesis based on Unit Selection for the Singing Synthesis Challenge 2016

Expressive Singing Synthesis based on Unit Selection for the Singing Synthesis Challenge 2016 Expressive Singing Synthesis based on Unit Selection for the Singing Synthesis Challenge 2016 Jordi Bonada, Martí Umbert, Merlijn Blaauw Music Technology Group, Universitat Pompeu Fabra, Spain jordi.bonada@upf.edu,

More information

Topic 11. Score-Informed Source Separation. (chroma slides adapted from Meinard Mueller)

Topic 11. Score-Informed Source Separation. (chroma slides adapted from Meinard Mueller) Topic 11 Score-Informed Source Separation (chroma slides adapted from Meinard Mueller) Why Score-informed Source Separation? Audio source separation is useful Music transcription, remixing, search Non-satisfying

More information

Music Representations. Beethoven, Bach, and Billions of Bytes. Music. Research Goals. Piano Roll Representation. Player Piano (1900)

Music Representations. Beethoven, Bach, and Billions of Bytes. Music. Research Goals. Piano Roll Representation. Player Piano (1900) Music Representations Lecture Music Processing Sheet Music (Image) CD / MP3 (Audio) MusicXML (Text) Beethoven, Bach, and Billions of Bytes New Alliances between Music and Computer Science Dance / Motion

More information

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Aric Bartle (abartle@stanford.edu) December 14, 2012 1 Background The field of composer recognition has

More information

Preview Only. A Holiday Encore for Band. Arranged by ROBERT W. SMITH (ASCAP) and MICHAEL STORY (ASCAP)

Preview Only. A Holiday Encore for Band. Arranged by ROBERT W. SMITH (ASCAP) and MICHAEL STORY (ASCAP) Band Expressions Series and to All a Good Night A Holiday Encore for Band Arranged by ROBERT W. SMITH (ASCAP) and MICHAEL STORY (ASCAP) 1 Conductor 8 C Flute 2 Oboe 4 1st B% Clarinet 4 2nd B% Clarinet

More information

advanced spectral processing

advanced spectral processing advanced spectral processing Jordi Janer Music Technology Group Universitat Pompeu Fabra, Barcelona jordi.janer @ upf.edu CDSIM UPF May 2014 hkp://mtg.upf.edu/ Outline 1. IntroducNon to spectral processing

More information

Allegretto we are the leaders in culture an eisteddfod with a difference

Allegretto we are the leaders in culture an eisteddfod with a difference Allegretto Provincial Eisteddfod Syllabus Carpe Diem Allegretto we are the leaders in culture an eisteddfod with a difference The Name and Mission of this Organization: The Eisteddfod is known as Allegretto

More information

CS 7643: Deep Learning

CS 7643: Deep Learning CS 7643: Deep Learning Topics: Stride, padding Pooling layers Fully-connected layers as convolutions Backprop in conv layers Dhruv Batra Georgia Tech Invited Talks Sumit Chopra on CNNs for Pixel Labeling

More information

Sub Kick This particular miking trick is one that can be used to bring great low-end presence to the kick drum.

Sub Kick This particular miking trick is one that can be used to bring great low-end presence to the kick drum. Kick Drum As the heartbeat of the contemporary drum kit, the kick drum sound we ve grown accustomed to hearing is both boomy and round on the bottom and has a nice, bright click in the high mid range.

More information

The Million Song Dataset

The Million Song Dataset The Million Song Dataset AUDIO FEATURES The Million Song Dataset There is no data like more data Bob Mercer of IBM (1985). T. Bertin-Mahieux, D.P.W. Ellis, B. Whitman, P. Lamere, The Million Song Dataset,

More information

Unifying Low-level and High-level Music. Similarity Measures

Unifying Low-level and High-level Music. Similarity Measures Unifying Low-level and High-level Music 1 Similarity Measures Dmitry Bogdanov, Joan Serrà, Nicolas Wack, Perfecto Herrera, and Xavier Serra Abstract Measuring music similarity is essential for multimedia

More information

Audio. Meinard Müller. Beethoven, Bach, and Billions of Bytes. International Audio Laboratories Erlangen. International Audio Laboratories Erlangen

Audio. Meinard Müller. Beethoven, Bach, and Billions of Bytes. International Audio Laboratories Erlangen. International Audio Laboratories Erlangen Meinard Müller Beethoven, Bach, and Billions of Bytes When Music meets Computer Science Meinard Müller International Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de School of Mathematics University

More information

AUSTRALIAN PIPE BAND COLLEGE PRELIMINARY DRUMMING SYLLABUS Written by Greg Bassani B.Ed., Dip.T., B.Tech., APBA Principal of Drumming, 2004.

AUSTRALIAN PIPE BAND COLLEGE PRELIMINARY DRUMMING SYLLABUS Written by Greg Bassani B.Ed., Dip.T., B.Tech., APBA Principal of Drumming, 2004. AUSTRALIAN PIPE BAND COLLEGE PRELIMINARY DRUMMING SYLLABUS 2004 Written by Greg Bassani B.Ed., Dip.T., B.Tech., APBA Principal of Drumming, 2004. Version 1.0/2004 G:\APBC DRUMMING SYLLABUS 2004_Preliminary.doc

More information

STRUCTURAL ANALYSIS AND SEGMENTATION OF MUSIC SIGNALS

STRUCTURAL ANALYSIS AND SEGMENTATION OF MUSIC SIGNALS STRUCTURAL ANALYSIS AND SEGMENTATION OF MUSIC SIGNALS A DISSERTATION SUBMITTED TO THE DEPARTMENT OF TECHNOLOGY OF THE UNIVERSITAT POMPEU FABRA FOR THE PROGRAM IN COMPUTER SCIENCE AND DIGITAL COMMUNICATION

More information

HST 725 Music Perception & Cognition Assignment #1 =================================================================

HST 725 Music Perception & Cognition Assignment #1 ================================================================= HST.725 Music Perception and Cognition, Spring 2009 Harvard-MIT Division of Health Sciences and Technology Course Director: Dr. Peter Cariani HST 725 Music Perception & Cognition Assignment #1 =================================================================

More information

Predicting the immediate future with Recurrent Neural Networks: Pre-training and Applications

Predicting the immediate future with Recurrent Neural Networks: Pre-training and Applications Predicting the immediate future with Recurrent Neural Networks: Pre-training and Applications Introduction Brandon Richardson December 16, 2011 Research preformed from the last 5 years has shown that the

More information

Computational Rhythm Similarity Development and Verification Through Deep Networks and Musically Motivated Analysis

Computational Rhythm Similarity Development and Verification Through Deep Networks and Musically Motivated Analysis NEW YORK UNIVERSITY Computational Rhythm Similarity Development and Verification Through Deep Networks and Musically Motivated Analysis by Tlacael Esparza Submitted in partial fulfillment of the requirements

More information

Using FR-4x Editor. Introduction. Installation. Connecting the FR-4x to the Editor

Using FR-4x Editor. Introduction. Installation. Connecting the FR-4x to the Editor Introduction FR-4x Editor is an editor for creating sets (SET) and user programs (UPG). This document explains how to edit the parameters and how to save the set or user program. (This document explains

More information

Music Understanding and the Future of Music

Music Understanding and the Future of Music Music Understanding and the Future of Music Roger B. Dannenberg Professor of Computer Science, Art, and Music Carnegie Mellon University Why Computers and Music? Music in every human society! Computers

More information

Low-Latency Instrument Separation in Polyphonic Audio Using Timbre Models

Low-Latency Instrument Separation in Polyphonic Audio Using Timbre Models Low-Latency Instrument Separation in Polyphonic Audio Using Timbre Models Ricard Marxer, Jordi Janer, and Jordi Bonada Universitat Pompeu Fabra, Music Technology Group, Roc Boronat 138, Barcelona {ricard.marxer,jordi.janer,jordi.bonada}@upf.edu

More information

Multidimensional analysis of interdependence in a string quartet

Multidimensional analysis of interdependence in a string quartet International Symposium on Performance Science The Author 2013 ISBN tbc All rights reserved Multidimensional analysis of interdependence in a string quartet Panos Papiotis 1, Marco Marchini 1, and Esteban

More information

Using the MPEG-7 Standard for the Description of Musical Content

Using the MPEG-7 Standard for the Description of Musical Content Using the MPEG-7 Standard for the Description of Musical Content EMILIA GÓMEZ, FABIEN GOUYON, PERFECTO HERRERA, XAVIER AMATRIAIN Music Technology Group, Institut Universitari de l Audiovisual Universitat

More information

Drum Source Separation using Percussive Feature Detection and Spectral Modulation

Drum Source Separation using Percussive Feature Detection and Spectral Modulation ISSC 25, Dublin, September 1-2 Drum Source Separation using Percussive Feature Detection and Spectral Modulation Dan Barry φ, Derry Fitzgerald^, Eugene Coyle φ and Bob Lawlor* φ Digital Audio Research

More information

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

Music Emotion Recognition. Jaesung Lee. Chung-Ang University Music Emotion Recognition Jaesung Lee Chung-Ang University Introduction Searching Music in Music Information Retrieval Some information about target music is available Query by Text: Title, Artist, or

More information

Tempo and Beat Analysis

Tempo and Beat Analysis Advanced Course Computer Science Music Processing Summer Term 2010 Meinard Müller, Peter Grosche Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Tempo and Beat Analysis Musical Properties:

More information

LEARNING A FEATURE SPACE FOR SIMILARITY IN WORLD MUSIC

LEARNING A FEATURE SPACE FOR SIMILARITY IN WORLD MUSIC LEARNING A FEATURE SPACE FOR SIMILARITY IN WORLD MUSIC Maria Panteli, Emmanouil Benetos, Simon Dixon Centre for Digital Music, Queen Mary University of London, United Kingdom {m.panteli, emmanouil.benetos,

More information

First Step Towards Enhancing Word Embeddings with Pitch Accents for DNN-based Slot Filling on Recognized Text

First Step Towards Enhancing Word Embeddings with Pitch Accents for DNN-based Slot Filling on Recognized Text First Step Towards Enhancing Word Embeddings with Pitch Accents for DNN-based Slot Filling on Recognized Text Sabrina Stehwien, Ngoc Thang Vu IMS, University of Stuttgart March 16, 2017 Slot Filling sequential

More information

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the

More information

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception LEARNING AUDIO SHEET MUSIC CORRESPONDENCES Matthias Dorfer Department of Computational Perception Short Introduction... I am a PhD Candidate in the Department of Computational Perception at Johannes Kepler

More information

Laughbot: Detecting Humor in Spoken Language with Language and Audio Cues

Laughbot: Detecting Humor in Spoken Language with Language and Audio Cues Laughbot: Detecting Humor in Spoken Language with Language and Audio Cues Kate Park katepark@stanford.edu Annie Hu anniehu@stanford.edu Natalie Muenster ncm000@stanford.edu Abstract We propose detecting

More information

CONTENT-BASED MELODIC TRANSFORMATIONS OF AUDIO MATERIAL FOR A MUSIC PROCESSING APPLICATION

CONTENT-BASED MELODIC TRANSFORMATIONS OF AUDIO MATERIAL FOR A MUSIC PROCESSING APPLICATION CONTENT-BASED MELODIC TRANSFORMATIONS OF AUDIO MATERIAL FOR A MUSIC PROCESSING APPLICATION Emilia Gómez, Gilles Peterschmitt, Xavier Amatriain, Perfecto Herrera Music Technology Group Universitat Pompeu

More information

MuseGAN: Multi-track Sequential Generative Adversarial Networks for Symbolic Music Generation and Accompaniment

MuseGAN: Multi-track Sequential Generative Adversarial Networks for Symbolic Music Generation and Accompaniment MuseGAN: Multi-track Sequential Generative Adversarial Networks for Symbolic Music Generation and Accompaniment Hao-Wen Dong*, Wen-Yi Hsiao*, Li-Chia Yang, Yi-Hsuan Yang Research Center of IT Innovation,

More information

Dr Kelly Jakubowski Music Psychologist October 2017

Dr Kelly Jakubowski Music Psychologist October 2017 Dr Kelly Jakubowski Music Psychologist October 2017 Overview Musical rhythm: Introduction Rhythm and movement Rhythm and language Rhythm and social engagement Introduction Engaging with music can teach

More information

Automatic Identification of Samples in Hip Hop Music

Automatic Identification of Samples in Hip Hop Music Automatic Identification of Samples in Hip Hop Music Jan Van Balen 1, Martín Haro 2, and Joan Serrà 3 1 Dept of Information and Computing Sciences, Utrecht University, the Netherlands 2 Music Technology

More information

Chairs: Josep Lladós (CVC, Universitat Autònoma de Barcelona)

Chairs: Josep Lladós (CVC, Universitat Autònoma de Barcelona) Session 3: Optical Music Recognition Chairs: Nina Hirata (University of São Paulo) Josep Lladós (CVC, Universitat Autònoma de Barcelona) Session outline (each paper: 10 min presentation) On the Potential

More information

This is why when you come close to dance music being played, the first thing that you hear is the boom-boom-boom of the kick drum.

This is why when you come close to dance music being played, the first thing that you hear is the boom-boom-boom of the kick drum. Unit 02 Creating Music Learners must select and create key musical elements and organise them into a complete original musical piece in their chosen style using a DAW. The piece must use a minimum of 4

More information

Classroom. Chapter 2: Lesson 12

Classroom. Chapter 2: Lesson 12 Classroom Chapter 2: Lesson 12 Adventus Incorporated, 2001 Chapter 2: The Grand Staff Lesson 12 For specific information on Piano Suite Premier and how to purchase, please visit one of the following links:

More information

Music Curriculum Kindergarten

Music Curriculum Kindergarten Music Curriculum Kindergarten Wisconsin Model Standards for Music A: Singing Echo short melodic patterns appropriate to grade level Sing kindergarten repertoire with appropriate posture and breathing Maintain

More information

Music Synchronization. Music Synchronization. Music Data. Music Data. General Goals. Music Information Retrieval (MIR)

Music Synchronization. Music Synchronization. Music Data. Music Data. General Goals. Music Information Retrieval (MIR) Advanced Course Computer Science Music Processing Summer Term 2010 Music ata Meinard Müller Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Music Synchronization Music ata Various interpretations

More information

CS 591 S1 Computational Audio

CS 591 S1 Computational Audio 4/29/7 CS 59 S Computational Audio Wayne Snyder Computer Science Department Boston University Today: Comparing Musical Signals: Cross- and Autocorrelations of Spectral Data for Structure Analysis Segmentation

More information