jsymbolic 2: New Developments and Research Opportunities

Similar documents
jsymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada

Style-independent computer-assisted exploratory analysis of large music collections

SIMSSA DB: A Database for Computational Musicological Research

EVALUATING THE GENRE CLASSIFICATION PERFORMANCE OF LYRICAL FEATURES RELATIVE TO AUDIO, SYMBOLIC AND CULTURAL FEATURES

Methodologies for Creating Symbolic Early Music Corpora for Musicological Research

Outline. Why do we classify? Audio Classification

Distributed Digital Music Archives and Libraries (DDMAL)

Specifying Features for Classical and Non-Classical Melody Evaluation

CSC475 Music Information Retrieval

Composer Style Attribution

MUSI-6201 Computational Music Analysis

A Basis for Characterizing Musical Genres

Creating a Feature Vector to Identify Similarity between MIDI Files

MUSIC CURRICULM MAP: KEY STAGE THREE:

Connecticut State Department of Education Music Standards Middle School Grades 6-8

Audio Feature Extraction for Corpus Analysis

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Feature-Based Analysis of Haydn String Quartets

MUSIC THEORY CURRICULUM STANDARDS GRADES Students will sing, alone and with others, a varied repertoire of music.

MUSICAL STRUCTURAL ANALYSIS DATABASE BASED ON GTTM

Music. Last Updated: May 28, 2015, 11:49 am NORTH CAROLINA ESSENTIAL STANDARDS

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

Introductions to Music Information Retrieval

Tool-based Identification of Melodic Patterns in MusicXML Documents

Automatic Music Clustering using Audio Attributes

ANNOTATING MUSICAL SCORES IN ENP

Largo Adagio Andante Moderato Allegro Presto Beats per minute

Representing, comparing and evaluating of music files

Curriculum Standard One: The student will listen to and analyze music critically, using the vocabulary and language of music.

METHODOLOGIES FOR CREATING SYMBOLIC CORPORA OF WESTERN MUSIC BEFORE 1600

Curriculum Standard One: The student will listen to and analyze music critically, using the vocabulary and language of music.

Melody Retrieval On The Web

Beethoven, Bach, and Billions of Bytes

Automatic Rhythmic Notation from Single Voice Audio Sources

6.UAP Project. FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System. Daryl Neubieser. May 12, 2016

Singer Traits Identification using Deep Neural Network

Video-based Vibrato Detection and Analysis for Polyphonic String Music

HS Music Theory Music

Curriculum Standard One: The student will listen to and analyze music critically, using the vocabulary and language of music.

Analysis and Clustering of Musical Compositions using Melody-based Features

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

A Planned Course Statement for. Music Theory, AP. Course # 760 Grade(s) 11, 12. Length of Period (mins.) 40 Total Clock Hours: 120

J-Syncker A computational implementation of the Schillinger System of Musical Composition.

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

PKUES Grade 10 Music Pre-IB Curriculum Outline. (adapted from IB Music SL)

A MANUAL ANNOTATION METHOD FOR MELODIC SIMILARITY AND THE STUDY OF MELODY FEATURE SETS

The purpose of this essay is to impart a basic vocabulary that you and your fellow

Musical Harmonization with Constraints: A Survey. Overview. Computers and Music. Tonal Music

Automatic Piano Music Transcription

Exploring the Design Space of Symbolic Music Genre Classification Using Data Mining Techniques Ortiz-Arroyo, Daniel; Kofod, Christian

2013 Assessment Report. Music Level 1

Figured Bass and Tonality Recognition Jerome Barthélemy Ircam 1 Place Igor Stravinsky Paris France

Music Performance Solo

Music Performance Ensemble

STRING QUARTET CLASSIFICATION WITH MONOPHONIC MODELS

Music Information Retrieval (MIR)

Topic 10. Multi-pitch Analysis

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

Copyright 2009 Pearson Education, Inc. or its affiliate(s). All rights reserved. NES, the NES logo, Pearson, the Pearson logo, and National

Music Genre Classification and Variance Comparison on Number of Genres

Unit 1. π π π π π π. 0 π π π π π π π π π. . 0 ð Š ² ² / Melody 1A. Melodic Dictation: Scalewise (Conjunct Diatonic) Melodies

Singer Recognition and Modeling Singer Error

QUALITY OF COMPUTER MUSIC USING MIDI LANGUAGE FOR DIGITAL MUSIC ARRANGEMENT

Stylistic features Antonio Vivaldi: Concerto in D minor, Op. 3 No. 11

CS229 Project Report Polyphonic Piano Transcription

Music Curriculum Kindergarten

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

Student Performance Q&A:

CALCULATING SIMILARITY OF FOLK SONG VARIANTS WITH MELODY-BASED FEATURES

Exploring Melodic Features for the Classification and Retrieval of Traditional Music in the Context of Cultural Source

I. Students will use body, voice and instruments as means of musical expression.

Articulation Clarity and distinct rendition in musical performance.

Palestrina Pal: A Grammar Checker for Music Compositions in the Style of Palestrina

Lesson Week: August 17-19, 2016 Grade Level: 11 th & 12 th Subject: Advanced Placement Music Theory Prepared by: Aaron Williams Overview & Purpose:

Assessment Schedule 2016 Music: Demonstrate knowledge of conventions in a range of music scores (91276)

AP Music Theory at the Career Center Chris Garmon, Instructor

Curriculum Standard One: The student will listen to and analyze music critically, using vocabulary and language of music.

Essentials Skills for Music 1 st Quarter

Sequential Association Rules in Atonal Music

Music Information Retrieval (MIR)

FREE TV AUSTRALIA OPERATIONAL PRACTICE OP- 59 Measurement and Management of Loudness in Soundtracks for Television Broadcasting

Transcription An Historical Overview

Semi-supervised Musical Instrument Recognition

Semi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis

INTERACTIVE GTTM ANALYZER

LESSON 1 PITCH NOTATION AND INTERVALS

II. Prerequisites: Ability to play a band instrument, access to a working instrument

Music Information Retrieval

Student Performance Q&A:

THE POTENTIAL FOR AUTOMATIC ASSESSMENT OF TRUMPET TONE QUALITY

Supervised Learning in Genre Classification

Sequential Association Rules in Atonal Music

Music and Text: Integrating Scholarly Literature into Music Data

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t

CPU Bach: An Automatic Chorale Harmonization System

Assessment Schedule 2017 Music: Demonstrate knowledge of conventions in a range of music scores (91276)

MUSIC PERFORMANCE: GROUP

MUSIC100 Rudiments of Music

Banff Sketches. for MIDI piano and interactive music system Robert Rowe

Missouri Educator Gateway Assessments

Transcription:

jsymbolic 2: New Developments and Research Opportunities Cory McKay Marianopolis College and CIRMMT Montreal, Canada

2 / 30 Topics Introduction to features (from a machine learning perspective) And how they can be useful for musicologists and music theorists jsymbolic2 What it is How it s useful to music theorists and musicologists

3 / 30 What are features? Pieces of information that can characterize something (e.g. a piece of music) in a (usually) simple way (Usually) numerical values Can be single values or can be vectors of related values Histograms are a common type of vector (Usually) represent a piece as a whole Or at least regularly spaced windows / musical segments within the piece

4 / 30 Chopin s Nocturne in B, Op. 32, No. 1 Average Note To Note Dynamics Change: 6.03 Chromatic Motion: 0.0769 Dominant Spread: 3 Harmonicity of Two Strongest Rhythmic Pulses: 1 Importance of Bass Register: 0.2 Interval Between Strongest Pitch Classes: 3 Most Common Pitch Class Prevalence: 0.433 Note Density: 3.75 Number of Common Melodic Intervals: 3 Number of Strong Pulses: 5 Orchestral Strings Fraction: 0 Overall Dynamic Range: 62 Pitch Class Variety: 7 Range: 48 Relative Strength of Most Common Intervals: 0.5 Size of Melodic Arcs: 11 Stepwise Motion: 0.231 Strength of Strongest Rhythmic Pulse: 0.321 Variability of Note Duration: 0.293 Variation of Dynamics: 16.4

5 / 30 Mendelssohn s Piano Trio No. 2 Average Note To Note Dynamics Change: 1.46 Chromatic Motion: 0.244 Dominant Spread: 2 Harmonicity of Two Strongest Rhythmic Pulses: 1 Importance of Bass Register: 0.373 Interval Between Strongest Pitch Classes: 7 Most Common Pitch Class Prevalence: 0.39 Note Density: 29.5 Number of Common Melodic Intervals: 6 Number of Strong Pulses: 6 Orchestral Strings Fraction: 0.56 Overall Dynamic Range: 22 Pitch Class Variety: 7 Range: 39 Relative Strength of Most Common Intervals: 0.8 Size of Melodic Arcs: 7.27 Stepwise Motion: 0.439 Strength of Strongest Rhythmic Pulse: 0.173 Variability of Note Duration: 0.104 Variation of Dynamics: 5.98

6 / 30 Feature value comparison Feature Nocturne Trio Average Note To Note Dynamic Change 6.03 1.46 Overall Dynamic Range 62 22 Variation of Dynamics 16.4 5.98 Note Density 3.75 29.5 Orchestral Strings Fraction 0 0.56 Variability of Note Duration 0.293 0.104 Chromatic Motion 0.077 0.244 Range 48 39

Relative Frequency Relative Frequency 7 / 30 Fifths pitch class histogram 0.3 Fifths Pitch Histogram: Four Seasons (Spring) by Vivaldi 0.3 Fifths Pitch Histogram: Sechs Kleine Klavierstücke by Schoenberg 0.25 0.25 0.2 0.2 0.15 0.15 0.1 0.1 0.05 0.05 0 0 1 2 3 4 5 6 7 8 9 10 11 0 0 1 2 3 4 5 6 7 8 9 10 11 Folded Pitch Class Folded Pitch Class

Relative Frequenc Relative Frequenc 8 / 30 Beat histogram Beat histograms use a technique called autocorrelation to calculate the relative strengths of different beat periodicities I Wanna Be Sedated by The Ramones (top) Several harmonic peaks with large spreads around them Round Midnight by Thelonious Monk (bottom) Only one strong peak, with a large low-level spread Histograms like this can be used directly, or other features may be derived from them e.g. peak statistics Beat Histogram: I Wanna Be Sedated by The Ramones 0.10 0.09 0.08 0.07 0.06 0.05 0.04 0.03 0.02 0.01 0.00 40 50 60 70 80 90 100 110 120 130 140 150 160 170 180 190 200 Beats Per Minute Beat Histogram: 'Round Mdinight by Thelonious Monk 0.30 0.25 0.20 0.15 0.10 0.05 0.00 40 50 60 70 80 90 100 110 120 130 140 150 160 170 180 190 200 Beats Per Minute

9 / 30 How can features be useful? Sophisticated searches of large musical databases e.g. find all pieces with no more than X amount of chromaticism, and less than Y amount of parallel motion ELVIS database + Musiclibs Using statistical analysis and visualization tools to study the empirical musical importance of various features when extracted from large datasets e.g. features based on instrumentation were most effective for distinguishing genres (McKay & Fujinaga 2005) Using machine learning to classify or cluster music Supervised or unsupervised learning e.g. identify the composers of unattributed musical pieces

10 / 30 Sample expert system if ( parallel_fifths == 0 && landini_cadences == 0 ) then composer Palestrina else composer Machaut

11 / 30 Sample supervised learning Supervised Learning Ockeghem Josquin

12 / 30 Sample supervised learning Supervised Learning Unknown (Ockeghem) Unknown (Josquin)

13 / 30 Sample supervised learning Supervised Learning Ockeghem Josquin Unknown (Ockeghem) Unknown (Josquin)

14 / 30 Sample supervised learning Supervised Learning Ockeghem Josquin Unknown (Ockeghem) Unknown (Josquin)

15 / 30 Sample unsupervised learning Unsupervised Learning Composer 1 Composer 2 Composer 3 Composer 4

16 / 30 Benefits of features and machine learning Can quickly perform consistent empirical studies involving thousands of pieces Can be applied to diverse types of music Can simultaneously consider thousands of features and their interrelationships And can statistical condense many features into low-dimensional spaces when needed No need to formally specify any heuristics or queries before beginning analyses Unless you want to, of course Can avoid (or validate) potentially incorrect ingrained biases and assumptions

17 / 30 jsymbolic s lineage Bodhidharma (2004) Specialized feature extraction and machine learning for genre classification research jsymbolic (2006) General-purpose feature extraction Part of jmir jsymbolic2 (2016) Bigger and better!

18 / 30 What does jsymbolic2 do? Extracts 158 features Some of these are multi-dimensional histograms, including: Pitch and pitch class histograms Melodic interval histogram Vertical interval histograms Chord types histogram Beat histogram Instrument histograms

19 / 30 jsymbolic2 s feature types (1/2) Instrumentation: What types of instruments are present and which are given particular importance relative to others? Texture: How many independent voices are there and how do they interact (e.g., polyphonic, homophonic, etc.)? Rhythm: Time intervals between the attacks of different notes Duration of notes What kinds of meters and rhythmic patterns are present? Rubato? Dynamics: How loud are notes and what kinds of dynamic variations occur?

20 / 30 jsymbolic feature types (2/2) Pitch Statistics: What are the occurrence rates of different pitches and pitch classes? How tonal is the piece? How much variety in pitch is there? Melody: What kinds of melodic intervals are present? How much melodic variation is there? What kinds of melodic contours are used? What types of phrases are used? Chords: What vertical intervals are present? What types of chords do they represent? How much harmonic movement is there?

21 / 30 How can you use jsymbolic2 Graphical user interface Command line interface Rodan workflow Java API

22 / 30 jsymbolic2 s file formats Input: MIDI MEI MusicXML (via Rodan workflow only) Output: ACE XML Weka ARFF CSV

23 / 30 jsymbolic2 s documentation Super-mega-ultra detailed manual At least compared to most academic software manuals In HTML Super-mega-ultra detailed Javadocs For programmers

24 / 30 jsymbolic2: More great things Windowed feature extraction Including overlapping windows Configuration files Pre-set feature choices Pre-set input and output choices More jmei2midi Most complete MEI to MIDI converter in the universe! General-purpose (not just for jsymbolic2) Specialized pipeline for transmitting relevant MEI data that cannot be represented in MIDI

25 / 30 Exploratory simple pilot study Josquin vs. Ockeghem composer identification / attribution 124 jsymbolic2 features extracted from the JRP data 105 Josquin pieces and 98 Ockeghem Achieved 89.7% classification accuracy 10-fold cross-validation Lots of room for improving results still further Only used simple SVM classifier with default settings No dimensionality reduction was used Both expert insights and automatic analysis can be applied Still more jsymbolic2 features to come Interesting future research applications: Determine which features are most effective Can analyze feature data both visually and statistically Apply trained classifiers to unattributed or uncertain pieces Expand scope to other composers

26 / 30 What you can do with jsymbolic Empirically study huge collections of music in new ways Search music databases based on feature values Analyze and visualize music based on feature values Use machine learning Design your own custom features jsymbolic2 is specifically designed to make it easy to add new custom features Easy to iteratively build increasingly complex features based on existing features Perform multimodal research Combine symbolic features with other features extracted from audio, lyrics and cultural data This improves results substantially! (McKay et al. 2010)

27 / 30 Use jsymbolic2 with jmir ACE: Meta-learning classification engine Bodhidharma MIDI, SLAC and Codaich: datasets jaudio: Audio feature extraction jlyrics: Extracts features from lyrical transcriptions jwebminer: Cultural feature extraction ACE XML: File formats Features, feature metadata, instance metadata and ontologies lyricfetcher: Lyric mining jmusicmetamanager: Metadata management jsongminer: Metadata harvesting jmirutilities: Infrastructure for conducting experiments jproductioncritic: Automated production error-checking

28 / 30 Research collaborations We would love to collaborate with music theorists and musicologists on their work We can help you apply and adapt jsymbolic to specific research projects We can help you come up with novel ways to study music

29 / 30 jsymbolic2: Currently in progress Final testing and debugging Annotation of all valid files in the ELVIS database with extracted features And Musiclibs, eventually Auto-annotation scripts MEI pre-modern notation Designing new features Requests welcome!

30 / 30 Acknowledgements Tristano Tenaglia Implemented almost all of the new jsymbolic2 code Véronique Lagacé ELVIS database integration scripts Ryan Bannon, Dr. Andrew Hankinson and Dr. Reiner Krämer On-site Rodan and ELVIS expertise in the lab Prof. Ichiro Fujinaga and Prof. Julie Cumming Grant application and project supervision superstardom The FRQSC and SSHRC Great financial generosity

Thanks for your attention E-mail: cory.mckay@mail.mcgill.ca jsymbolic2: github.com/ddmal/jsymbolic2 jmir: jmir.sourceforge.net