Music Complexity Descriptors. Matt Stabile June 6 th, 2008

Similar documents
Curriculum Standard One: The student will listen to and analyze music critically, using the vocabulary and language of music.

Music Representations

Tempo and Beat Analysis

Automatic music transcription

Topics in Computer Music Instrument Identification. Ioanna Karydi

Analysis, Synthesis, and Perception of Musical Sounds

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)

Curriculum Standard One: The student will listen to and analyze music critically, using vocabulary and language of music.

The song remains the same: identifying versions of the same piece using tonal descriptors

MUSI-6201 Computational Music Analysis

Topic 10. Multi-pitch Analysis

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t

Concert halls conveyors of musical expressions

LOUDNESS EFFECT OF THE DIFFERENT TONES ON THE TIMBRE SUBJECTIVE PERCEPTION EXPERIMENT OF ERHU

MUSIC CURRICULM MAP: KEY STAGE THREE:

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

CSC475 Music Information Retrieval

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng

Measurement of overtone frequencies of a toy piano and perception of its pitch

EMERGENT SOUNDSCAPE COMPOSITION: REFLECTIONS ON VIRTUALITY

Music Representations

MOTIVATION AGENDA MUSIC, EMOTION, AND TIMBRE CHARACTERIZING THE EMOTION OF INDIVIDUAL PIANO AND OTHER MUSICAL INSTRUMENT SOUNDS

Curriculum Standard One: The student will listen to and analyze music critically, using the vocabulary and language of music.

Assessment Schedule 2016 Music: Demonstrate knowledge of conventions in a range of music scores (91276)

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

Curriculum Standard One: The student will listen to and analyze music critically, using the vocabulary and language of music.

Music for Alto Saxophone & Computer

Influence of timbre, presence/absence of tonal hierarchy and musical training on the perception of musical tension and relaxation schemas

Pitch Perception. Roger Shepard

Speech To Song Classification

I. LISTENING. For most people, sound is background only. To the sound designer/producer, sound is everything.!tc 243 2

Creating a Feature Vector to Identify Similarity between MIDI Files

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University

Automatic Music Clustering using Audio Attributes

Musical Acoustics Lecture 15 Pitch & Frequency (Psycho-Acoustics)

Outline. Why do we classify? Audio Classification

Supervised Learning in Genre Classification

MELODY EXTRACTION FROM POLYPHONIC AUDIO OF WESTERN OPERA: A METHOD BASED ON DETECTION OF THE SINGER S FORMANT

MHSIB.5 Composing and arranging music within specified guidelines a. Creates music incorporating expressive elements.

Efficient Computer-Aided Pitch Track and Note Estimation for Scientific Applications. Matthias Mauch Chris Cannam György Fazekas

LESSON 1 PITCH NOTATION AND INTERVALS

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

Title Music Grade 4. Page: 1 of 13

LEVELS IN NATIONAL CURRICULUM MUSIC

LEVELS IN NATIONAL CURRICULUM MUSIC

Singing accuracy, listeners tolerance, and pitch analysis

Quarterly Progress and Status Report. Violin timbre and the picket fence

Title Music Grade 3. Page: 1 of 13

Towards Music Performer Recognition Using Timbre Features

Drunken Sailor The Melody

Transcription of the Singing Melody in Polyphonic Music

Curriculum Standard One: The student will listen to and analyze music critically, using the vocabulary and language of music.

UNIVERSITY OF DUBLIN TRINITY COLLEGE

Classification of Timbre Similarity

TYING SEMANTIC LABELS TO COMPUTATIONAL DESCRIPTORS OF SIMILAR TIMBRES

The Tone Height of Multiharmonic Sounds. Introduction

STRUCTURAL CHANGE ON MULTIPLE TIME SCALES AS A CORRELATE OF MUSICAL COMPLEXITY

Assessment Schedule 2017 Music: Demonstrate knowledge of conventions in a range of music scores (91276)

MUSI 1900 Notes: Christine Blair

Pitch. The perceptual correlate of frequency: the perceptual dimension along which sounds can be ordered from low to high.

CS229 Project Report Polyphonic Piano Transcription

Psychoacoustics. lecturer:

CTP 431 Music and Audio Computing. Basic Acoustics. Graduate School of Culture Technology (GSCT) Juhan Nam

CONTENT-BASED MELODIC TRANSFORMATIONS OF AUDIO MATERIAL FOR A MUSIC PROCESSING APPLICATION

ON THE USE OF PERCEPTUAL PROPERTIES FOR MELODY ESTIMATION

Tempo and Beat Tracking

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

PREDICTING THE PERCEIVED SPACIOUSNESS OF STEREOPHONIC MUSIC RECORDINGS

Audio Feature Extraction for Corpus Analysis

CTP431- Music and Audio Computing Musical Acoustics. Graduate School of Culture Technology KAIST Juhan Nam

Analyzer Documentation

Robert Alexandru Dobre, Cristian Negrescu

Grade 4 Music Curriculum Maps

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

The Human Features of Music.

Semi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis

BIG IDEAS. Music is a process that relies on the interplay of the senses. Learning Standards

MEMORY & TIMBRE MEMT 463

NATIONAL and KENTUCKY ACADEMIC STANDARDS in MUSIC

Music Study Guide. Moore Public Schools. Definitions of Musical Terms

Instrumental Performance Band 7. Fine Arts Curriculum Framework

use individual notes, chords, and chord progressions to analyze the structure of given musical selections. different volume levels.

Music Segmentation Using Markov Chain Methods

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

GCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam

Boulez. Aspects of Pli Selon Pli. Glen Halls All Rights Reserved.

A perceptual assessment of sound in distant genres of today s experimental music

Transcription An Historical Overview

We realize that this is really small, if we consider that the atmospheric pressure 2 is

Grade Level 5-12 Subject Area: Vocal and Instrumental Music

Experiments on musical instrument separation using multiplecause

A New Method for Calculating Music Similarity

The Mathematics of Music and the Statistical Implications of Exposure to Music on High. Achieving Teens. Kelsey Mongeau

PERFORMING ARTS. Head of Music: Cinzia Cursaro. Year 7 MUSIC Core Component 1 Term

Advanced Orchestra Performance Groups

A Need for Universal Audio Terminologies and Improved Knowledge Transfer to the Consumer

Smooth Rhythms as Probes of Entrainment. Music Perception 10 (1993): ABSTRACT

Connecticut State Department of Education Music Standards Middle School Grades 6-8

Scoregram: Displaying Gross Timbre Information from a Score

Transcription:

Music Complexity Descriptors Matt Stabile June 6 th, 2008

Musical Complexity as a Semantic Descriptor Modern digital audio collections need new criteria for categorization and searching. Applicable to: Music Collections Media Asset Management Systems (Sound Effects Libraries) Problem: Collections too large for manual assignment of descriptors by humans Solution: Automatic computation of descriptors based on the audio file itself

Existing Applications SIMAC: Semantic Interaction with Music Audio Contents

Existing Applications FindSounds Palette Uses Sound Matching Technology to search by sound prototype Also can search by name, description, category, genre, source, copyright, format, size, number of channels, resolution, sample rate, duration, key, and tempo Could use further descriptors to refine search

Defining Music Complexity Sebastian Streich Music Complexity: A Multi Faceted Description of Audio Content, 2007 Complexity of music as a high level, intuitive attribute, which can be experienced directly or indirectly by the active listener Streich defines musical complexity as that property of a musical unit that determines how much effort the listener has to put into following and understanding it

Complexity Facets Finnas (1989) states that unusual harmonies and timbres, irregular tempi and rhythms, unexpected tone sequences and variations in volume raise the level of perceived complexity in music. Facets of music are at least partly independent Complex rhythms with no melodic voice Unexpected volume and timbre changes with simple melody and chord sequences Useful to analyze these facets separately to obtain better complexity descriptors

Complexity Facets: Song Level (Streich) Acoustic Complexities Dynamic: loudness evolution within a track Spatial: disparity of the stereo channels Tonal Complexity Melodic and Harmonic complexity Most difficult due to imperfect transcription Timbral Complexity Timbral texture of a track, # different instr. Rhythmic Complexity Danceability

Methods and Algorithms Timbre Complexity Methods (Streich): Unsupervised HMMs using MFCCs Produce finite set of timbre models for a given input signal Complexity measure: number of models created = number of different inst. textures Too computationally expensive (Repeated training on top of feature extraction) Too unsupervised! Are the HMMs created perceptually meaningful? LZ77 Compression Gain Uses Timbre Symbols to apply entropy estimation from information theory Models human memory: 3 5s chunks of audio are used Timbre Symbols: Bass, Presence, Spectral Roll off, Spectral Flatness Measure Complexity Measure: low compression factor = source entropy low, complexity low Problem: Only computationally practical with coarse quantization, but this limits accuracy. Can guarantee different symbols different perceptual impression, but not the reverse. Spectral Envelope Matching Chosen method, based on idea that it takes a change of at least 4 db in higher harmonics and 10 db in low harmonics to distinguish timbre of two tones. (Winckel)

Timbre/Dynamics Algorithm Loudness Estimation & Spectral Envelope Matching Utilizes functions from Palmpak s Matlab MA Toolbox

Pre Processing FFT w/hann Window Normalized Power Spectrum Terhardt s Outer Ear Frequency Weighting Bark Scale (critical bands of hearing) Heuristic Spreading Function (Spectral Masking)

Sonogram Plots Sone scale: linear correspondence to human loudness perception

Dynamic Complexity Relates to properties of the loudness evolution In terms of Abruptness and Rate of Changes in dynamic level. Dynamic range and time scope important After Pre-processing: Total loudness estimate for each frame, Mmax = band with max loudness: Complexity: Average fluctuation of successive loudness values: Comp Values: a) 0.134 c) 0.304 b) 0.247 d) 0.488

Timbral Complexity Spectral Envelope Matching Avoids hard quantization necessary in LZ77 by using human perception of timbre. (4 db in higher harmonics and 10 db in low harmonics to distinguish timbre of two tones) Compares changes in spectral envelope rather than single harmonics Complexity Measure: Counts amount of timbral change in a given temporal window and then extract a complexity number Bandwise comparison with preceding frames, 6dB threshold, reaches back 80ms to 4s, approx. human memory Band Loudness Similarities Complexity Measure: % frames with similarity = 0 a) Bagpipe: 8.5% b) Symphony: 11.3% c) Rap: 4.8%

Complexity on Samples Started off aiming to calculate timbral complexity and dynamic complexity of a sample individually Influence of dynamics on timbre became apparent Dynamic level + technique of performance (i.e. vibrato, pizzicato, spiccato) strongly effects timbre of the tone Proposed method: Use Spectral Envelope Matching average fluctuation of similar frames combined with amount the sign of the derivative changes.

Timbre/Dynamics Algorithm: Sample Level MedianFiltering Average Fluctuation of Similar Frames/ Derivation Sign Change Counting

Results Reference Samples: Triangle wave, Sine Wave No perceived change in Timbre Avg Frames Fluct: = 1 Deriv sign change = 0 Loudness Evolution Sonogram of Sine Oscillator Band Similarity Plot

Results Samples: Cello and Harp playing at varying dynamic levels Avg Frames Fluct: = 0.8 Deriv sign change = 97 Loudness Evolution Sonogram of Cello playing mf Band Similarity Plot

Results Timbral Complexity Results: Avg. Frame Fluc Deriv Sine: 1 0 Triangle 1 0 Saw(Arp) 3.9 53 Cello: p = 1.05 130 mf = 0.8 97 f = 0.76 159 Harp: p = 2.4 403 mf = 4.06 635 f = 6.7 461 Band loudness similarities for Cello Avg. Frame Fluc reveals how extreme the timbral jumps are Derivative sign change count indicator of overall number of changes in timbre Categories: - No Timbral Change = Sine/Triangle - Periodic Timbral Change = Saw - Measurable Non-Periodic Timbral Change = Cello/Harp - No Similar Frames (Timbre always changing ) = Some Crazy Signal

Conclusions Need to find robust calculation of timbral complexity number Ex: Streich Timbral Complexity: % frames with sim. = 0 Can definitely be useful as attributes for refined sorting of samples Would need surveys completed to find correlation to human perception and then tuned accordingly.