Learning Musicianship for Automatic Accompaniment

Similar documents
A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS

SPECTRAL LEARNING FOR EXPRESSIVE INTERACTIVE ENSEMBLE MUSIC PERFORMANCE

Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment

Computer Coordination With Popular Music: A New Research Agenda 1

MidiFind: Fast and Effec/ve Similarity Searching in Large MIDI Databases

More About Regression

Gus (Guangyu) Xia , NYU Shanghai, Shanghai, Tel: (412) Webpage:

DELTA MODULATION AND DPCM CODING OF COLOR SIGNALS

Motion Video Compression

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur

QCN Transience and Equilibrium: Response and Stability. Abdul Kabbani, Rong Pan, Balaji Prabhakar and Mick Seaman

Correlation to the Common Core State Standards

Adaptive reference frame selection for generalized video signal coding. Carnegie Mellon University, Pittsburgh, PA 15213

Video coding standards

2D Interleaver Design for Image Transmission over Severe Burst-Error Environment

COMP 249 Advanced Distributed Systems Multimedia Networking. Video Compression Standards

6.UAP Project. FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System. Daryl Neubieser. May 12, 2016

1/ 19 2/17 3/23 4/23 5/18 Total/100. Please do not write in the spaces above.

Restoration of Hyperspectral Push-Broom Scanner Data

DIGITAL COMMUNICATION

Jumpstarters for Math

Welcome Accelerated Algebra 2!

Koester Performance Research Koester Performance Research Heidi Koester, Ph.D. Rich Simpson, Ph.D., ATP

Chapter 2 Introduction to

Research Topic. Error Concealment Techniques in H.264/AVC for Wireless Video Transmission in Mobile Networks

1 Overview. 1.1 Nominal Project Requirements

Recording Form. Part One: Oral Reading. Recording Form. Saving Up Level M Fiction

Temporal Error Concealment Algorithm Using Adaptive Multi- Side Boundary Matching Principle

Creating a Feature Vector to Identify Similarity between MIDI Files

Experiments on musical instrument separation using multiplecause

Section 001. Read this before starting!

Automatic Construction of Synthetic Musical Instruments and Performers

Video-based Vibrato Detection and Analysis for Polyphonic String Music

Comment #147, #169: Problems of high DFE coefficients

CS229 Project Report Polyphonic Piano Transcription

ESTIMATING THE ERROR DISTRIBUTION OF A TAP SEQUENCE WITHOUT GROUND TRUTH 1

PS User Guide Series Seismic-Data Display

Part I: Graph Coloring

in the Howard County Public School System and Rocketship Education

Musical Form. Module 2 of Music: Under the Hood. John Hooker Carnegie Mellon University. Osher Course September 2018

Overview: Video Coding Standards

Bridging the Gap Between CBR and VBR for H264 Standard

CARLETON UNIVERSITY. Facts without theory is trivia. Theory without facts is bull 2607-LRB

CPU Bach: An Automatic Chorale Harmonization System

A Novel Approach towards Video Compression for Mobile Internet using Transform Domain Technique

UC San Diego UC San Diego Previously Published Works

Logic Design II (17.342) Spring Lecture Outline

Music Alignment and Applications. Introduction

Automatic Piano Music Transcription

Predicting the immediate future with Recurrent Neural Networks: Pre-training and Applications

Appeal decision. Appeal No France. Tokyo, Japan. Tokyo, Japan. Tokyo, Japan. Tokyo, Japan. Tokyo, Japan

Exercises. ASReml Tutorial: B4 Bivariate Analysis p. 55

Analysis of local and global timing and pitch change in ordinary

TWO-FACTOR ANOVA Kim Neuendorf 4/9/18 COM 631/731 I. MODEL

Finding Alternative Musical Scales

Analysis of Packet Loss for Compressed Video: Does Burst-Length Matter?

MID-TERM EXAMINATION IN DATA MODELS AND DECISION MAKING 22:960:575

Classroom. Chapter 4: Lesson 22

An Overview of Video Coding Algorithms

Perceptual dimensions of short audio clips and corresponding timbre features

Enhancing Switch Use and Scanning for People with Physical Impairments. Introduction Heidi Koester, Ph.D. Part 2. What We ll Cover

ECG Denoising Using Singular Value Decomposition

Implementation of an MPEG Codec on the Tilera TM 64 Processor

NETFLIX MOVIE RATING ANALYSIS

Joint Optimization of Source-Channel Video Coding Using the H.264/AVC encoder and FEC Codes. Digital Signal and Image Processing Lab

Setting Energy Efficiency Requirements Using Multivariate Regression

Processing data with Mestrelab Mnova

Computer Music Journal, Volume 38, Number 2, Summer 2014, pp (Article)

SUMMIT LAW GROUP PLLC 315 FIFTH AVENUE SOUTH, SUITE 1000 SEATTLE, WASHINGTON Telephone: (206) Fax: (206)

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

PACKET-SWITCHED networks have become ubiquitous

10GBASE-R Test Patterns

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University

Joint source-channel video coding for H.264 using FEC

I. Model. Q29a. I love the options at my fingertips today, watching videos on my phone, texting, and streaming films. Main Effect X1: Gender

A Novel Macroblock-Level Filtering Upsampling Architecture for H.264/AVC Scalable Extension

LEARNING TO CONTROL A REVERBERATOR USING SUBJECTIVE PERCEPTUAL DESCRIPTORS

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon

A STUDY OF ENSEMBLE SYNCHRONISATION UNDER RESTRICTED LINE OF SIGHT

Week 14 Music Understanding and Classification

Visual Encoding Design

H.261: A Standard for VideoConferencing Applications. Nimrod Peleg Update: Nov. 2003

Drift Compensation for Reduced Spatial Resolution Transcoding

North Carolina Standard Course of Study - Mathematics

AUTOMATIC ACCOMPANIMENT OF VOCAL MELODIES IN THE CONTEXT OF POPULAR MUSIC

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception

NON-LINEAR EFFECTS MODELING FOR POLYPHONIC PIANO TRANSCRIPTION

Structured training for large-vocabulary chord recognition. Brian McFee* & Juan Pablo Bello

Luma Adjustment for High Dynamic Range Video

GBA 327: Module 7D AVP Transcript Title: The Monte Carlo Simulation Using Risk Solver. Title Slide

Chapter 27. Inferences for Regression. Remembering Regression. An Example: Body Fat and Waist Size. Remembering Regression (cont.)

4:1 Mux Symbol 4:1 Mux Circuit

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

University of Bristol - Explore Bristol Research. Peer reviewed version. Link to published version (if available): /ISCAS.2005.

FAST SPATIAL AND TEMPORAL CORRELATION-BASED REFERENCE PICTURE SELECTION

A Bayesian Network for Real-Time Musical Accompaniment

Film Grain Technology

EE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function

Joint bottom-up/top-down machine learning structures to simulate human audition and musical creativity

Bite Size Brownies. Designed by: Jonathan Thompson George Mason University, COMPLETE Math

Transcription:

Learning Musicianship for Automatic Accompaniment Gus (Guangyu) Xia Roger Dannenberg School of Computer Science Carnegie Mellon University

2 Introduction: Musical background Interaction Expression Rehearsal

3 Introduction: Technical background Musicianship Expressive Performance Interaction Score following and automatic accompaniment Expressive Interactive Performance

Introduction: Problem definition For interactive music performance, how can we build artificial performers that automatically improve their ability to sense and coordinate with human musicians expression with rehearsal experience? How to interpret the music based on the expression of human musicians? How to distill models from rehearsals? What are the limits of validity of the learned models? How many rehearsals are needed? We start from piano duets, focusing on expressive timing and expressive dynamics. 4

5 Outline Introduction Data Collection Methods Demos Conclusion & Future Work

6 Current Data Collection Musicians: 10 music master students play duet pieces in 5 pairs. Music pieces: 3 pieces of music are selected, Danny boy, Serenade (by Schubert), and Ashokan Farewell. Each pair performs every piece of music 7 times. Recording settings: Recorded by electronic pianos with MIDI output.

7 Outline Introduction Data Collection Methods Demos Conclusion & Future Work

Method Overview From local to general Local: low-dimensional feature space, only apply to certain notes General: high-dimensional feature space, apply to the whole piece of music Base line: Score Following Automatic Accompaniment Real time Real time predicted next note s performance time next note s reference time Reference time Reference time 8

Method (1): Note-specific approach Idea: Expressive timings of the notes are linearly correlated. Predict the expressive timing of 2 nd piano by the expressive timing of 1 st piano. = [,,, ] =,,, Model: = + "#$( ) 9

10 Result: Note-specific approach Mean Absolute Error: BL: 0.098 Note-8: 0.087 Note-34: 0.060 Time residual (sec) 0.25 0.2 0.15 0.1 0.05 BL Note 8 Note 34 0 0 10 20 30 40 50 60 Score time (sec)

Method (2): Rhythm-specific approach Idea: Notes with same score rhythm context share parameters. Introduces an extra dummy variable to encode the score rhythm context of each note. = [,,, ] =,,, Model: "( =,) "( +,) "#$( ) 11

12 Result: Rhythm-specific approach Mean Absolute Error: BL: 0.098 Rhythm-4: 0.084 Rhythm-8: 0.067 Time residual (sec) 0.25 0.2 0.15 0.1 0.05 BL Rhythm 4 Rhythm 8 0 0 10 20 30 40 50 60 Score time (sec)

13 Method (3): General feature approach Idea: Make the model more general. Predict the expressive timing by considering more than score rhythm context. =,,, =,,, Model: =

14 Regularization: Group Lasso Idea: Reduces the burden for training. Discover the dominant features that could predict the expressive timings. Solve: min " +

15 Result: General feature approach Mean Absolute Error: (ONLY 4 training pieces) BL: 0.098 LR: 0.072 Glasso: 0.059 Time residual (sec) 0.25 0.2 0.15 0.1 0.05 BL LR Glasso 0 0 10 20 30 40 50 60 Score time (sec)

16 Method (4): LDS approach Idea: Add another regularization by adjacent notes. Lower dimensional hidden mental states that control the expressive timings. u t- 1 u t u t+1 z t- 1 z t z t+1 y t- 1 y t y t+1 Model: = + + ~(0, ) = + + ~(0, )

17 Result: LDS (horizontal regularization) Mean Absolute Error: (ONLY 4 training pieces) BL: 0.085 LR: 0.072 LDS: 0.067 Time residual (sec) 0.25 0.2 0.15 0.1 0.05 BL LR LDS 0 0 10 20 30 40 50 60 Score time (sec)

18 A Global View Time residual(sec) 0.18 0.16 0.14 0.12 0.1 0.08 0.06 0.04 0.02 Serenade Danny boy Ashokan Farewell 0 BL Note 4 Rhythm 4 Glasso 4 Note 8 Rhythm 8 Glasso 8 Note 34 Rhythm 34Glasso 34 Methods and training size

19 Outline Introduction Data Collection Methods Demos Conclusion & Future Work

20 Some Initial Audio Demo Base Line: Note-specific approach: 34 training examples General feature approach, group lasso: 4 training examples

21 Future Work Cross-piece models Performer-specific models Online learning and decoding Plugin with music robots

22 Conclusion An artificial performer for interactive performance Learn musicianship from rehearsal experience A combination of expressive performance and automatic accompaniment Much better prediction just based on 4 rehearsals

23 Q&A

Spectral Learning(1): Oblique projections u t- 1 u t u t+1 z t- 1 z t z t+1 y t- 1 y t y t+1 ( ) = [ ] We don t know the future. Partially explain future observations based on the history [ 0] 0 24

25 Spectral Learning(2): state estimation u t- 1 u t u t+1 z t- 1 z t z t+1 = " y t- 1 y t y t+1 States estimation by SVD = Σ = (Σ )(Σ ) Moreover, enforce a bottleneck by throwing out near-zero singular values and corresponding columns in U and V.

26 Spectral Learning(3): Estimate parameter u t- 1 u t u t+1 z t- 1 z t z t+1 y t- 1 y t y t+1 = + + ~(0, ) = + + ~(0, ) Based on estimated hidden states, the parameters could be estimated from the following equation: = +