REALTIME ANALYSIS OF DYNAMIC SHAPING

Similar documents
Quarterly Progress and Status Report. Perception of just noticeable time displacement of a tone presented in a metrical sequence at different tempos

Concert halls conveyors of musical expressions

Scoregram: Displaying Gross Timbre Information from a Score

Director Musices: The KTH Performance Rules System

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

Loudness and Sharpness Calculation

CS229 Project Report Polyphonic Piano Transcription

DIFFERENCES IN TRAFFIC NOISE MEASUREMENTS WITH SLM AND BINAURAL RECORDING HEAD

CHILDREN S CONCEPTUALISATION OF MUSIC

INTERMEDIATE STUDY GUIDE

Simple Harmonic Motion: What is a Sound Spectrum?

Analysis of local and global timing and pitch change in ordinary

Quarterly Progress and Status Report. Matching the rule parameters of PHRASE ARCH to performances of Träumerei : a preliminary study

From Score to Performance: A Tutorial to Rubato Software Part I: Metro- and MeloRubette Part II: PerformanceRubette

Psychoacoustic Evaluation of Fan Noise

A prototype system for rule-based expressive modifications of audio recordings

Enhancing Music Maps

Using the new psychoacoustic tonality analyses Tonality (Hearing Model) 1

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE

Analytic Comparison of Audio Feature Sets using Self-Organising Maps

ESP: Expression Synthesis Project

10 Visualization of Tonal Content in the Symbolic and Audio Domains

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm


Visual and Aural: Visualization of Harmony in Music with Colour. Bojan Klemenc, Peter Ciuha, Lovro Šubelj and Marko Bajec

ATOMIC NOTATION AND MELODIC SIMILARITY

ADVANCED STUDY GUIDE

A QUANTIFICATION OF THE RHYTHMIC QUALITIES OF SALIENCE AND KINESIS

Study on the Sound Quality Objective Evaluation of High Speed Train's. Door Closing Sound

Music NORTH WARREN CLUSTER GRADE K - 3

Table 1 Pairs of sound samples used in this study Group1 Group2 Group1 Group2 Sound 2. Sound 2. Pair

Music. Curriculum Glance Cards

Widmer et al.: YQX Plays Chopin 12/03/2012. Contents. IntroducAon Expressive Music Performance How YQX Works Results

Quarterly Progress and Status Report. Replicability and accuracy of pitch patterns in professional singers

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE

Music Representations

Augmentation Matrix: A Music System Derived from the Proportions of the Harmonic Series

On time: the influence of tempo, structure and style on the timing of grace notes in skilled musical performance

ASD JHS CHOIR ADVANCED TERMS & SYMBOLS ADVANCED STUDY GUIDE Level 1 Be Able To Hear And Sing:

THE SOUND OF SADNESS: THE EFFECT OF PERFORMERS EMOTIONS ON AUDIENCE RATINGS

Introduction. Figure 1: A training example and a new problem.

La Salle University. I. Listening Answer the following questions about the various works we have listened to in the course so far.

Music Theory: A Very Brief Introduction

Computational Models of Expressive Music Performance: The State of the Art

Quarterly Progress and Status Report. Musicians and nonmusicians sensitivity to differences in music performance

Mensuration of a kiss The drawings of Jorinde Voigt

Synthetic and Pseudo-Synthetic Music Performances: An Evaluation

Unit 2. WoK 1 - Perception

An Integrated Music Chromaticism Model

Spectral Sounds Summary

Years 3 and 4 standard elaborations Australian Curriculum: Music

Psychoacoustics. lecturer:

Authors: Kasper Marklund, Anders Friberg, Sofia Dahl, KTH, Carlo Drioli, GEM, Erik Lindström, UUP Last update: November 28, 2002

Analysing Room Impulse Responses with Psychoacoustical Algorithms: A Preliminary Study

Colour-influences on loudness judgements

Measurement of overtone frequencies of a toy piano and perception of its pitch

A PRELIMINARY COMPUTATIONAL MODEL OF IMMANENT ACCENT SALIENCE IN TONAL MUSIC

MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC

Construction of a harmonic phrase

The Development of a Synthetic Colour Test Image for Subjective and Objective Quality Assessment of Digital Codecs

2013 Music Style and Composition GA 3: Aural and written examination

A Beat Tracking System for Audio Signals

Composition I Music theory

LOUDNESS EFFECT OF THE DIFFERENT TONES ON THE TIMBRE SUBJECTIVE PERCEPTION EXPERIMENT OF ERHU

Chapter Five: The Elements of Music

Unobtrusive practice tools for pianists

Music Fundamentals. All the Technical Stuff

MANOR ROAD PRIMARY SCHOOL

Colour Reproduction Performance of JPEG and JPEG2000 Codecs

Sound design strategy for enhancing subjective preference of EV interior sound

Modeling memory for melodies

Expressive information

Quarterly Progress and Status Report. An attempt to predict the masking effect of vowel spectra

Music 175: Pitch II. Tamara Smyth, Department of Music, University of California, San Diego (UCSD) June 2, 2015

Rhona Hellman and the Munich School of Psychoacoustics

The quality of potato chip sounds and crispness impression

Perceiving temporal regularity in music

MPEG-7 AUDIO SPECTRUM BASIS AS A SIGNATURE OF VIOLIN SOUND

HST 725 Music Perception & Cognition Assignment #1 =================================================================

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)

Goebl, Pampalk, Widmer: Exploring Expressive Performance Trajectories. Werner Goebl, Elias Pampalk and Gerhard Widmer (2004) Introduction

This guide gives details of the effects available on the FX selection DMX channels 15 and 17 in the MAC Aura.

Smooth Rhythms as Probes of Entrainment. Music Perception 10 (1993): ABSTRACT

SUBJECT VISION AND DRIVERS

A Case Based Approach to the Generation of Musical Expression

New Mexico. Content ARTS EDUCATION. Standards, Benchmarks, and. Performance GRADES Standards

TO HONOR STEVENS AND REPEAL HIS LAW (FOR THE AUDITORY STSTEM)

Proceedings of Meetings on Acoustics

Music Performance Panel: NICI / MMM Position Statement

The purpose of this essay is to impart a basic vocabulary that you and your fellow

FC Cincinnati Stadium Environmental Noise Model

Determination of Sound Quality of Refrigerant Compressors

Zooming into saxophone performance: Tongue and finger coordination

A Meta-Theoretical Basis for Design Theory. Dr. Terence Love We-B Centre School of Management Information Systems Edith Cowan University

Common assumptions in color characterization of projectors

Frontiers of Optoelectronics Instruction for Authors

Instance and System: a Figure and its 2 18 Variations

Sound visualization through a swarm of fireflies

Subjective Emotional Responses to Musical Structure, Expression and Timbre Features: A Synthetic Approach

PRESCOTT UNIFIED SCHOOL DISTRICT District Instructional Guide January 2016

Transcription:

REALTIME ANALYSIS OF DYNAMIC SHAPING Jörg Langner Humboldt University of Berlin Musikwissenschaftliches Seminar Unter den Linden 6, D-10099 Berlin, Germany Phone: +49-(0)30-20932065 Fax: +49-(0)30-20932183 E-mail: jllangner@aol.com Reinhard Kopiez (Music Conservatoire Hannover) Christian Stoffel Martin Wilz (University of Cologne) Paper given at the 6th International Conference on Music Perception and Cognition, Keele, England, 5.-10. August 2000. Do not cite without permission. Introduction Compared to research on timing, the field of musical dynamics is a neglected parameter in performance research. For example, despite a focus on musical rhythm, timing and performance, the latest edition of Deutsch s (1999) survey over a whole discipline, The psychology of music does not even contain a sub-chapter on dynamics. Other research literature is widely distributed. We cannot say whether this situation is due to a lack of interest but would rather assume that it is due to a lack of adequate research methods, which prevents a deeper understanding of the nature of dynamics. To sum up we can formulate some important research topics: Although the history of performance practice shows the increasingly important role of dynamic shaping for conveying expression in music, we know only very little about the relationship between musical form and musical dynamics. Based on musical experience we can say that e.g. a Bruckner-Symphony is unimaginable without the form-generating force of dynamics. The relationship between timing and dynamics and its importance for musical perception is unclear: they either exist in a hierarchical relationship (e.g. with a dominance of timing over

2 dynamics), or are of equal importance. In the first case we could assume that dynamics have only a small effect on a global level and a greater effect on a more local level this contradicts musical experience; in the second case the problem of redundancy is evoked: why should we take care of a second expressive parameter (dynamics), if expression is already mediated by the domain of timing? Which methods of analysis of dynamics and of obvious presentation of results are available? This question concerns the field of performance analysis as well as that of educational application. Only a highly obvious and easy manageable analysis of dynamics will be accepted by the majority of instrumental teachers. This means a special need for realtime methods of analysis and presentation. Some answers to the above mentioned questions can be found in the literature: as one of the founding authors, Riemann (1884) published a treatise on musical phrasing which concentrated exclusively on the role of dynamics and rubato. His simple assumption was that dynamics and rubato are coupled, and that the development of an eight bar long musical phrase is shaped simultaneously by a crescendo and an accelerando until the climax of the phrase. This more global perspective of dynamics seems to be more plausible. Huron s (1991; 1992) perception theory of ramp archetypes fits well into this perspective. Huron (1991) calculated a mean length of 4.3 bars for crescendi and of 5.8 bars for decrescendi using a sample of 537 works or movements of 14 composers with a total of 85476 bars. The same idea of a simple coupling of the two parameters can be found in Todd s (1990) model of musical expression: the faster the louder, the slower the softer (p. 3540). We don t believe in such a simple, rule-based relationship of parameters and assume that this perspective meets only a part of musical reality. This approach allows only a very limited view. However, as Friberg (1991) tried to show, it is possible to generate decent synthesized performances by use of such a rule-based system. We would like to try a different approach: referring to the Theory of oscillating systems (TOS) by Langner (1999) we hypothesize that the timing and dynamics of a performance are shaped on multiple levels, including local and global layers. Local layers concern the dynamic shaping for example from note to note or from measure to measure; global layers on the other hand are connected to the relationship of dynamics between larger sections or subsections of a piece of

3 music. (Such multi-level structure can also be found in other domains of performance; for the analysis of timing see Langner & Kopiez 1995). An adequate method of performance analysis should preserve the full information contained in the performance data (without any reduction), and as Langner (1999, pp. 153-155; 1997) demonstrated, this multi-dimensional character of dynamic shaping is in strong concordance with musical perception and experience. If possible, analysis of dynamics should be done in realtime. Our demonstration uses recently developed software, and shows by use of graphical output in the form of so-called Dynagrams some possibilities for application. Method The procedure will now be described, based on the assumption that the piece of music to be analysed is available as complete audio file. Following that is an outline of the modifications made when the analysis works through the music step by step in realtime. (a) Non-realtime procedure The starting point for the procedure is the digitized audio signal. From this (step 1) the loudness curve of the piece is calculated this means that at regular intervals in the piece a loudness value is allocated in Sone units. For this purpose a particular computer programme was used which was developed by Bernhard Feiten & Markus Spitzer (Technological University of Berlin) on commission from the Hochschule für Musik und Theater Hannover (see also Langner, Kopiez & Feiten 1998, pp. 18 20). This programme is based on Zwicker s Model of loudness (Zwicker & Fastl 1990, pp. 197 214), which guarantees close proximity to the perceived loudness, and produces superior quality to the simple use of decibel values. This loudness curve (step 2) was then subjected to multiple smoothing out processes of varied strength. This smoothing out was achieved through the inclusion, when measuring at a particular point of time, of not only the loudness value at exactly this point but also the surrounding values, thus creating a mean measurement. This surround of the point is also termed the window for calculation. The wider the calculation window, the stronger the smoothing out effect. If one were to take in an extreme case the length of the entire piece of music on a window, there would be only one

4 mean value for the whole piece, and the smoothing out would therefore be at its maximum. In contrast, a very narrow window would produce a smoothing out very similar to the original loudness curve. There are many interim steps between these extremes. Concrete graphic examples are to be found in Langner (1997). A strong smoothing out of the loudness curve only represents differences (that is deviation from a horizontal line) when the appropriate wide-ranging dynamic shaping exists; precisely these varied strength smoothing out cases allow multi-layer analysis as mentioned in the introduction. Our procedure uses a wide spectrum of various window sizes. The exact range can be selected within the programme. A frequently applied setting contains 37 different windows, sized between 0.25 and 128 seconds (in logarithmic steps). Finally (step 3) the gradients in every smoothing out curve are calculated at each point in time. The procedure is similar to that of the first derivative in mathematics and physics; here though the gradients according to each separate window size for each smoothing out curve were calculated. The effect is that the strongly smoothed out curves (which generally show much weaker fluctuation) can obtain just as steep gradients as the more weakly smoothed out. There is a sense of equal treatment under the contrasting smoothing realignments. Following from this change to gradients, the analytical perspective changes from focusing upon loudness to focusing upon loudness changes that is from loud/soft to crescendo/decrescendo. (This change in perspective stood the test of previous analyses; the final decision as to whether loudness or loudness changes are actually represented has not yet been reached). The output (step 4) is produced in graph form showing the gradients referred to, in a so-called Dynagram (see fig. 1 and fig. 2). The time axis shows the horizontal axis; the window size is represented on the vertical axis. Red colouring signifies crescendo (the more intense the red the stronger the crescendo); green colouring then shows decrescendo (the more intense the green, the stronger the decrescendo). (b) Realtime procedure During the realtime version, the audio signal is recorded through a microphone link to the computer. The procedure described above is carried out in the same way. In order to create the Dynagram it must be considered that in calculating the smoothing out, a certain surrounding area of a point in

5 time has constantly to be included. One must almost look into the future for the weaker smoothing out to a lesser extent, for the stronger one to a greater extent. (Such looking into the future can be considered with retrospective re-interpretation of what is heard and is from this point of view plausible). As far as the procedure is concerned, the following is relevant: the Dynagram can only be calculated retrospectively with inconsequential delay for the small window sizes, but with considerable delay for the large window sizes. The data points of a realtime Dynagram thus appear on the screen not in vertical axis form, but approximately as a diagonal line. Results Figures 1 and 2 show the Dynagram of both a professional and a non-professional performance of Erik Satie s Gymnopédie No.1. The generally more intense loudness pattern from the professional pianist is noticeable. Particularly remarkable is also the greater intensity in the larger window area; the shading in this area reveals correspondence of loudness organization to formal structure. The composition namely is made of two identical parts, each of which again consists of two almost equal sized sections. WINDOW SIZE s 1. 2. 4. 8. 16. 32. 64. 128. 0 50 100 150 TIME s Fig. 1: Dynagram of a professional performance of Erik Satie s Gymnopédie No.1. The different colours have the following meaning: intense red = strong crescendo, pale red = weak crescendo, white = constant loudness, pale green = weak decrescendo, intense green = strong decrescendo. The dynamic shaping reflects clearly the formal structure of the composition (the formal breaks are marked in the upper horizontal frame).

6 WINDOW SIZE s 1. 2. 4. 8. 16. 32. 64. 128. 0 50 100 150 200 TIME Fig. 2: Dynagram of a non-professional performance of Erik Satie s Gymnopédie No.1. The dynamic shaping is not as strong as in the professional performance and reflects the formal structure of the composition less clearly. s (To demonstrate these structural segments, the start of each section is marked in the upper horizontal frame of the Dynagram.The starting points of the two main parts are black; the less strong formal devisions are in contrast coloured grey). It is clear from the Dynagram of the professional player s version that each of the four formal sections is covered from a red-green pair in the window size area from 8 to 16 s, just as the two halves in the vicinity of 32 s. Clearly the professional pianist is capable of pointedly marking the structure of the composition through control of loudness (in other words with arches from crescendi and decrescendi); the nonprofessional meanwhile achieves this only to a smaller extent. Further examples fitting to the procedure showed the Dynagrams to be a way of making visible in particular more extensive loudness shaping of a performance. The analysis of a recording of a movement a Bruckner Symphony (conducted by Günter Wand) for instance revealed a build-up spanning some 20 minutes from start to the final climax of the piece. Discussion and perspectives of application The analyses carried out up to this point give reason to believe that in the Dynagrams important attributes and qualitative characteristics can be registered. In particular the absence or presence of wide-ranging loudness shaping is clearly visualized. This can be seen in the lower part of the Dynagram, the area for larger window sizes. These characteristics of the procedure render it a useful

7 tool for performance research. It brings into play a new analytical tool with which the issues briefly raised in the introduction those of the role and significance of dynamics can be tackled in a new light. The option of carrying out the analysis in realtime, simply with a microphone and standard computer, opens up a whole new perspective for application in instrumental lessions and musicians practice time: the Dynagram is displayed on a monitor while the pupil plays, and facilitates both pupil and teacher instant analysis of strength and weakness. The Dynagrams can be saved and printed out, and thus kept for comparison at a later date; a pupil s progress can in this way be documented. The pupil additionally has a mean to self-analysis. Further possibilities for the analysis of music will be enabled by realtime transfer of several additional procedures. See here the contribution from Langner Rhythm, Periodicity and Oscillation in the accompanying volume, in which an online-compatible process encompassing rhythmical qualities is outlined. Both papers are part of a wider research project which reaches out over all areas of music, and in particular covers analytical processes for harmony and melody (Langner 1999, pp. 156 157). References Deutsch, D. (Ed.) (1999). The psychology of music. 2nd edition. New York: Academic Press. Friberg, A. (1991). Generative rules for music performance: A formal description of a rule system. Computer Music Journal, 15(2), 56 71. Huron, D. (1991). The ramp archetype: A score-based study on 14 piano composers. Psychology of Music, 19, 33 45. Huron, D. (1992). The ramp archetype and the maintenance of passive auditory attention. Music Perception, 10(1), 83 92. Langner, J. (1997). Multidimensional dynamic shaping. In A. Gabrielsson (Ed.), Proceedings of the third triennial ESCOM conference, Uppsala, Sweden, 7 12 June, 713 718. Langner, J. (1999). Musikalischer Rhythmus und Oszillation. Eine theoretische und empirische Erkundung. [Musical rhythm and oscillation. A theoretical and empirical investigation]. Dissertation, Hochschule für Musik und Theater Hannover. (A printed version of this dissertation, including a comprehensive abstract in English, will be published by Peter Lang Verlag, Frankfurt/Main in 2000 or 2001).

8 Langner, J. & Kopiez, R. (1995). Oscillations triggered by Schumann s Träumerei : Towards a new method of performance analysis based on a Theory of oscillating systems (TOS). In A. Friberg & J. Sundberg (Eds.), Proceedings of the KTH Symposium on Grammars for music performance, Stockholm, May 27, 45 58. Langner, J., Kopiez, R. & Feiten, B. (1998). Perception and representation of multiple tempo hierarchies in musical performance and composition. In R. Kopiez & W. Auhagen (Eds.), Controlling creative processes in music (pp. 13 35). Frankfurt a.m.: P.Lang. Riemann, H. (1884). Musikalische Dynamik und Agogik. [Musical dynamics and agogics]. Hamburg: Rather. Todd, N.P. McAngus (1990). The dynamics of dynamics: A model of musical expression. Journal of the Acoustical Society of America, 91(6), 3540 3550. Zwicker, E. & Fastl, H. (1990). Psychoacoustics. Berlin: Springer.