Quarterly Progress and Status Report. A singer s expression of emotions in sung performance

Similar documents
Quarterly Progress and Status Report. Replicability and accuracy of pitch patterns in professional singers

Quarterly Progress and Status Report. Formant frequency tuning in singing

Quarterly Progress and Status Report. Musicians and nonmusicians sensitivity to differences in music performance

Quarterly Progress and Status Report. Perception of just noticeable time displacement of a tone presented in a metrical sequence at different tempos

Making music with voice. Distinguished lecture, CIRMMT Jan 2009, Copyright Johan Sundberg

MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC

Director Musices: The KTH Performance Rules System

Loudness and Pitch of Kunqu Opera 1 Li Dong, Johan Sundberg and Jiangping Kong Abstract Equivalent sound level (Leq), sound pressure level (SPL) and f

Quarterly Progress and Status Report. Matching the rule parameters of PHRASE ARCH to performances of Träumerei : a preliminary study

Quarterly Progress and Status Report. An attempt to predict the masking effect of vowel spectra

A Computational Model for Discriminating Music Performers

Musicians Adjustment of Performance to Room Acoustics, Part III: Understanding the Variations in Musical Expressions

Music Representations

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 FORMANT FREQUENCY ADJUSTMENT IN BARBERSHOP QUARTET SINGING

Artificial Social Composition: A Multi-Agent System for Composing Music Performances by Emotional Communication

Rhythm and Melody Aspects of Language and Music

I. LISTENING. For most people, sound is background only. To the sound designer/producer, sound is everything.!tc 243 2

A prototype system for rule-based expressive modifications of audio recordings

AUD 6306 Speech Science

Version 5: August Requires performance/aural assessment. S1C1-102 Adjusting and matching pitches. Requires performance/aural assessment

Quarterly Progress and Status Report. Intonation preferences for major thirds with non-beating ensemble sounds

Quarterly Progress and Status Report. X-ray study of articulation and formant frequencies in two female singers

Florida Performing Fine Arts Assessment Item Specifications for Benchmarks in Course: M/J Chorus 3

1. Introduction NCMMSC2009

THE SOUND OF SADNESS: THE EFFECT OF PERFORMERS EMOTIONS ON AUDIENCE RATINGS

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

The Tone Height of Multiharmonic Sounds. Introduction

We realize that this is really small, if we consider that the atmospheric pressure 2 is

Article: Phrasing...Speaking in Musical Sentences.Blue Grass News, official Journal of the KY Music Educators Association

RHYTHM. Simple Meters; The Beat and Its Division into Two Parts

Importance of Note-Level Control in Automatic Music Performance

Kent Academic Repository

Standard 1: Singing, alone and with others, a varied repertoire of music

SAMPLE ASSESSMENT TASKS MUSIC GENERAL YEAR 12

Instrumental Music II. Fine Arts Curriculum Framework

UNIVERSITY OF DUBLIN TRINITY COLLEGE

ASD JHS CHOIR ADVANCED TERMS & SYMBOLS ADVANCED STUDY GUIDE Level 1 Be Able To Hear And Sing:

Electronic Musicological Review

Amplitude and Loudness 1

Music Curriculum Glossary

6.5 Percussion scalograms and musical rhythm

Marion BANDS STUDENT RESOURCE BOOK

Timbre blending of wind instruments: acoustics and perception

Assessment may include recording to be evaluated by students, teachers, and/or administrators in addition to live performance evaluation.

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

On the contextual appropriateness of performance rules

œ Æ œ. œ - œ > œ^ ? b b2 4 œ œ œ œ Section 1.9 D Y N A M I C S, A R T I C U L A T I O N S, S L U R S, T E M P O M A R K I N G S

increase by 6 db each if the distance between them is halved. Likewise, vowels with a high first formant, such as /a/, or a high second formant, such

A PRELIMINARY COMPUTATIONAL MODEL OF IMMANENT ACCENT SALIENCE IN TONAL MUSIC

Western Classical Tradition. Music for voices: operas and songs

Quantitative Emotion in the Avett Brother s I and Love and You. has been around since the prehistoric eras of our world. Since its creation, it has

Proceedings of Meetings on Acoustics

Pitch-Synchronous Spectrogram: Principles and Applications

Quarterly Progress and Status Report. Acoustic analysis of three male voices of different quality

Quarterly Progress and Status Report. Is the musical retard an allusion to physical motion?

Largo Adagio Andante Moderato Allegro Presto Beats per minute

Quarterly Progress and Status Report. Voice source characteristics in different registers in classically trained female musical theatre singers

Laboratory Assignment 3. Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB

JOURNAL OF BUILDING ACOUSTICS. Volume 20 Number

Temporal summation of loudness as a function of frequency and temporal pattern

EVTA SESSION HELSINKI JUNE 06 10, 2012

Burton s Sonatina for Flute & Piano: A Practice Guide

Sofia Dahl Cognitive and Systematic Musicology Lab, School of Music. Looking at movement gesture Examples from drumming and percussion Sofia Dahl

Quarterly Progress and Status Report

Connecticut State Department of Education Music Standards Middle School Grades 6-8

Instrumental Music III. Fine Arts Curriculum Framework. Revised 2008

K-12 Performing Arts - Music Standards Lincoln Community School Sources: ArtsEdge - National Standards for Arts Education

La Salle University. I. Listening Answer the following questions about the various works we have listened to in the course so far.

Computer Coordination With Popular Music: A New Research Agenda 1

Topic 10. Multi-pitch Analysis

Subjective evaluation of common singing skills using the rank ordering method

Smooth Rhythms as Probes of Entrainment. Music Perception 10 (1993): ABSTRACT

Structural Communication

Affective Sound Synthesis: Considerations in Designing Emotionally Engaging Timbres for Computer Music

PSYCHOACOUSTICS & THE GRAMMAR OF AUDIO (By Steve Donofrio NATF)

Efficient Computer-Aided Pitch Track and Note Estimation for Scientific Applications. Matthias Mauch Chris Cannam György Fazekas

Measurement of overtone frequencies of a toy piano and perception of its pitch

Introduction to Vocal Music: The development of Secular Song

Norman Public Schools MUSIC ASSESSMENT GUIDE FOR GRADE 8

Quarterly Progress and Status Report. Violin timbre and the picket fence

How We Sing: The Science Behind Our Musical Voice. Music has been an important part of culture throughout our history, and vocal

Instruments can often be played at great length with little consideration for tiring.

The purpose of this essay is to impart a basic vocabulary that you and your fellow

Acoustic Analysis of Beethoven Piano Sonata Op.110. Yan-bing DING and Qiu-hua HUANG

Tempo and Beat Analysis

Grade Level 5-12 Subject Area: Vocal and Instrumental Music

Instrumental Music II. Fine Arts Curriculum Framework. Revised 2008

ST. JOHN S EVANGELICAL LUTHERAN SCHOOL Curriculum in Music. Ephesians 5:19-20

Semi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis

Real-Time Control of Music Performance

Saint Patrick High School

Instrumental Performance Band 7. Fine Arts Curriculum Framework

SAMPLE ASSESSMENT TASKS MUSIC CONTEMPORARY ATAR YEAR 12

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng

Analysis of the effects of signal distance on spectrograms

Musical Bits And Pieces For Non-Musicians

HIGHLANDS CHOIR SEMESTER EXAM REVIEW. Whole Half Quarter Eighth Sixteenth. Whole Half Quarter Eighth Sixteenth

Digital audio and computer music. COS 116, Spring 2012 Guest lecture: Rebecca Fiebrink

Step by Step: Standards-Based Assessment in General Music

WHAT IS BARBERSHOP. Life Changing Music By Denise Fly and Jane Schlinke

Transcription:

Dept. for Speech, Music and Hearing Quarterly Progress and Status Report A singer s expression of emotions in sung performance Sundberg, J. and warsson, J. and Hagegård, H. journal: STL-QPSR volume: 35 number: 2-3 year: 1994 pages: 081-092 http://www.speech.kth.se/qpsr

STL-QPSR 2-31] 994 A singer's expression of emotions in sung performance Johan Sundberg, Jenny warsson & Hiikan Hagegiird Abstract A professional singer (co-author HH) sang a set of music excerpts fiom the Lieder and opera repertoires in two deliberately contrasting fashions, () as in a concert situation and (2) in an emotionally neutral way, as if the singer did not engage emotionally in the poem. The two renderings were evaluated by expert panels with respect to degree of emotional involvement and with regard to the type of emotional ambience characterizing the renderings. On the basis of these tests certain excerpts were selected for acoustic analysis. Surprisingly, no clear differences were found with regard to departures Porn nominal tone duration, thus suggesting that such departures are related to other aspects of espressivity than emotionality, such as the marking of musical structure. Agitated renderings seemed characterized by louder voice, faster tempo, and higher rate of loudness variation than the nonagitated renderings. Emotional espressivity seemed further associated with great FO modulations during consonants and often also with a lowering of vowel formant frequencies. ntroduction Emotional expression is an essential aspect of artistic singing; listening to a good performance can be as exiting as it is agonising to listen to a performance with a neutral or an inappropriate expression. Emotional expressivity also constitutes a highly relevant ingredient in speech communication. Yet, the expressive aspects of speech and particularly singing has attracted relatively little attention in research (see Scherer 1994 for a review). By and large, the results have shown that pitch characteristics have a very powerful effect on perceived emotionality in speech. Studies of emotional speech do not explain how emotional expressivity is induced in sung performances, where the pitch parameter is restricted by the score. Some studies have been devoted to expressivity in singing. Kotlyar & Morozov (1976) asked eleven singers to sing a set of music excerpts imparting five different "emotional meanings" on the same music excerpts: joy, sorrow, anger, fear, and neutral. They measured syllable duration, micropauses between syllables, mean SPL, tone rise and decay time and found characteristic combinations in these measures for the various emotions. They concluded that temporal and dynamic characteristics are extremely important to vocal expression of emotions. Scherer & Oschinsky (1977) studied the relevance of amplitude modulation, tempo, harmonic content, tonality, rhythm as well as the level, contour and variation of pitch on emotion attributions to sentence-like sound sequences and musical melodies. The most important cues were found in the tempo and in the overtone content. n this investigation we attempted to elucidate how emotions are expressed in sung performance by comparing deliberately neutral and expressive versions of the same

STL-QPSR 2-311 994 music examples. n this report some major trends will be described, leaving detailed analysis to future studies. Experiment An international baritone soloist, highly experienced in both opera and Lieder singing (co-author HH) agreed to perform a set of music examples in two different ways, with an expression which he perceived as appropriate and in a manner as void of emotional expression as possible. For obvious reasons, he experienced the latter task as somewhat exotic, lying beyond his musical expertise, though still possible. The recordings were made on a DAT recorder in a large studio hall at the Hagegirden Music Center, Gunnarsbyn, Sweden. A small microphone was fastened to the singer's glasses so as to ensure a constant distance. No piano accompaniment was used. The exerpts are listed in Table 1. They were selected so as to potentially represent a variety of different emotional ambiances, such as fear, joy, sadness, love, hate, security, tenderness. The singer was not informed about these emotional interpretations of the excerpts. Table 1. List of music excerpts Composer Folktune Folktune G. Mahler F. Mendelsohn F. Schubert F. Schubert F. Schubert F. Schubert F. Schubert F. Schubert R. Schumann R. Schumann R. Schumann R. Strauss G. Verdi Song title "Byssan lull", verse 1 "Vi gi over daggstankta berg", bars 1-8 "Lieder eines fahrenden Gesellen," song # 3, bars 5-1 1 "Paulus", Aria # 18, bars 5-13 "Erlkonig", bars 4 1-50 "Erlkonig", bars 72-79 "Du bist die Ruh", bars 8-1 5 "Wanderers Nachtlied", bars 3-14 "Das Wandern", verse 1, bars 12-20 "Nahe des Geliebten", bars 3-8 "Liederkreis" V "Schone Fremde", bars 16-24 V "n der Fremde", bars 25-37 X1 "Fruhlingsnacht", bars 18-26 "Dichterliebe" V, bars 3 1-42 "Dichterliebe" V, bars 12-18 "Zueignung", bars 2 1-29 "Falstaff', Ford's Monologue, bars 24-3 1 Excerpt Name By ssan Daggstankta* Gluhend Messer* Mendelsohn Siehst Vater Du* Mein Vater* Du bist die Ruh Wanderers Nachtlied Das Wandern Nahe des Geliebten Es funkeln * Als miisste* Und der Mond* Es schweben Wie Du auch* Zueignung* Ford's Monlogue* Perceptual evaluations Two listening tests were carried out to evaluate the emotional quality of the performances recorded. n test 1 the excerpts were arranged in random order on a tape.

STL-QPSR 2-31] 994 Table 2. Overview of results of acoustic analysis of the differences between the expressive and neutral versions of the excerpts. For each parameter, the values observed for the neutral versions are given together with the difference in percent @om this value in the expressive version. The tempo is given as the number of the excerpt's shortest note value (SNV) per second, the loudness in terms of the sound level (SL) averaged over the entire excerpt, and the short-term variability of loudness (Mean dsl/dt) in terms of the time derivative of sound level averaged over the entire excerpt. The leftmost column shows the predominant emotions of the expressive versions according to Listening test 2. The excerpts are ordered according to the loudness variability value. Excerpts marked with a star (*) possess an agitated or lively musical character. Excerpt Tempo Mean SL Mean dsl/dt Neutr Expr Neutr Expr Neutr Expr Predominant emotion SNV/s DgfS(%) (db) Dgff (db/ms) Diff (<3 of 5 votes) (%) (%) Ford's Monologue * 6.1-4.9 75.9-0.7 0.014 93 Hateful (4), Angry (3) Mein Vater * 5.7 0 70.7 2.5 0.022 73 Hateful (4), Angry (3) ch hub ein * 5.8-4.3 76.9 2.6 0.038 39 Secure (3), Happy (3), Sad (3) Wie du auch * 8.7-6.9 69.6 3.2 0.027 30 Angry (4), Hateful (3) Und der Mond * 4.0 0 59.3 2.7 0.032 28 Happy (4), Loving (4), Secure (: Zueignung * 5.3-16.9 71.7-0.2 0.016 19 Secure (5), Loving (4) Daggstankta * 7.9 3.8 55.8 1.7 0.124 2 Happy (5), Secure (5) Nahe des geliebten 6.8-26.5 75.8-1.2 0.015-7 Secure (5), Loving (5) Mendelsohn 4.5-20.0 77.6 0.8 0.014-14 Secure (5), Loving (5), Sad (3) Es schweben 8.0-14.0 78.5-1.2 0.018-22 Loving (5), Secure (4) Wanderers Nachtlied 2.6-23.1 73.2-1.3 0.007-29 Loving (4), Secure (3) Du bist die Ruh 2.5-20.0 76.8-4.5 0.014-50 Secure (4), Loving (4) Vowel-to-vowel durations were measured from spectrograms and compared with the score's nominal durations, defined from the data on the mean tempo. The departures from the nominal durations tended to form patterns that were mostly similar for the neutral and expressive durations, a finding nicely replicating results from an experiment recently reported by Drake & Palmer (1993). These patterns seemed to reflect two principles, lengthening of short notes, and marking of musical phrase structure. This marking adhered to a principle previously described by Todd (1985); a phrase is typically marked by starting somewhat slowly, then accelerating, and again decelerating towards the end of the phrase, so that a parabolic tempo curve is produced, see Figure la. Applying these two principles to the nominal durations of two of the excerpts yielded an astonishingly good match of the singer's performances in both the neutral and the expressive rendering, as can be seen in Figure lb and c. Thus, independent of emotional expressivity, the singer's marking of phrases can be rather accurately described using the same coefficients. Greater lengthenings of short notes were sometimes observed in the expressive versions. Somewhat surprisingly, no examples were found of lengthening of notes presenting particularly important words in the poem.

STL-QPSR 2-311 994 0.50 NOTE POSTON 0 20 40 60 80 100 120 140 160 180 200 NOTE POSTON Figure. Upper panel: Parabola pattern used for increasing the durations (DR) of notes in a subphrase and a phrase (left and right half of the graph), see Todd (1 985). Middle and lower panels: Deviations @om nominal durations (DR) observed for the various syllables in excerpts "Es schweben" (middle) and "Mendelsohn" (lower). Open circles and Jilled diamonds refer to the neutral and expressive versions, and the open diamonds and dotted lines refer to model generated data, using the parabola pattern shown in the upper panel and the principle of lengthening of short notes. 0.50 0 10 20 30 40 50 60 70 80 90 NOTE POSTON

STL-QPSR 2-311 994 Overall sound level was determined as the mean sound level over the entire excerpt as measured by the SWELL program (Ternstrom, 1991). The means are given in Table 2. Note that this mean increases not only when vocal loudness is high, but also when a song is performed legato, with few and short intertone pauses. The latter effect is reflected in the low value obtained for Daggstankta, which the singer sang in a staccato-like manner. Expressive versions characterised as secure and loving showed lower means than in neutral, as they were sung softer. The opposite effect was noted in most versions characterised as hateful or happy. Comparison of the long term variations of sound level between the neutral and expressive renderings revealed no clear differences, thus suggesting that crescendos and diminuendos were not used for the purpose of imparting emotional expression. n excerpt "Und der Mond", however, the singer made a crescendo at the end of the piece only in the emotive version. This probably reflected the content of the poem; in this part of the song the text presents the essence of the poem. Thus, in this case the singer seemed to use long term variation of sound level for the purpose of enhancing aspects of the text. The short-term variability ofsound level was evaluated by computing the time derivative of the sound-level curve, smoothed with a 20 Hz low-pass filter. The means of this derivative are listed in Table 2, where the excerpts were ordered according to this mean. The difference in this parameter was considerable, ranging from +93% to -50%. ncreases were found in the expressive versions characterised as hateful or happy while decreases were found for most secure and loving versions. ndeed, all agitated excerpts were found at the upper part of the table while all excerpts portraying a nonagited emotion are found in the lower part. The source of a great variability was mostly an emphatic pronunciation of consonants causing a great amplitude modulation in the expressive versions. Figure 2 offers a typical example, where in the expressive version examples of a great amplitude modulation can be observed, e.g., during the consonant [r] in "strahlst". The above observations show that the short-term variability in sound level is an important means of emotional expression in singing. FO-patterns were analysed by means of the double-peak-picking strategy available in the SWELL program. n agitated excerpts, the extent of the vibrato undulations was mostly considerably greater in the expressive than in the neutral versions, while in the nonagitated excerpts the two versions were more similar. Typical examples are shown in Figure 3. This suggests that the vibrato extent is used for expressive purposes in both singing and instrument playing (Gabrielsson, 1993). Local departures from the nominal FO values were observed in many expressive versions. These departures often smoothed pitch jumps by means of short-term portamenti, i.e., glissando-like pitch changes, but sometimes FO simply made a deep dip between tones so that the new tone onset started from a very low value. Such portamenti frequently occurred on voiced consonants, apparently for the reason of emphasis.

STL-QPSR 2-311 994 ri wie d u au ch sl r a hl sl - TlME (s) TlME (s) Figure 2. Comparison of audio signal, sound level and FO in the neutral (upper panel) and expressive (lower panel) versions of the first tones in excerpt "Wie du auch strahlst". Note the amlitude modulation in the Jirst three syllables and during the [r] sound in the expressive version.

FUNDAMENTAL FREQUENCY (Hz) A & & g- g E g g g g g g g FUNDAMENTAL FREQUENCY (Hz) N N N N W W ~~~~~~~~k?~~

STL-QPSR 2-311 994 According to listening experiment 2 the expressive version collected all subjects' votes for one or more specific emotions in five excerpts (see Table 2). Using spectrographic analysis (Sonagraph DSP 5500) the four lowest formant frequencies were determined for most vowels in these excerpts. Figure 4 compares the data for the neutral and expressive versions. Occasionally differences were observed for F 1 and F2 and in all such cases the value measured in the emotive rendering was lower than that measured in the neutral rendering. For F 1 and F2 the differences occurred in the ranges of 0.5-0.8 khz and 1-2 khz, respectively. Hence, the lower values in the emotive versions reflected a darker vowel quality. F3 and in particular F4 frequently showed clearly lower values in the emotive renderings. This would mirror a "coloring" of the voice timbre which would correspond to a more enhanced singer's formant. The result suggests that voice timbre can be used for the purpose of emotional expression in singinp F11F2, F38cF4 Fn N NEUTRAL (Hz) Figure 4. Comparisons of of F, F2, F3 and F4 of vowels occurring in the neutral and expressive versions offive of the excerpts. Symbols refer to formant numbers. Discussion n this investigation we attempted to use as musically realistic situations as possible, thus taking advantage of the singer's special competence. We avoided asking him to perform the same excerpt with a number of different emotions, because the musical competence of a musician would include the ability to portray by acoustical means the

STL-QPSR 2-31 1994 particular emotional ambiance embedded by the composer in the song while dressing a song in an inappropriate ambiance is not necessarily part of a musician's expertise. Our results revealed two major performance trends. One was the marking of the musical structure by means of patterned departures from nominal duration. These departures were rather similar in the neutral and expressive versions. The second trend was to use tempo, hndamental frequency modulation, loudness and rate of loudness variation for expressive purposes. These factors seemed to vary in accordance with the overall musical character of the excerpt. As mentioned before, the excerpts can be grossly divided into two groups, agitated and nonagitated. The tempo tended to be considerably slower in the expressive renderings of the nonagitated excerpts. With regard to loudness, most expressive versions of the agitated excerpts were sung louder while those of the nonagitated versions were sung softer. With regard to rate of loudness variation, expressive versions of all agitated excerpts showed increased rates while those of the nonagitated excerpts showed decreased rates. These observations suggest that the polarity agitatedlnonagitated is a factor of major relevance to expresssion in singing. However, signs of specific emotions may very well hide in small details, such as formant transitions at CV and VC boundaries. Synthesis would offer a powerful tool in revealing the emotional relevance of such details in future research. By and large, our observations agree with findings made by Kotlyar & Morozov (1976). For instance, they found that sad was characterised by long durations of syllables, i.e., a slow tempo, that the "average voice pressure level" was high in anger and also in happy, and that syllable onsets and decays were fast in anger. Furthermore, our results reveal striking parallels between expression in speech and singing. Scherer (1994), summarising previous research of expression of emotions in speech, notes that anger and joy are associated with a high FO variability and also that sadness is associated with a low vocal intensity and a low rate of articulation. Hypothesising that FO variability in speech is translated into vibrato extent in singing, these observations seem to compare well with our findings. The departures from nominal duration showed that short notes were often lengthened in both the neutral and the expressive versions. Previous studies of musical performance of instrumental music has revealed the opposite trend, i. e., that short notes are shortened (Sundberg, 1993), but a detailed analysis did not show one single example of this principle in the singer's expressive versions, not even in the agitated excerpts. Lengthening of short notes may be idiomatic to singing, although singers and instrumentalists mostly perform music together and can hence be expected to adhere to similar performance principles. The relevance of the text to the artist's sung performance is interesting. Palmer and Kelly (1992) analysed syllable durations in sung performances and observed greater lengthenings of nouns than adjectives under certain conditions. We did not find any cases of lengthening of semantically important syllables in our material. t might be rewarding to study such aspects hrther by comparing a singer's sung performance with an actors' reciting of the same text. t would also be interesting to compare expressive

STL-QPSR 2-311 994 performances of songs sung with its text and as a vocalise, i.e., as a song with sustained vowels replacing the text. One single singer was used in the present investigation, so the results do not allow general conclusions. On the other hand, the fact that this singer is successful in an international carreer obviously makes his methods of expression a worthwhile object of scientific study. The singer's success also suggests that his methods of conveying expression in singing are widely accepted. Moreover, our study revealed striking parallels between expression in speech and singing. Still, we believe that some characteristics of his art may be personal. For example, it was interesting that in listening test 2, there was a predominance of loving and secure, while scared and sad were rarely chosen alternatives. This investigation was a preliminary attempt to study emotional expression in singing, and our findings have resulted in suggestions for a number of future studies which could complement the description of emotional expression in singing. Thereby, we think that our method of comparing neutral with expressive versions of the same excerpt is useful. The generality of the findings should be tested in synthesis experiments and by analysing other singers' methods for expressing emotions. Conclusions An agitated versus nonagitated overall character of the song seems to be highly relevant to the expression of emotions in singing. Fast tempo, high vocal loudness and, in particular, high rate of loudness variation were used in agitated excerpts while the opposite was observed in nonagitated excertps. By and large, these characteristics agree with those previously found in speech. Expression seemed further promoted by great fundamental frequency modulations during consonant segments, and sometimes also by a lowering of the vowel formant frequencies. Departures from nominal tone durations, on the other hand, seemed rather independent of emotional expression, apparently depending more on the musical structure. A more detailed analysis, preferably supported by synthesis experiments, is likely to further elucidate the expression of emotions in singing. Acknowledgement The authors gratefully acknowledge the participation of Lars Fryden and Anders Friberg in the analysis of the singer's departures from nominal duration. The research was supported by a grant from the Swedish Research Council for Engineering Sciences. References Drake, C. & Palmer, C. (1993). "Accent structure in music performance", Music Perception 10, pp. 343-378. Gabrielsson, A. (1993). "Emotional expression in music performance". Paper given at the Stockholm Music Acoustics Conference, July 1993.

STL-QPSR 2-3/1994 Kotlyar, G. & Morozov, V. (1976). "Acoustical correlates of the emotional content of vocalised speech", Soviet Physics Acoustics 22, pp. 208-21 1. Palmer, C. & Kelly, M.H. (1992). "Linguistic prosody and musical meter in song", Journal of Memory and Language 3 1, pp. 525-542. Scherer, K. (1994). "Expression of emotions in voice and music". Manuscript submitted for publication. Scherer, K. & Oschinsky, J. (1977). "Cue utilization in emotion attribution from auditory stimuli", Motivation and Emotion, pp. 33 1-346. Sundberg, J. (1993). "How can music be expressive?", Speech Communication 13, pp. 239-253. Ternstrom, S. (1991). "Sound SWELL manual", Soundswell, Solna, Sweden. Todd, N. (1985). "A model of expressive timing in tonal music", Music Perception 3, pp. 33-59.