Electric brain responses reveal gender di erences in music processing

Similar documents
Interaction between Syntax Processing in Language and in Music: An ERP Study

Effects of musical expertise on the early right anterior negativity: An event-related brain potential study

Children Processing Music: Electric Brain Responses Reveal Musical Competence and Gender Differences

Untangling syntactic and sensory processing: An ERP study of music perception

Untangling syntactic and sensory processing: An ERP study of music perception

Overlap of Musical and Linguistic Syntax Processing: Intracranial ERP Evidence

Shared Neural Resources between Music and Language Indicate Semantic Processing of Musical Tension-Resolution Patterns

Neural substrates of processing syntax and semantics in music Stefan Koelsch

Music perception in cochlear implant users: an event-related potential study q

Short-term effects of processing musical syntax: An ERP study

Effects of Unexpected Chords and of Performer s Expression on Brain Responses and Electrodermal Activity

Melodic pitch expectation interacts with neural responses to syntactic but not semantic violations

What is music as a cognitive ability?

I. INTRODUCTION. Electronic mail:

PSYCHOLOGICAL SCIENCE. Research Report

Effects of Musical Training on Key and Harmony Perception

Bach Speaks: A Cortical Language-Network Serves the Processing of Music

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and

Auditory processing during deep propofol sedation and recovery from unconsciousness

Neuroscience Letters

ARTICLE IN PRESS. Neuroscience Letters xxx (2014) xxx xxx. Contents lists available at ScienceDirect. Neuroscience Letters

Are left fronto-temporal brain areas a prerequisite for normal music-syntactic processing?

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and

Event-Related Brain Potentials (ERPs) Elicited by Novel Stimuli during Sentence Processing

Syntactic expectancy: an event-related potentials study

Communicating hands: ERPs elicited by meaningful symbolic hand postures

Musical scale properties are automatically processed in the human auditory cortex

Semantic integration in videos of real-world events: An electrophysiological investigation

Making psycholinguistics musical: Self-paced reading time evidence for shared processing of linguistic and musical syntax

23/01/51. Gender-selective effects of the P300 and N400 components of the. VEP waveform. How are ERP related to gender? Event-Related Potential (ERP)

Abnormal Electrical Brain Responses to Pitch in Congenital Amusia Isabelle Peretz, PhD, 1 Elvira Brattico, MA, 2 and Mari Tervaniemi, PhD 2

Auditory semantic networks for words and natural sounds

The e ect of musicianship on pitch memory in performance matched groups

Object selectivity of local field potentials and spikes in the macaque inferior temporal cortex

BOOK REVIEW ESSAY. Music and the Continuous Nature of the Mind: Koelsch s (2012) Brain and Music. Reviewed by Timothy Justus Pitzer College

Structural Integration in Language and Music: Evidence for a Shared System.

Hemispheric asymmetry in the perception of musical pitch structure

Estimating the Time to Reach a Target Frequency in Singing

Stefan Koelsch: Brain and Music: A Contribution to the Investigation of Central Auditory Processing with a new Electrophysiological Approach.

Affective Priming. Music 451A Final Project

Stewart, Lauren and Walsh, Vincent (2001) Neuropsychology: music of the hemispheres Dispatch, Current Biology Vol.11 No.

Affective Priming Effects of Musical Sounds on the Processing of Word Meaning

Non-native Homonym Processing: an ERP Measurement

Neural evidence for a single lexicogrammatical processing system. Jennifer Hughes

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

What Can Experiments Reveal About the Origins of Music? Josh H. McDermott

I like my coffee with cream and sugar. I like my coffee with cream and socks. I shaved off my mustache and beard. I shaved off my mustache and BEARD

Expressive timing facilitates the neural processing of phrase boundaries in music: evidence from event-related potentials

How Order of Label Presentation Impacts Semantic Processing: an ERP Study

Harmony and tonality The vertical dimension. HST 725 Lecture 11 Music Perception & Cognition

Supplemental Material for Gamma-band Synchronization in the Macaque Hippocampus and Memory Formation

MUSICAL TENSION. carol l. krumhansl and fred lerdahl. chapter 16. Introduction

Dimensions of Music *

The Influence of Explicit Markers on Slow Cortical Potentials During Figurative Language Processing

ELECTROPHYSIOLOGICAL INSIGHTS INTO LANGUAGE AND SPEECH PROCESSING

Cross-modal Semantic Priming: A Timecourse Analysis Using Event-related Brain Potentials

Brain oscillations and electroencephalography scalp networks during tempo perception

Piano training enhances the neural processing of pitch and improves speech perception in Mandarin-speaking children

In press, Cerebral Cortex. Sensorimotor learning enhances expectations during auditory perception

With thanks to Seana Coulson and Katherine De Long!

NIH Public Access Author Manuscript Psychophysiology. Author manuscript; available in PMC 2014 April 23.

Impaired learning of event frequencies in tone deafness

Electrophysiological Evidence for Early Contextual Influences during Spoken-Word Recognition: N200 Versus N400 Effects

Construction of a harmonic phrase

THE INTERACTION BETWEEN MELODIC PITCH CONTENT AND RHYTHMIC PERCEPTION. Gideon Broshy, Leah Latterner and Kevin Sherwin

Lutz Jäncke. Minireview

Syntax in a pianist s hand: ERP signatures of embodied syntax processing in music

MEMORY IN MUSIC AND EMOTIONS

Semantic combinatorial processing of non-anomalous expressions

Connecting sound to meaning. /kæt/

Pre-Processing of ERP Data. Peter J. Molfese, Ph.D. Yale University

Event-Related Brain Potentials Reflect Semantic Priming in an Object Decision Task

Supporting Online Material

The N400 Event-Related Potential in Children Across Sentence Type and Ear Condition

Consciousness and Cognition

Sensitivity to musical structure in the human brain

The N400 and Late Positive Complex (LPC) Effects Reflect Controlled Rather than Automatic Mechanisms of Sentence Processing

PDF hosted at the Radboud Repository of the Radboud University Nijmegen

Music Lexical Networks

Distortion and Western music chord processing. Virtala, Paula.

Timbre-speci c enhancement of auditory cortical representations in musicians

The Role of Prosodic Breaks and Pitch Accents in Grouping Words during On-line Sentence Processing

Musical structure modulates semantic priming in vocal music

RP and N400 ERP components reflect semantic violations in visual processing of human actions

The Relationship Between Auditory Imagery and Musical Synchronization Abilities in Musicians

Can Music Influence Language and Cognition?

The Time Course of Orthographic and Phonological Code Activation Jonathan Grainger, 1 Kristi Kiyonaga, 2 and Phillip J. Holcomb 2

Individual differences in prediction: An investigation of the N400 in word-pair semantic priming

Two Neurocognitive Mechanisms of Semantic Integration during the Comprehension of Visual Real-world Events

The Interplay between Prosody and Syntax in Sentence Processing: The Case of Subject- and Object-control Verbs

Degree of Musical Expertise Modulates Higher Order Brain Functioning

Right Hemisphere Sensitivity to Word and Sentence Level Context: Evidence from Event-Related Brain Potentials. Seana Coulson, UCSD

DATA! NOW WHAT? Preparing your ERP data for analysis

MEANING RELATEDNESS IN POLYSEMOUS AND HOMONYMOUS WORDS: AN ERP STUDY IN RUSSIAN

Influence of tonal context and timbral variation on perception of pitch

The Power of Listening

DOI: / ORIGINAL ARTICLE. Evaluation protocol for amusia - portuguese sample

PICTURE PUZZLES, A CUBE IN DIFFERENT perspectives, PROCESSING OF RHYTHMIC AND MELODIC GESTALTS AN ERP STUDY

Changes in emotional tone and instrumental timbre are reflected by the mismatch negativity

Processing new and repeated names: Effects of coreference on repetition priming with speech and fast RSVP

Transcription:

BRAIN IMAGING Electric brain responses reveal gender di erences in music processing Stefan Koelsch, 1,2,CA Burkhard Maess, 2 Tobias Grossmann 2 and Angela D. Friederici 2 1 Harvard Medical School, Boston,USA; 2 Max Planck Institute of Cognitive Neuroscience, Stephan Str.1a, 04103 Leipzig, Germany CA,2 Corresponding Author and Address: koelsch@cns.mpg.de Received 25 January 2003; accepted 6 February 2003 DOI: 10.1097/01.wnr.0000065762.60383.67 The present study investigates gender di erences in the functional organization of the brain for music processing. In the language domain, males appear to have greater left hemisphere control than females. Despite some overlap of neural structures and processes for the perception of music and language, gender di erences of musical functions have so far not been reported. Data sets of three previous music experiments with event-related brain potentials (ERPs) were pooled and re-analyzed. Results demonstrate that an electrophysiological correlate of music-syntactic processing (ERAN, or music-syntactic MMN) is generated bilaterally in females, and with right hemispheric predominance in males.the present ndings indicate that gender di erences for the analysis of auditory information are not restricted to processes in the linguistic domain such as syntax, semantics, and phonology. NeuroReport 14 :709^713 c 2003 Lippincott Williams & Wilkins. Key words: Auditory processing; EEG; ERAN; ERP; Gender di erences; MMN; Music; N5 INTRODUCTION The brains of females and males have been reported to differ both structurally and functionally [1]. For the processing of language, brain imaging studies indicate gender differences in both anterior (inferior frontal gyrus) [2,3] and posterior (temporal) [4,5] language areas, activations of language functions being more preponderant in the left hemisphere in males, and more bilateral in females. These findings concur with lesion studies reporting that the incidence of aphasia is considerably higher in males than females with left-hemisphere lesions [6], supporting the notion of greater left hemisphere control of verbal functions in men than in women [1]. Similar gender differences have been reported for children in dichotic listening paradigms [7], and the prevalence of several language disorders in boys relative to girls has been taken to reflect the greater asymmetry of language functions in the male brain [1,6,8]. Despite considerable overlap of neural structures and processes underlying the perception of language and music [9 15], gender differences for the processing of music have so far not been reported. In order to investigate gender differences in regard to music processing, we reanalyzed the data sets of three recent experiments [15,16]. In these experiments, music processing was investigated using EEG and event-related brain potentials (ERPs). Chord sequences were presented to participants who did not have formal musical training (non-musicians). Each sequence consisted of five chords, and sequences were composed in a way that a musical context was built up towards the end of each sequence. Infrequently, an inappropriate harmony was presented at either the third or the fifth position of a chord sequence (Fig. 1). These inappropriate harmonies elicit a relatively early electric brain response with negative polarity, an amplitude maximum around 200 ms, and a right anterior scalp distribution. As a working term, this effect was labelled the early right anterior negativity (ERAN) [15]. The ERAN is taken to reflect the violation of a musical sound expectancy; the generation of these expectancies is based on (implicit) knowledge of complex musical regularities: in the mentioned experiments regularities of major minor tonal music (e.g. classical Western European music) which are described by music theory. Deviant acoustic events that are not presented within a complex musical context do not elicit an ERAN [16]. The ERAN can be elicited pre-attentively [17], suggesting that musical expectancies are automatically generated during the perception of a musical context. Moreover, the ERAN is larger in subjects with explicit musical long-term training (musicians) than in non-musicians [18], reflecting that a more precise representation of musical regularities leads to more specific musical expectancies. The neural generators of the ERAN have been localized in the inferior fronto-lateral cortex (inferior part of the pars opercularis, Brodmann s area 44, in the left hemisphere often denoted as Broca s area) [14]. The ERAN is usually followed by a negative potential around 500 ms (denoted N5) [15,17,18], which is taken to reflect processes of harmonic integration. The locations of the neural 0959-4965 c Lippincott Williams & Wilkins Vol 14 No 5 15 April 2003 709

S. KOELSCH ETAL. under computerized control via MIDI on a synthesizer. In 10% of the sequences, an in-key chord at the second, third, fourth or fifth position was played by another instrument than piano (e.g. harpsichord, celesta, marimba). In Experiment 2 [15], stimuli were the same as in Experiment 1, except that Neapolitan chords were replaced by half-tone clusters. Clusters represent a higher degree of violation than Neapolitans; consequently, ERAN and N5 elicited by clusters had a larger amplitude than when elicited by Neapolitans (for details see [15]). In Block 2 [16], stimuli were the same as in Experiment 1, except that (a) 255 sequences were presented, (b) no secondary dominants were presented at the second position, (c) no deviant instruments were presented, and (d) the probability was 20% for a Neapolitan at the third, and 20% for a Neapolitan at the fifth position. Fig.1. Examples of chord sequences consisting of in-key chords only (top row), containing a Neapolitan chord at the third and fth position (middle row), and containing clusters at the third and fth position (bottom row). The arrows indicate Neapolitans and clusters. generators of the N5 are not yet clear; a recent fmri study indicates that, in addition to fronto-lateral, superior temporal areas are activated during the processing of unexpected musical events [19]. The aim of the present study was to determine whether ERAN and N5 differ between males and females. In order to increase the signal-to-noise ratio of the data, we pooled 62 data sets from three previous EEG experiments in which ERAN and N5 were elicited. MATERIALS AND METHODS Subjects: Data sets from 62 right-handed subjects (age 19 31 years, 31 females) were analyzed: the data sets from Experiment 1 (n ¼ 18) and Experiment 2 (n ¼ 18) of Koelsch et al. [15], and the 26 data sets from Block 2 of Koelsch et al. [16]. All subjects were non-musicians with no formal musical training (apart from normal school education). Stimuli: In Experiment 1 [15], 172 chord sequences were presented, each consisting of five chords. The first chord was the tonic of the following cadence, chords at the second position were: tonic, mediant, submediant, subdominant, dominant to the dominant, secondary dominant to mediant, secondary dominant to submediant, secondary dominant to supertonic; at the third position: subdominant, dominant, dominant six-four chord, Neapolitan sixth chord; at the fourth position: dominant seventh chord; at the fifth position: tonic or Neapolitan sixth chord. In 25% of the sequences, a Neapolitan chord was presented at the third position, and in another 25% a Neapolitan was presented at the fifth position of a chord sequence. Sequences with Neapolitan chords and in-key sequences were randomly intermixed. Each variation of a Neapolitan chord was presented at the third position in one sequence, and at the fifth position in another. Presentation time of chords 1 4 was 600 ms, chord 5 was presented for 1200 ms. All chords had the same decay of loudness, i.e. Neapolitan chords at the third and at the fifth position were on average physically identical (within the first 600 ms). Chords were played Procedure: In Experiments 1 and 2 [15], participants were only informed about the deviant instruments, not about the Neapolitan chords or the clusters. Participants were instructed to ignore the harmonies, to count the deviant instruments, and to report the number of deviant instruments when asked by pressing a response button. After an inquiry, the following chord sequences were from another key (resulting in seven sub-blocks from a different key). During Block 2 [16], subjects were playing a video-game under the instruction to ignore all acoustic stimuli. Data analysis: Continuous EEG data from 25 scalp locations of the 10 20 system were referenced off-line to the algebraically mean of left and right mastoid electrodes. From the continuous raw EEGs, trials were averaged offline; the epoch-length was 1100 ms, including a pre-stimulus baseline from 100 to 0 ms (with respect to the onset of a chord). ERPs elicited by appropriate (in-key) and inappropriate harmonies (clusters and Neapolitans cumulated) were averaged separately for the third and the fifth position, and separately for men and women. Data sets of the three experiments were pooled, the collapsing of the data sets is justified by the similarities in stimulation and paradigm: all experiments had (a) identical in-key chords, (b) identical timing, (c) in all experiments harmonically unexpected musical events were presented within a musical context with (d) similar probablilities at (e) identical positions within the chord sequences (third and fifth position). Note that the unexpected harmonies investigated in the mentioned experiments (clusters and Neapolitans) elicit both ERAN and N5, which are the subject of investigation in the present study. For statistical analyses, two regions of interest (ROI) were computed [20]: a region of left frontal (F7, F3, FT7, FC3) and a region of right frontal electrodes (F4, F8, FC4, FT8). These ROIs correspond to the ROIs analyzed in the previous studies [15,16]. The computation of ROIs is a means to avoid loss of statistical power when ANOVAs are used to quantify potentials measured by adjacent electrodes [20]. Differences between mean amplitudes of these regions were tested by conducting repeated measurement ANOVAs. Global ANO- VAs were conducted with the factors chord type (appropriate inappropriate harmonies), lateralization (left right hemisphere), and gender. Time windows for statistical 710 Vol 14 No 5 15 April 2003

ELECTRIC BRAIN RESPONSES REVEALGENDER DIFFERENCES IN MUSIC PROCESSING analyses were chosen around the amplitude maxima of effects. In the time windows analyzed, effects had a stable scalp topography. After the statistical evaluation, ERPs were for presentation low-pass filtered at 10 Hz (41 points, FIR). Females F7 F8 RESULTS Figure 2 shows the ERPs elicited at the fifth position of the chord sequences, separately for males and females. The data indicate that the inappropriate harmonies elicited (in both males and females) an early as well as a late negativity. In both groups, the amplitude of the early negativity (the ERAN) was maximal around 200 ms after the onset of the chord, the late negativity (the N5) peaked around 550 ms. The mean amplitudes of these effects did not differ between groups. Whereas the early negativity had a clear right hemispheric preponderance in males, it was bilateral in females (Figs. 2 and 3). This difference in scalp topography was not due to larger potentials over the right hemisphere of males, but due to smaller potentials in males over the left hemisphere. These observations were supported by a statistical analysis of the scalp topographies. A global ANOVA for a time window from 130 to 280 ms revealed an effect of chord type (F(1,60) ¼ 58.49, p o 0.0001), indicating that the effect of the inappropriate harmonies is significant), an interaction between factors chord type and hemisphere (F(1,60) ¼ 5.66, p o 0.02), indicating that this effect was lateralized), and an interaction between factors chord type, lateralization, and gender (F(1,60) ¼ 4.24, p o 0.05), indicating that the lateralization of the effect differed between males and females). No further two-way interactions were found. To validate the three-way interaction, the same analysis was performed after amplitude normalization [21]. In this analysis the interaction remained significant (F(1,60) ¼ 4.96, p o 0.03). Consequently, it has to be concluded that the scalp topography of the early negative effect elicited by the harmonically inappropriate chords differed in its lateralization between genders. When data sets were analyzed separately for each experiment, the three-way interaction (chord type lateralization gender) was not clearly significant in any single experiment, but in two of the three experiments the estimated p-value was o 0.07. When analyzing the pooled data sets separately for males and females, data for males show a clear interaction between the factors chord type and hemisphere (F(1,30) ¼ 5.37, p o 0.005), indicating that the effect was lateralized), whereas data for females show no interaction (p 4 0.75). Although the late negativity seems to be larger in females, and probably stronger lateralized in males, these differences were statistically not significant. An ANOVA for a time window from 450 to 600 ms revealed an effect of chord type (F(1,60) ¼ 28.37, p o 0.0001), but no two- or three-way interaction. As already shown in the data from the single experiments [15,16], early and late negativities elicited by the inappropriate harmonies were significantly smaller at the third position than at the fifth position, because the musical context was built up to a smaller degree in the middle of the sequence, leading to less specific expectancies (compared to Males 5 5 µv 0.2 0.4 FT7 F7 FT7 s 0.6 0.8 1.0 Legend : In-key chord Neapolitans FT8 F8 FT8 Fig. 2. ERP waveforms of the electric potentials elicited at the fth position of the chord sequences by appropriate (solid line) and inappropriate harmonies (Neapolitan chords and clusters, dashed line), averaged across females (top) and males (bottom). Electrodes of the left column (F7, FT7) represent left anterior leads, electrodes of the right colum (F8, FT8) represent right anterior leads. In both groups, the inappropriate harmonies elicited an early negativity (long arrow) which was followed by a late negativity (short arrow).the early negativity had a bilateral scalp distribution in females and was preponderant over the right hemisphere in males. Whereas the amplitude of the early negativity did not di er between genders over the right hemisphere, the potential was smaller over the left hemisphere in males. Vol 14 No 5 15 April 2003 711

S. KOELSCH ETAL. maxima Fig. 3. Scalp distribution of the early e ect of inappropriate harmonies, separately for females (left) and males (right).the gures show isopotential lines of the di erence ERPs (appropriate subtracted from inappropriate harmonies) 200 ms after the onset of the stimuli. The thick line indicates the zero-potential line, the dashed lines anterior to the thick line indicate negative potential values, the solid lines posteriorly indicate positive values; the electrical potential di erence displayed by two adjacent isopotential lines is 0.2 mv. The potential maps show that the early e ect is lateralized to the right in males and bilateral in females (the arrows indicate the negative potential maxima of the e ects). at the end of the sequence, where a strong expectancy for the tonic chord was established) [15 17]. DISCUSSION The present study reveals that the early negativity elicited by inappropriate harmonies presented within a musical context has a bilateral potential distribution over the scalp in females and a right-lateralized distribution in males. The early negativity has been taken to reflect a musical sound expectancy violation, the musical expectancies being generated during the perception of a preceding musical context, based on (implicit) knowledge of complex musical regularities [22 24] (the implicit knowledge being most presumably culture specific). The system of musical regularities of major minor tonal music is described by music theory (with respect to the succession of chord functions, the harmonic relationships between chords and keys, part writing, etc.) and may be denoted as a musical syntax [13 15]. The term musical syntax does not imply that it is a linguistic syntax in musical terms; rather it points to the fact that music is structured according to complex regularities (a feature reminiscent of language). With respect to the differences in scalp topography between genders it is interesting to note that the ERAN has been suggested to be generated in the inferior frontolateral cortex (BA 44), which is in the left hemisphere part of Broca s area [14] (this observation was supported by a recent fmri study [19]). The present results are not taken as evidence for a gender-specific modulation of the ERAN in BA44, as claims about generators cannot be made based on scalp topography alone. However, both functional [2 4] and anatomical [25,26] gender differences have been reported for BA 44: in functional imaging studies, activations of language functions in this region were more strongly leftlateralized in males than in females (who showed more bilateral activations) [2 4]. Anatomical studies indicate that the cell density of BA 44 is lower in the right hemisphere in males, but not in females [25], and that Broca s area is proportionally larger in the female brain [26]. The functional and structural gender differences have been taken as evidence for a more symmetrical control of verbal functions in women and, correspondingly, to a more lateralized control in men. The gender differences reported in previous studies thus correspond with the present observation that the early music negativity is lateralized in males and bilateral in females, although with a hemispheric weighting different from that of language. It is highly plausible that the observed gender difference is due to a gender-specific hemispheric weighting in the recruitment of neural ressources, e.g. a gender-wise difference in the strength or number of generators; this interpretation concurs with the mentioned results of previous functional imaging studies [2 4]. A gender-wise difference in the orientation of identical generators in each hemisphere is rather unlikely, because this would require gross anatomical gender differences in gyrification and sulcification (note that the present data were obtained with EEG). With this respect, the present data strongly suggest that females process music-syntactic information bilaterally, and males with a right-hemispheric predominance. Note, however, that the early negativity of males was also present over the left hemisphere, indicating that males did not exclusively employ the right hemisphere for the processing of the musical information. Note that the present study, as well as previous studies [14 18], indicate some similarities and differences between the ERAN and the mismatch negativity (MMN) [27]. An important difference is that the amplitude of the ERAN depends on a regularity-based generation of harmonic expectancies for functionally appropriate chords [14,16,17], whereas the MMN is broadly assumed to reflect sensory memory operations [27]. In addition, the present study shows that the ERAN differs in its scalp distribution between males and females, which has (at least so far) not been reported for the MMN. However, the ERAN and the MMN also share a number of features [15 17], including the sensitivity to stimuli that do not match with a previously presented group of events. Because the early negativity was lateralized only in males, the term early right anterior negativity falls short as a descriptor for the effect in general. Given the similarities between ERAN and MMN, and the findings that the ERAN is sensitive to violations of complex musical regularities, the ERAN has previously been suggested to be understood as a music-syntactic MMN [18]. CONCLUSION The present results indicate that relatively early electric brain activity elicited by inappropriate harmonies within a musical sequence is distributed bilaterally over the scalp in females, and lateralized to the right hemisphere in males. This finding reveals gender differences for the processing of musical information, indicating that gender differences in auditory information processing are not restricted to processes in the linguistic domain such as syntax, semantics, and phonology. Interestingly, because of the female brain s ability to process language bilaterally, it appears that (a) females can better compensate aphasic impairments 712 Vol 14 No 5 15 April 2003

ELECTRIC BRAIN RESPONSES REVEALGENDER DIFFERENCES IN MUSIC PROCESSING [6], (b) girls have fewer language disorders than boys [8], and (c) females have facilitated language and verbal learning [28]. The gender difference observed in the present study might contribute to these effects, because the processing of musical elements of speech is an important aspect of speech perception and language acquisition [29,30]. Finally, on the basis of the present findings one should predict differential impairment in syntactic music processing between men and women suffering from lesions within the right inferior fronto-lateral cortex. REFERENCES 1. Kimura D. Sex and Cognition. Cambridge: MIT Press; 1999. 2. Pugh K, Shaywitz B, Shaywitz S et al. Brain 119, 1221 1238 (1996). 3. Jaeger J, Lockwood A, Valin R et al. Neuroreport 9, 2803 2807 (1998). 4. Kansaku K, Yamaura A and Kitazawa S. Cerebr Cortex 10, 866 872 (2000). 5. Phillips M, Lowe M, Dzemidzic M et al. Radiology 220, 202 207 (2001). 6. McGlone J. Behav Brain Sci 3, 215 263 (1980). 7. Moulden J and Persinger M. Percept Mot Skills 90, 893 898 (2000). 8. Hier D. Bull Orton Soc 29, 74 83 (1979). 9. Platel H, Price C, Baron J et al. Brain 120, 229 243 (1997). 10. Liegeois-Chauvel C, Peretz I, Babaie M et al. Brain 121, 1853 1867 (1998). 11. Peretz I, Kolinsky R, Tramo M et al. Brain 117, 1283 1301 (1994). 12. Zatorre R, Evans A and Meyer E. J Neurosci 14, 1908 1919 (1994). 13. Patel AD, Gibson E, Ratner J et al. J Cogn Neurosci 10, 717 733 (1998). 14. Maess B, Koelsch S, Gunter T et al. Nature Neurosci 4, 540 545 (2001). 15. Koelsch S, Gunter T, Friederici AD et al. J Cogn Neurosci 12, 520 541 (2000). 16. Koelsch S, Gunter T, Schröger E et al. Neuroreport 12, 1385 1389 (2001). 17. Koelsch S, Schröger E and Gunter T. Psychophysiology 39, 1 11 (2002). 18. Koelsch S, Schmidt B and Kansok J. Psychophysiology 39, 657 663 (2002). 19. Koelsch S, Gunter T, von Cramon DY et al. Neuroimage 17, 956 966 (2002). 20. Oken B and Chiappa K. Ann Neurol 19, 493 494 (1986). 21. McCarthy G and Wood C. Electroencephalogr Clin Neurophysiol 69, 218 233 (1985). 22. Krumhansl C and Kessler E. Psychol Rev 89, 334 368 (1982). 23. Bharucha J and Krumhansl C. Cognition 13, 63 102 (1983). 24. Tillmann B, Bharucha J and Bigand E. Psychol Rev 107, 885 913 (2000). 25. Amunts K, Schleicher A, Burgel U et al. J Comp Neurol 412, 319 341 (1999). 26. Harasty J, Double K, Halliday G et al. Arch Neur 54, 171 176 (1997). 27. Schröger E. Behav Res Meth Instr Comp 30, 131 145 (1998). 28. Kramer JH, Delis DC, Kaplan E et al. Neuropsychology 11, 577 584 (1997). 29. Steinhauer K, Alter K and Friederici AD. Nature Neurosci 2, 191 196 (1999). 30. Papous ek H. In: Sloboda J and Deliege I (eds). Musical Beginnings. Oxford: Oxford University Press; 1996. Vol 14 No 5 15 April 2003 713