Auditory scene analysis

Similar documents
Pitch Perception and Grouping. HST.723 Neural Coding and Perception of Sound

AUD 6306 Speech Science

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

Pitch perception for mixtures of spectrally overlapping harmonic complex tones

Voice segregation by difference in fundamental frequency: Effect of masker type

9.35 Sensation And Perception Spring 2009

Pitch. The perceptual correlate of frequency: the perceptual dimension along which sounds can be ordered from low to high.

Harmony and tonality The vertical dimension. HST 725 Lecture 11 Music Perception & Cognition

Do Zwicker Tones Evoke a Musical Pitch?

A DISSERTATION SUBMITTED TO THE FACULTY OF THE GRADUATE SCHOOL OF THE UNIVERSITY OF MINNESOTA BY

2018 Fall CTP431: Music and Audio Computing Fundamentals of Musical Acoustics

HST 725 Music Perception & Cognition Assignment #1 =================================================================

Springer Handbook of Auditory Research. Series Editors: Richard R. Fay and Arthur N. Popper

Effect of harmonic rank on sequential sound segregation

I. INTRODUCTION. Electronic mail:

The presence of multiple sound sources is a routine occurrence

Music Perception & Cognition

Brian C. J. Moore Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, England

Quarterly Progress and Status Report. Violin timbre and the picket fence

Musical Acoustics Lecture 15 Pitch & Frequency (Psycho-Acoustics)

Timbre perception

Perceptual Considerations in Designing and Fitting Hearing Aids for Music Published on Friday, 14 March :01

Pitch is one of the most common terms used to describe sound.

2 Autocorrelation verses Strobed Temporal Integration

Demonstrations. to accompany Bregman s. Auditory Scene Analysis. The perceptual organization of sound MIT Press, 1990

Temporal Envelope and Periodicity Cues on Musical Pitch Discrimination with Acoustic Simulation of Cochlear Implant

Music Perception with Combined Stimulation

UNIVERSITY OF DUBLIN TRINITY COLLEGE

Neural Correlates of Auditory Streaming of Harmonic Complex Sounds With Different Phase Relations in the Songbird Forebrain

Behavioral and neural identification of birdsong under several masking conditions

Topic 1. Auditory Scene Analysis

The Tone Height of Multiharmonic Sounds. Introduction

CTP431- Music and Audio Computing Musical Acoustics. Graduate School of Culture Technology KAIST Juhan Nam

The perception of concurrent sound objects through the use of harmonic enhancement: a study of auditory attention

Informational masking of speech produced by speech-like sounds without linguistic content

Psychoacoustics. lecturer:

CTP 431 Music and Audio Computing. Basic Acoustics. Graduate School of Culture Technology (GSCT) Juhan Nam

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

Polyrhythms Lawrence Ward Cogs 401

Brain.fm Theory & Process

MEMORY & TIMBRE MEMT 463

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and

Our Perceptions of Music: Why Does the Theme from Jaws Sound Like a Big Scary Shark?

Melody: sequences of pitches unfolding in time. HST 725 Lecture 12 Music Perception & Cognition

Pitch Perception. Roger Shepard

EMS : Electroacoustic Music Studies Network De Montfort/Leicester 2007

THE PSYCHOACOUSTICS OF MULTICHANNEL AUDIO. J. ROBERT STUART Meridian Audio Ltd Stonehill, Huntingdon, PE18 6ED England

Music Source Separation

Pitch: The Perceptual Ends of the Periodicity; but Of What Periodicity?

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

What is music as a cognitive ability?

Rhythm: patterns of events in time. HST 725 Lecture 13 Music Perception & Cognition

Consonance, 2: Psychoacoustic factors: Grove Music Online Article for print

Consonance and Pitch

Auditory Stream Segregation (Sequential Integration)

I. INTRODUCTION. 1 place Stravinsky, Paris, France; electronic mail:

Quarterly Progress and Status Report. Perception of just noticeable time displacement of a tone presented in a metrical sequence at different tempos

Acoustic Scene Classification

Perception and cortical neural coding of harmonic fusion in ferrets. (Dated: March 18, 2007)

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Proceedings of Meetings on Acoustics

Tempo and Beat Analysis

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng

MAutoPitch. Presets button. Left arrow button. Right arrow button. Randomize button. Save button. Panic button. Settings button

Physics and Neurophysiology of Hearing

Lecture 2 What we hear: Basic dimensions of auditory experience

Asynchronous Preparation of Tonally Fused Intervals in Polyphonic Music

Creative Computing II

Simple Harmonic Motion: What is a Sound Spectrum?

Analysis, Synthesis, and Perception of Musical Sounds

Speech Recognition and Signal Processing for Broadcast News Transcription

Automatic music transcription

Topics in Computer Music Instrument Identification. Ioanna Karydi

Music BCI ( )

We realize that this is really small, if we consider that the atmospheric pressure 2 is

2005 BY THE REGENTS OF THE UNIVERSITY OF CALIFORNIA. The Influence of Pitch Interval on the Perception of Polyrhythms

Timbre blending of wind instruments: acoustics and perception

Topic 10. Multi-pitch Analysis

Musical Illusions Diana Deutsch Department of Psychology University of California, San Diego La Jolla, CA 92093

University of California Press is collaborating with JSTOR to digitize, preserve and extend access to Music Perception: An Interdisciplinary Journal.

1aAA14. The audibility of direct sound as a key to measuring the clarity of speech and music

EE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function

Digital audio and computer music. COS 116, Spring 2012 Guest lecture: Rebecca Fiebrink

"The mind is a fire to be kindled, not a vessel to be filled." Plutarch

Getting Started with the LabVIEW Sound and Vibration Toolkit

A Behavioral Study on the Effects of Rock Music on Auditory Attention

Auditory Illusions. Diana Deutsch. The sounds we perceive do not always correspond to those that are

Measurement of overtone frequencies of a toy piano and perception of its pitch

Auditory streaming of amplitude modulated sounds in the songbird forebrain

August Acoustics and Psychoacoustics Barbara Crowe Music Therapy Director. Notes from BC s copyrighted materials for IHTP

Informational Masking and Trained Listening. Undergraduate Honors Thesis

聲音有高度嗎? 音高之聽覺生理基礎. Do Sounds Have a Height? Physiological Basis for the Pitch Percept

Proceedings of Meetings on Acoustics

Smooth Rhythms as Probes of Entrainment. Music Perception 10 (1993): ABSTRACT

Quarterly Progress and Status Report. An attempt to predict the masking effect of vowel spectra

Hugo Technology. An introduction into Rob Watts' technology

Making music with voice. Distinguished lecture, CIRMMT Jan 2009, Copyright Johan Sundberg

HST Neural Coding and Perception of Sound. Spring Cochlear Nucleus Unit Classification from Spike Trains. M.

INTRODUCTION J. Acoust. Soc. Am. 107 (3), March /2000/107(3)/1589/9/$ Acoustical Society of America 1589

The Beat Alignment Test (BAT): Surveying beat processing abilities in the general population

Transcription:

Harvard-MIT Division of Health Sciences and Technology HST.723: Neural Coding and Perception of Sound Instructor: Christophe Micheyl Auditory scene analysis Christophe Micheyl

We are often surrounded by various sound sources. Some of importance to us; others, a nuisance. SINGER Figures by MIT OCW.

The waves from these sources mingle before reaching our ears. SINGER Figures by MIT OCW.

The result is a complex acoustic mixture. Figures removed due to copyright reasons.

The auditory system must disentangle the mixture to permit (or at least facilitate) source identification Figures removed due to copyright reasons.

Solution: Figures removed due to copyright reasons.

Some of the questions that we will address: - What tricks does the auditory system use to analyze complex scenes? - What neural/brain processes subtend these perceptual phenomena? - Why do hearing-impaired listeners have listening difficulties in the presence of multiple sound sources?

Why is this important? -Understand how the auditory system works in real-life (the system was probably not designed primarily to process isolated sounds) -Build artificial sound-processing systems that can do ASA like us (speaker separation for speech recognition, instrument separation for music transcription, content-based indexing in audio recordings, ) - or help us do it better (sound pre-processing for intelligent hearing aids, enhanced speech-in-noise understanding, )

Bottom-up and top-down mechanisms Bottom-up (or primitive ) mechanisms -partition the sensory input based on simple stimulus properties -largely automatic (pre-attentive) -probably innate or acquired early during infancy Top-down (or schema-based ) mechanisms -partition the input based on stored object representations (prototypes) -heavily dependent upon experience/knowledge

The basic laws of perceptual organization courtesy of: the Gestalt-psychology school promixity similarity closure continuity etc

Top-down Figure removed due to copyright reasons.

Sequential and simultaneous mechanisms Sequential mechanisms (auditory streaming ) Figures removed due to copyright reasons.

Sequential and simultaneous mechanisms Simultaneous mechanisms Level Clarinet Voice Figure removed due to copyright reasons. Freq

Outline I. Simultaneous ASA processes -Harmonicity - Onset/offset - Co-modulation II. Sequential ASA processes - Auditory streaming

Harmonicity Many important sounds are harmonic (vowels of speech, most musical sounds, animal calls, ) Level 200 400 600 800 1000 1200 Frequency F0 2F0 3F0 4F0 5F0 6F0 Does the auditory system exploit this physical property to group/segregate frequency components?

Harmonic fusion Harmonic complexes are generally perceived as one sound Level stimulus several components several frequencies percept 1 sound 1 pitch 200 400 600 800 1000 1200

Deviations from harmonicity promote segregation If a harmonic is mistuned by > 2-3%, it stands out perceptually (Moore et al., 1985, 1986; Hartmann et al., 1990 ) stimulus percept Level 1 sound pitch = 200 Hz 200 400 600 800 1000 1200 2 sounds harmonic, pitch=200hz + pure tone, 618 Hz 200 400 618 800 1000 1200 Frequency

Frequency in tune Demonstration From: Bregman (1990) Auditory scene analysis MIT Press Demo CD 1.2kHz Time

Level Influence of harmonic grouping/segregation on other aspects of auditory perception Mistuning a harmonic near a formant can affect the perceived identity of a vowel 1 st formant peak Darwin & Gardner (1986) /I/ /e/ /e/ Frequency

Mechanisms of harmonicity-based grouping? Spectral: the harmonic sieve (Duifhuis et al., 1982) Components that pass through the sieve are grouped; those that don t are excluded Level Frequency Level Level Frequency Frequency

Actual mechanisms of harmonicity-based grouping? Harmonics above the 10 th can generally not be heard out (Moore et al., 1985) This suggests a role of peripheral frequency selectivity, because harmonics above the 10 th are generally unresolved in the cochlea: The cochlea as a filter bank Level 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Frequency 45 45 Simulated Spectral EPs: Level (db) 40 35 30 25 20 F0 = 200Hz Level (db) 40 35 30 25 20 F0 = 400Hz 15 15 10 10 5 1000 2000 3000 4000 5000 6000 Frequency (Hz) 5 1000 2000 3000 4000 5000 6000 Frequency (Hz)

Mechanisms of harmonicity-based grouping? Temporal: across-channel synchrony (Roberts & Brunstrom, 2001) Components that elicit synchronous neural responses are grouped Level Mistuned harmonic Time Above 2000 Hz, harmonics become increasingly harder to hear out (Hartmann et al., 1990) This suggests a contribution of temporal mechanisms, because phase locking breaks down at high frequencies

An aside: harmonicity or equal spectral spacing? Grouping/segregation of spectral components is based not solely on harmonicity, but also on spectral spacing Roberts & Bregman (1991) Level Frequency-shifted complex (inharmonic, but equally spaced components) 200Hz Shifting the frequency of a component in a shifted complex makes it stand out 250 450 650 850 1050 1250 1450 Odd-numbered harmonics + 1 even-numbered harmonic 200Hz 200Hz The even-numbered harmonic stands out more than the neighboring odd-numbered harmonics 200 600 800 1000 1400 Freq. But the utility of a specific spectral-spacing-based grouping mechanism is questionable

F0-based segregation of whole harmonic complexes Level stimulus percept 1 sound 200 Hz Sound A harmonic F0=200 Hz + 1 sound 240 Hz Sound B harmonic F0=240 Hz = A+B inharmonic? Frequency

Double vowels Two (synthetic) vowels with different F0s played simultaneously Level vowel A F0=100 Hz /o/ vowel B F0=140 Hz /e/? A+B Frequency

Double vowels Can listeners use F0 differences to sort out the frequency components? vowel A F0=100 Hz /o/ /e/ vowel B F0=140 Hz harmonics corresponding to one F0 harmonics corresponding to the other F0

Concurrent vowels F0 differences facilitate the identification of concurrent vowels (Scheffers, 1983; Assmann & Summerfield, 1990; ) Figure removed due to copyright reasons. Please see: Assmann, and Summerfield. J. Acoust. Soc. Am. 88 (1990): 680-687. (but note %-correct well above chance even with no F0 difference, providing evidence for a role of template-based mechanisms) This also works with whole sentences (Brokx & Nooteboom, 1982)

Actual mechanisms of harmonicity-based grouping? Harmonics above the 10 th can generally not be heard out (Moore et al., 1985) This suggests a role of peripheral frequency selectivity, because harmonics above the 10 th are generally unresolved in the cochlea: The cochlea as a filter bank Level 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Frequency 45 45 Simulated Spectral EPs: Level (db) 40 35 30 25 20 F0 = 200Hz Level (db) 40 35 30 25 20 F0 = 400Hz 15 15 10 10 5 1000 2000 3000 4000 5000 6000 Frequency (Hz) 5 1000 2000 3000 4000 5000 6000 Frequency (Hz)

Influence of frequency resolution on the F0-based segregation of concurrent complexes Example simulated spectral excitation patterns in response to harmonic complex target, maskers, and target+masker mixtures at different F0s F0 400 Hz resolved harmonics F0 100 Hz unresolved harmonics 50 40 Tgt+Msk Tgt Msk 50 40 Tgt+Msk Tgt Msk Level (db) 30 20 Level (db) 30 20 10 10 0 1000 2000 3000 4000 5000 6000 Frequency (Hz) resulting EP displays some peaks 0 1000 2000 3000 4000 5000 6000 Frequency (Hz) resulting EP displays no peaks

Influence of frequency resolution on the F0-based segregation of concurrent complexes Freq (Carlyon, 1996; Micheyl & Oxenham, 2004) Pitch is going up Freq Target Masker Time? Time F0-based segregation does not work if all frequency components are unresolved

Influence of frequency resolution on the F0-based segregation of concurrent complexes Yet, in principle, it is possible to segregate two periodic components falling into the same peripheral auditory filter using some temporal mechanism (harmonic cancellation model, de Cheveigné et al., 1992; timing nets, Cariani, 2001) + Temporal analysis (e.g., timing net or AC) Time Our results (Micheyl & Oxenham, 2004) and those of Carlyon (1996) indicate that the auditory system makes very limited (if any) use of this temporal strategy for segregating simultaneous harmonic complexes

Implications for hearing-impaired listeners Cochlear damages Loss of frequency selectivity (broadened auditory filters) Reduced resolvability of frequency components Reduced ability to perceptually isolate simultaneous harmonic sounds Reduced ability to extract the individual properties (loudness, pitch, timbre) of these sounds

Onset time Frequency components that start together tend to fuse together stimulus percept Frequency 1 sound

Onset time Onset asynchronies promote perceptual segregation stimulus percept Frequency 1 sound 2 sounds Time

Influence of onset grouping/segregation on other aspects of auditory perception De-synchronizing a harmonic near a formant can affect perceived vowel identity Level 1 st formant peak Darwin (1984); Darwin & Sutherland (1984) Freq /I/ /e/ /e/ Freq 40ms Time

Demonstration of onset asynchrony and vowel identity Frequency From: Bregman (1990) Auditory scene analysis MIT Press Demo CD ee en? Time

Co-modulation. I. Frequency modulation (FM) When the F0 of a harmonic sound changes, all of its harmonics change frequency coherently Frequency (linear scale) Time

Co-modulation. I. Frequency modulation (FM) Coherent FM promotes the fusion of harmonics Darwin et al. (1994) stimulus percept Frequency 2 sounds same amount of mistuning 1 sound Time

Frequency FM-based grouping - Demo 1 FM can make harmonics stand out From: Bregman (1990) Auditory scene analysis MIT Press Demo CD Time

FM-based grouping - Demo 2 Incoherent FM promotes segregation Frequency From: Bregman (1990) Auditory scene analysis MIT Press Demo CD Time

Is it FM or harmonicity? Frequency harmonic Carlyon (1991) Condition 1 inharmonic Which sound contains the incoherent FM? 2 nd sound Condition 2 inharmonic inharmonic? Time

Co-modulation. II. Amplitude modulation Current evidence in favor of a genuine AM-based grouping/segregation mechanism is weak, at best Out-of phase AM generally results in onset asynchronies (leading to the question: is it really AM phase or rather onset asynchrony?) Out-of phase AM results in some spectral components being well audible while the others are not, at certain times (leading to the question: is the pop-out due to AM or enhanced SNR?)

Auditory streaming What is it? Description and demonstration of the phenomenon

Frequency f B B A A A A Time

Frequency 1 stream gallop B B A A A A Time

Frequency B B f A A A A Time

Frequency 2 streams! one high and slow, the other low and fast B B A A A A Time

Frequency B B A A A A Time

A basic pre-requisite for any neural correlate of streaming: depend on both df and dt df temporal coherence boundary always 2 streams 1 or 2 streams always 1 stream fission boundary dt Tone repetition rate

1.0 0.8 0.6 0.4 0.2 0.0 0 1 2 3 4 5 6 7 8 9 Time (s) f 1 ST 3 ST 6 ST 9 ST Probability '2 streams' response

Build-up up Probability '2 streams' response 1.0 0.8 0.6 0.4 0.2 0.0 0 1 2 3 4 5 6 7 8 9 Time (s) f 1 ST 3 ST 6 ST 9 ST

Traditional explanations for the build-up «Neurophysiological» explanation Neural adaptation of coherence/pitch pitch-motion detectors (Anstis & Saida, 1985) «Cognitive» explanation The default is integration (1 stream); the brain needs to accumulate evidence that there is more than 1 stream before declaring «2 streams» (Bregman,, 1978, 1990, )

Asymptote Probability '2 streams' response 1.0 0.8 0.6 0.4 0.2 0.0 0 1 2 3 4 5 6 7 8 9 Time (s) f 1 ST 3 ST 6 ST 9 ST

1.0 Probability '2 streams' response 0.8 0.6 0.4 0.2 Percept is switching f 1 ST 3 ST 6 ST 9 ST 0.0 0 1 2 3 4 5 6 7 8 9 Time (s)

Ambiguous stimuli and bi-stable percepts Necker s cube Rubin s vase-faces Figures removed due to copyright reasons. have been used successfully in the past to demonstrate neural/brain correlates of visual percepts e.g., Logothetis & Schall (1989), Leopold & Logothetis (1996),..

Streaming How does it work? Theories and computational models

The channeling theory Hartmann and Johnson (1991) Music Percept. Level Frequency

The channeling theory Hartmann and Johnson (1991) Music Percept. 1 stream Level AB Frequency

The channeling theory Hartmann and Johnson (1991) Music Percept. 2 streams Level A B Frequency

Streaming How does it really work? Neural mechanisms

Behavioral evidence that streaming occurs in - monkey monkey (Izumi,, 2002) - bird (Hulse et et al., 1997; McDougall-Shackleton et al,, 1998) - fish fish (Fay,, 1998)

1 stream Stim.. X Resp.. P Stimulus parameters 1 stream Stim.. Y Resp.. Q

Single/few/multi-unit intra-cortical recordings Monkeys: Fishman et al. (2001) Hear. Res. 151, 167-187 Bats: Kanwal, Medvedev, Micheyl (2003) Neural Networks Figures removed due to copyright reasons. Please see: Fishman et al. (2001) At low repetition rates, units respond to both on- and off-bf tones At high repetition rates, only on-bf tone response is visible

Is peripheral chanelling the whole story?

Sounds that excite the same peripheral channels can yield streaming Vliegen & Oxenham (1999) Vliegen,, Moore, Oxenham (1999) Grimault, Micheyl, Carlyon et al.. (2001) Grimault, Bacon, Micheyl (2002) Roberts, Glasberg,, Moore (2002)...

Streaming with complex tones Amplitude F0 400Hz 800Hz 1200Hz F0 150Hz 450Hz 300Hz F0 Frequency

Streaming based on F0 differences Frequency Frequency B B B B Time Time F0 B B Time

Streaming based on F0 differences F0 A =100Hz F0 B = F0 A +1.5oct = 283Hz 125 ms

Auditory spectral excitation pattern evoked by bandpass-filtered harmonic complex 45 40 F0 = 400Hz Level (db) 35 30 25 20 15 10 5 1000 2000 3000 4000 5000 6000 Frequency (Hz)

Auditory spectral excitation pattern evoked by bandpass-filtered harmonic complex 45 40 F0 = 200Hz Level (db) 35 30 25 20 15 10 5 1000 2000 3000 4000 5000 6000 Frequency (Hz)

Auditory spectral excitation pattern evoked by bandpass-filtered harmonic complex 45 40 F0 = 100Hz Level (db) 35 30 25 20 15 10 5 1000 2000 3000 4000 5000 6000 Frequency (Hz)

F0 A =100Hz F0 B = F0 A +1.5oct = 283Hz

F0-based streaming with unresolved harmonics is possible... Vliegen & Oxenham (1999); Vliegen, Moore, Oxenham (1999) Grimault, Micheyl, Carlyon et al. (2000)...but the effect is weaker than with resolved harmonics Grimault, Micheyl, Carlyon et al. (2000) 1.0 F0(A) = 250 Hz Probability of "2 streams" response 0.8 0.6 0.4 0.2 0.0 Resolved Resolved Unresolved Unresolved Low region High region -6 0 6 12 18 F0 difference (semitones) From: Grimault et al. (2000) JASA 108, 263-

AM-rate-based streaming Grimault, Bacon, Micheyl (2002) WN AM @ 80 Hz, 50% WN AM @ 160 Hz, 50%

AM-rate-based streaming Grimault, Bacon, Micheyl (2002) fam A =80Hz fam B =fam A +1.5oct=226 Hz

Phase-based streaming Roberts, Glasberg, Moore (2002) Harmonics in sine phase φ(n)=0 Harmonics in alternating-phase φ(n)=0 for odd n φ(n)=0 for even n

Phase-based streaming Roberts, Glasberg, Moore (2002) F0 A =100Hz/SINE F0 B = 100Hz/ALT

Conclusion: The formation of auditory streams is determined primarily by peripheral frequency selectivity, but some streaming may be produced even by sounds that excite the same peripheral channels

Does streaming influence other aspects of auditory perception?

Stream segregation can help... Improved recognition of intervleaved melodies Dowling (1973), Dowling et al. (1987), Hartmann & Johnson (1991), Vliegen & Oxenham (1999), Iverson (1995), Cusack & Roberts (2000), Bey & McAdams (2002) Frequency Target Interferer? Time

Stream segregation can help... Improved (pitch) discrimination of target tones separated by extraneous tones Jones, Macken, Harries (1997) Micheyl & Carlyon (1998) Gockel, Carlyon, & Micheyl (1999) F0 Target Interferers? Time

Stream segregation can harm... Detrimental effect on temporal order identification Bregman & Campbell (1971) Frequency? Time

Stream segregation can harm... Loss of fine temporal relationships Brochard, Drake, Botte, & McAdams (1999) Cusack & Roberts (2000) Roberts, Glasberg, & Moore (2003) Frequency Standard Signal Standard? Signal Time

References Books, reviews on ASA: - Darwin CJ & Carlyon RP (1995) Auditory grouping. In: Hearing (Ed. BJ Moore), Acad. Press, NY - Bregman (1990) Auditory scene analysis. MIT Press, Cambridge MA. Misc: - Darwin CJ, Ciocca V. (1992) Grouping in pitch perception: effects of onset asynchrony and ear of presentation of a mistuned component. J Acoust Soc Am. 91, 3381-3390. - Darwin CJ, Gardner RB. (1986) Mistuning a harmonic of a vowel: grouping and phase effects on vowel quality. J Acoust Soc Am. 79, 838-845. On the neural mechanisms streaming: - Fishman YI et al. (2001) Neural correlates of auditory stream segregation in primary auditory cortex of the awake monkey. Hear Res. 151, 167-187. Computer models of sound segregation: - Cariani PA (2001) Neural timing nets. Neural Netw. 14, 737-753 - de Cheveigne A, et al. (1995) Identification of concurrent harmonic and inharmonic vowels: a test of the theory of harmonic cancellation and enhancement. J Acoust Soc Am. 97, 3736-3748. - Assmann PF, Summerfield Q. (1990) Modeling the perception of concurrent vowels: vowels with different fundamental frequencies. J. Acoust. Soc. Am. 88, 680-697.