Sound Ontology for Computational Auditory Scene Analysis

Similar documents
Interfacing Sound Stream Segregation to Recognition - Preliminar Several Sounds Si

158 ACTION AND PERCEPTION

Music Understanding At The Beat Level Real-time Beat Tracking For Audio Signals

Beat Tracking based on Multiple-agent Architecture A Real-time Beat Tracking System for Audio Signals

AUD 6306 Speech Science

Acoustic Scene Classification

Topic 1. Auditory Scene Analysis

Musical Instrument Recognizer Instrogram and Its Application to Music Retrieval based on Instrumentation Similarity

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

Hidden Markov Model based dance recognition

Speech and Speaker Recognition for the Command of an Industrial Robot

Topics in Computer Music Instrument Identification. Ioanna Karydi

Topic 10. Multi-pitch Analysis

A prototype system for rule-based expressive modifications of audio recordings

Automatic Laughter Detection

THE importance of music content analysis for musical

HUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

MUSICAL INSTRUMENT IDENTIFICATION BASED ON HARMONIC TEMPORAL TIMBRE FEATURES

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

Joint bottom-up/top-down machine learning structures to simulate human audition and musical creativity

A Survey on: Sound Source Separation Methods

Music Segmentation Using Markov Chain Methods

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES

Multiple instrument tracking based on reconstruction error, pitch continuity and instrument activity

2 2. Melody description The MPEG-7 standard distinguishes three types of attributes related to melody: the fundamental frequency LLD associated to a t

A REAL-TIME SIGNAL PROCESSING FRAMEWORK OF MUSICAL EXPRESSIVE FEATURE EXTRACTION USING MATLAB

Keywords Separation of sound, percussive instruments, non-percussive instruments, flexible audio source separation toolbox

Automatic Laughter Detection

Automatic Construction of Synthetic Musical Instruments and Performers

Music Radar: A Web-based Query by Humming System

Improving Frame Based Automatic Laughter Detection

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

Transition Networks. Chapter 5

Application Of Missing Feature Theory To The Recognition Of Musical Instruments In Polyphonic Audio

MUSI-6201 Computational Music Analysis

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

Proposal for Application of Speech Techniques to Music Analysis

On Human Capability and Acoustic Cues for Discriminating Singing and Speaking Voices

Arts, Computers and Artificial Intelligence

Frankenstein: a Framework for musical improvisation. Davide Morelli

A Beat Tracking System for Audio Signals

Melodic Outline Extraction Method for Non-note-level Melody Editing

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors

Department of Electrical & Electronic Engineering Imperial College of Science, Technology and Medicine. Project: Real-Time Speech Enhancement

Phone-based Plosive Detection

... A Pseudo-Statistical Approach to Commercial Boundary Detection. Prasanna V Rangarajan Dept of Electrical Engineering Columbia University

A Bayesian Network for Real-Time Musical Accompaniment

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

SINCE the lyrics of a song represent its theme and story, they

Piano Transcription MUMT611 Presentation III 1 March, Hankinson, 1/15

Music Information Retrieval with Temporal Features and Timbre

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

Computational Modelling of Harmony

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

A System for Acoustic Chord Transcription and Key Extraction from Audio Using Hidden Markov models Trained on Synthesized Audio

1. INTRODUCTION. Index Terms Video Transcoding, Video Streaming, Frame skipping, Interpolation frame, Decoder, Encoder.

Subjective Similarity of Music: Data Collection for Individuality Analysis

Chord Classification of an Audio Signal using Artificial Neural Network

Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies

Efficient Computer-Aided Pitch Track and Note Estimation for Scientific Applications. Matthias Mauch Chris Cannam György Fazekas

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon

EXPLORING THE USE OF ENF FOR MULTIMEDIA SYNCHRONIZATION

Comparison Parameters and Speaker Similarity Coincidence Criteria:

Expressive Singing Synthesis based on Unit Selection for the Singing Synthesis Challenge 2016

Singing voice synthesis based on deep neural networks

Interacting with a Virtual Conductor

A Case Based Approach to the Generation of Musical Expression

Proceedings of Meetings on Acoustics

Figured Bass and Tonality Recognition Jerome Barthélemy Ircam 1 Place Igor Stravinsky Paris France

Automatic music transcription

Computer Coordination With Popular Music: A New Research Agenda 1

Musical Creativity. Jukka Toivanen Introduction to Computational Creativity Dept. of Computer Science University of Helsinki

Predicting Performance of PESQ in Case of Single Frame Losses

Musical Instrument Identification based on F0-dependent Multivariate Normal Distribution

EMERGENT SOUNDSCAPE COMPOSITION: REFLECTIONS ON VIRTUALITY

CTP431- Music and Audio Computing Music Information Retrieval. Graduate School of Culture Technology KAIST Juhan Nam

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

An Audio-based Real-time Beat Tracking System for Music With or Without Drum-sounds

Take a Break, Bach! Let Machine Learning Harmonize That Chorale For You. Chris Lewis Stanford University

Perceptual Considerations in Designing and Fitting Hearing Aids for Music Published on Friday, 14 March :01

Recognising Cello Performers Using Timbre Models

BayesianBand: Jam Session System based on Mutual Prediction by User and System

Enhancing Music Maps

2. Measurements of the sound levels of CMs as well as those of the programs

CPU Bach: An Automatic Chorale Harmonization System

Sentiment Extraction in Music

Recognising Cello Performers using Timbre Models

SYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS

An Introduction to Description Logic I

FULL-AUTOMATIC DJ MIXING SYSTEM WITH OPTIMAL TEMPO ADJUSTMENT BASED ON MEASUREMENT FUNCTION OF USER DISCOMFORT

Outline. Why do we classify? Audio Classification

POLYPHONIC INSTRUMENT RECOGNITION USING SPECTRAL CLUSTERING

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

Study of White Gaussian Noise with Varying Signal to Noise Ratio in Speech Signal using Wavelet

Singer Traits Identification using Deep Neural Network

1. Introduction NCMMSC2009

WE ADDRESS the development of a novel computational

Transcription:

From: AAAI-98 Proceedings. Copyright 1998, AAAI (www.aaai.org). All rights reserved. Sound Ontology for Computational Auditory Scene Analysis Tomohiro Nakatanit and Hiroshi G. Okuno NTT Basic Research Laboratories Nippon Telegraph and Telephone Corporation 3-l Morinosato-Wakamiya, Atsugi, Kanagawa 243-0198, JAPAN okuno@nue.org Abstract This paper proposes that sound ontology should be used both as a common vocabulary for sound representation and as a common terminology for integrating various sound stream segregation systems. Since research on computational auditory scene analysis (CASA) focuses on recognizing and understanding various kinds of sounds, sound stream segregation which extracts each sound stream from a mixture of sounds is essential for CASA. Even if sound stream segregation systems use a harmonic structure of sound as a cue of segregation, it is not easy to integrate such systems because the definition of a harmonic structure differs or the precision of extracted harmonic structures differs. Therefore, sound ontology is needed as a common knowledge representation of sounds. Another problem is to interface sound stream segregation systems with applications such as automatic speech recognition systems. Since the requirement of the quality of segregated sound streams depends on applications, sound stream segregation systems must provide a flexible interface. Therefore, sound ontology is needed to fulfill the requirements imposed by them. In addition, the hierarchical structure of sound ontology provides a means of controlling top-down and bottom-up processing of sound stream segregation. Introduction Sound is gathering attention as important media for multi-medal communications, but is less utilized as input media than characters or images. One reason is the lack of a general approach to recognize auditory events from a mixture of sounds. Usually, people hear a mixture of sounds, and people with normal hearing can segregate sounds from the mixture and focus on a particular voice or sound in a noisy environment. This capability is known as the cocktail party e#ect (Cherry 1953). Perceptual segregation of sounds, called auditoy scene analysis, has been studied by psychoacoust;p onrl Ib aii,u yuyllr y,lryurbw nc~,ohr.nh~,&~nl IG;Ubad.bAICILU roonavrhnro fn,. I I III Ib mnva th,, IIcb~I IVJ. fnrtrr uy +Current address: NTT Multimedia Business Department, nak@mbd.mbc.ntt.co.jp Copyright 01998, American Association for Artificial Intelligence (www.aaai.org). All rights reserved. years. Although many observations have been analyzed and reported (Bregman 1990), it is only recently that researchers have begun to use computer modeling of auditory scene analysis. This emerging research area is called computational auditory scene analysis (CASA) (Brown and Cooke 1992; Cooke et al. 1993; Nakatani, Okuno, and Kawabata 1994a; Rosenthal and Okuno 1998), and its goal is the understanding of an arbitrary sound mixture including non-speech sounds and music. Computers need to be able to decide which parts of a mixed acoustic.-- sifznal ~~v~~.- are..~ relevant ~-- to a =-.-.-----.- oarticnlar ~. nurnose..-r _L- - which part should be interpreted as speech, for example, and which should be interpreted as a door closing, an air conditioner humming, or another person interrupting. CASA focuses on the computer modeling and implementation for the understanding of acoustic events One of its main research topics of CASA is sound stream segregation. In particuiar, CASA focuses on a general model and mechanism of segregating various kinds of sounds, not limited to specific kinds of sounds, from a mixture of sounds. Sound stream segregation can be used as a frontend for automatic speech recognition in real-world environments (Okuno, Nakatani, and Kawabata 1996). As seen in the cocktail-party effect, humans have the ability to selectively attend to a sound from a particular source, even when it is mixed with other sounds. Current automatic speech recognition systems can understand clean speech well in relatively noiseless laboratory environments, but break down in more realistic, noisier environments. Speech enhancement is essential to enable automatic speech recognition to work in such environments. Conventional approaches to speech enhancement are classified as noise reduction, speaker adaptation, and other robustness techniques (Minami and Furui 1995). Speech stream segregation is a novel approach to speech enhancement? and works as the frontend system for automatic speech recognition just as hearing aids for hearing impaired people. Of course, speech stream segregation as a front-end for ASR is the first step toward more robust ASR. The 1004 Understanding Sound

nd rd -m-rsical sound in trumental sound 3ec ach -speaker sex, age, I I orchestra rmonic structure fundamental frequency rmonic structure fundamental frequency -single t tone h AM, FM modulation power spectrum rmonic structure voiced consonant u voiced consonant... -..qchiner;;;rm;tource -natural -... sound... 11 represents the part-of hierarchy of the abstract sound class. Figure 1: Example of sound ontology. next step may be to integrate speech stream segregation and ASR by exploiting top-down and bottom-up processings. Sound ontology is important in musical and speech stream segregation on the following aspects: 1. integrating sound stream segregation systems, 2. interfacing sound stream segregation systems with applications, and 3. integrating bottom-up and top-down processings in sound stream segregation. For example, speech stream segregation system and musical stream, segregation system are combined to develop a system that can recognize both speeches and music from a mixture of voiced announcements and background music. Although the both systems use a harmonic structure of sounds as a cue of segregation, harmonic structures extracted by one system can not be utilized by the other system because the definition of a harmonic structure and the precision of extracted harmonic structure differ. Therefore, sound ontology is needed to share the similar information by the both systems. In addition, sound ontology is expected to play an important role in making the system expandable and scalable., The second example is one of AI challenges, that is, the problem of listening to three simultaneous speeches which is proposed as a challenging problem for AI and CASA (Okuno, Nakatani, and Kawabata 1997). This CASA challenge requires sound stream segregation systems to be interfaced with automatic speech recognition systems. In order to attain better performance of speech recognition, extracted speech streams should fulfill the requirements on the input of automatic speech recognition systems. Some automatic speech recognition systems use power spectrum as a cue of recognition, while others use LPC spectrum. Therefore, speech stream segregation systems should improve the property of extracted speech streams that is required by a particular automatic speech recognition system. If one sound stream segregation system is designed to be interfaced with these two kinds of automatic speech recognition systems, sound ontology may be used as a design model for such a system. The third example is that speech stream segregation should incorporate two kinds of processings, primitive segregation of speech stream and schema-based segregation of speech streams. Primitive segregation may be considered as bottom-up processing by means of lower level properties of sound such harmonic structures, timbre, loudness, AM or FM modulation. Schema- Sound Understanding 1005

based segregation may be considered as top-down processing by means of learned constraints. It includes a memory based segregation that uses a memory of specific speaker s voices, and a semantic based segregation that uses contents of speeches. The important issue is how to integrate the both processing and sound ontology may be used as a driving power of integration. The rest of the paper is organized as follows. Section 2 presents a sound ontology and Section 3 discusses its usage with respect to the above-mentioned three issues. Section 4 presents the ontology-based integration of speech stream and musical stream segregation systems. Section 5 and 6 discuss related work and concluding remarks. Sound Ontology Sound ontology is composed of sound classes, definitions of individual sound attributes, and their relationships. It is defined hierarchically by using the following two attributes: l Part-of hierarchy - a hierarchy based on the inclusion relation in sound l Is-a hierarchy - a hierarchy based on the abstrac- Some part of sound ontology concerning speech and music is shown in Fig. 1. Other parts may be defined on demand of segregating corresponding sounds. In this paper, we focus on speech and musical stream segregation. A box depicts a Part-of hierarchy of basic sound classes for speech and music, which is composed of four layers of sound classes. A sound source in the figure is a temporal sequence of sounds generated by a single sound source. A sound source group is a set of sound sources that share some common characteristics as music. A single tone is a sound that continues without any durations of silence, and it has some low-level attributes such as harmonic structure. In each layer, an upper class is composed of lower classes which are components that share some common characteristics. For example, a harmonic stream is composed of frequency components that have harmonic relationships. The Is-a hierarchy can be constructed using any abstraction level. For example, voice, female voice, the voice of a particular woman, and the woman s nasal voice form an Is-a hierarchy. (Not specified in Fig. 1.) With sound ontology, each class has some attributes, such as fundamental frequency, rhythm, and timbre. A lower class in the Is-a hierarchy inherits the attributes of its upper classes by default unless another special specification is given. in other words, an abstract sound class has attributes that are common to more concrete sound classes. In addition, an actually generated sound, such as uttered speech, is treated as an instance of a sound class. In this representation, segregating a sound stream means generating an instance of a sound class and extracting its attributes from an input sound mixture. Proposed Usage of Sound Ontology Ontology-Based Integration Sound stream segregation should run incrementally, not in batch, because it is usually combined with applications and is not a stand-alone system. For incremental processing, we exploit not only sharing of information but also sharing of processing. Therefore, for integration of existing segregation systems they are first decomposed into processing modules by using the sound ontology as a common specification. In this paper, we take an example of integrating speech and musical stream segregation. The rough procedure of modularization is as follows. First, existing segregation systems are divided into processing modules, each of which segregates a class of sound in sound ontology, such as, harmonic structure, voiced segment, and musical note. Then, these modules are combined to segregating streams according to their identified sound types. Obviously, such an integration requires the proce- dure of inter&,inn b&wen - different - -_- 1 kinds -L nf -. mndnle!s. To specify the relationships between sounds, a relation class is defined between two sound classes. It is represented by a pair of sound classes, such as [speech, musical note]. A relation class has the same two hierarchical structures as sound classes defined as follows: if and only if both classes of a relation class are at a level higher than those of another relation class in the Is-a hierarchy (or in the Part-of hierarchy), the former is at a higher level. In the Is-a hierarchy, processing modules and interaction modules are inherited from a upper level to a lower level unless other modules are specified at some lower class. Ontology-based sound stream segregation is executed by generating instances of sound classes and extracting their attributes. This process can be divided into four tasks: 1. find new sound streams in a sound mixture, 2. identify classes of individual sound streams, 3. extract attributes of streams according to their sound classes, and 4. extract interaction between streams according to their relation classes. Since classes of sound streams are usually not given in advance, sound streams are treated initially as instances of abstract classes. These streams are refined to more concrete classes as more detailed attributes are extracted. At the same time, these streams are identified as components of stream groups at a higher level of the Part-of hierarchy, such as the musical stream. As a result, the attributes of these streams are extracted more precisely by using operations specific to 1006 Understanding Sound

concrete classes and by using attributes group streams. extracted for Sound Ontology Based Interfacing The CASA challenge requires speech stream segregation systems to interface with automatic speech recognition systems. First, speech stream segregation system that extracts each speech stream from a mixture of sounds. Then each extracted speech stream is recognized by a conventional automatic speech recognition system. As such an approach, Binaural Harmonic Stream Segregation (Bi-HBSS) was used as a speech stream segregation (Okuno, Nakatani, and Kawabata 1996). By taking the structure of Vowel (V) + Consonant (C) + Vowel (V) of speech into consideration, a speech stream was extracted by the foiiowing three successive subprocesses: 1. Harmonic stream fragment extraction, 2. Harmonic grouping, and 3. Residue substitution. The first two subprocesses reconstruct the harmonic parts of speech and calculates the residue by subtracting all extracted harmonic parts for the input. Since any major attributes for extracting non-harmonic parts have not been known yet, it is reasonable to substitute the residue for non-harmonic parts. Since Bi-HBSS takes binaural sounds (a pair of sounds recorded by a dummy head microphone) as inputs, it uses the direction of the sound source for segregation. They reported that the recognition performance with the Hidden Markov Model based automatic speech recognition system called HMM-LR (Kita, Kawabata, and Shikano 1990) is better when the residue of all directions is substituted than when the residue of the sound source direction (Okuno, Nakatani, and Kawabata 1996). This interface is not, however, valid for another automatic speech recognition system. Our experiment with one of the popular commercial automatic speech recognition systems, HTK (Young et al. 1996), shows that the interface won t work well and that if the residue of the sound source direction is used by residue substitution the recognition performance is improved. That is, the reason why Bi-HBSS does not work well with HTM is that the cues of recognition used by HMM-LR and HTK differs. HMM-LR uses only power spectrum and ignores input signals of weak power, while HTK uses LPC spectrum and the power of input is automatically normalized. Therefore, weak harmonic structures inciuded in the residue that are usuauy ignored by HMM- LR causes HTK s poor recognition performance. Thus, speech stream segregation systems should generate an output appropriate to successive processing. Since the system architecture of sound stream segregation is constructed on the basis of sound ontology, adaptive output generation is also realized by the same architecture. c Part-of hierarchy a Is-a hierarchy (Sound class In inner circle Is a kind of sound class in outer circle) Figure 2: Hierarchical structure of sound ontology. Integration of Bottom-up and Top-down Processing The representation of single tone components also has analysis conditions under which the system extracts the attributes of individual components. This is because the quantitative meaning of sound attributes may differ by analysis means, such as FFT, auditory filter bank, LPC analysis, and by analysis parameters. For example, interface agents use a common abstract terminology harmonic structure in exchanging infor- mation on voiced segments and on musical notes because voiced segments and musical notes both are a kind of harmonic structure based on sound ontology. Of course, voiced segments and musical notes do not ontir-1-r ULLYUV J hsxm Y.S. the AA ES~P Y-A&A.. attrihjltes CYYYIAYL.YVY. Tn &&. fact. *-, the.._ t,ime **ss1 nat- II terns of voiced segments are much more complex than those of musical notes, so that some harmonic components appear and/or disappear according to their phoneme transitions, while the piano exhibits quite marked deviations from harmonicity during its attack period. Moreover, many frequency components of a musical note are often impossible to discriminate from those of other notes and thus are treated as a set of harmonic sounds, because they continuously overlap each other in a chord. These attributes should be handled somewhat differently in each system. Therefore, the specific sound classes of a harmonic structure (i.e., a voiced segment or a musical note) also have attributes specific to the classes. Thus, sound ontology enables segregation systems to share information on extracted sound streams by providing common representation of sounds and correspondence between different analysis conditions. Sound Understanding 1007

processings specific For simplicity, we adopt a heuristics on the fundamental frequency pattern of a harmonic stream. A Harmonic sound is recognized as a musical note if the standard deviation of the fundamental frequency, denoted gfo, for n millisecond from the beginning of the stream satisfies the following inequality: qj.6 < c, where n and c are constants, and & is the average fundamental frequency. Otherwise, the sound is treated as a part of speech. Figure 3: Common and specific processing moduies for speech and musical stream segregation. Ontology-Based Speech and Musical Stream Segregation a,..,a,,.-,,,-..+:.-.,,,#.,,,,:,, -,.4..1,, D LlLIU mzg;r CfjcLbl U p L L.c;aDlrrg lll U U1I;D Speech and musical stream segregation systems are integrated on the basis of sound ontology. As mentioned before, two systems, Bi-HBSS and OPTIMA (Kashino et al. 1995) are decomposed into primitive processing modules as follows: l l Processing modules for speech stream segregation: 1. voice segregation, 2. unvoiced consonants (residual signal after di harmonic sounds are subtracted from input signal) extraction, 3. sequential grouping of voice, and 4. unvoiced consonant interpolation. Processing modules for musical stream segregation: 1. note extraction, 2. identification of sound sources (instruments), 3. rhythm extraction, 4. code extraction, and 5. knowledge sources that store musical information statistics. The relationship between processing modules are l L~ ehnmm 111 in Bier A Lb v. 9 Snm~ u,,ial -* DTP r~rnmnn -v.^..y -) while others are specific to speech or musical stream segregation. Three interfacing modules are designed to integrate the modules; discriminator of voice and musical notes, primitive fundamental frequency tracer, and mediator of single tone tracers. Discriminator of voice and musical note Many signal processing algorithms have been presented to distinguish SoUrid classes, such as discriminant any- ysis and the subspace method (Kashino et al. 1995). Primitive fundamental frequency tracer A primitive fundamental frequency tracer is redesigned, although such a tracer is implicitly embedded in -- Bi-HBSS -----i or Optima. The primitive fundamental frequency tracer extracts a harmonic structure at each time frame as follows: 1. the fundamental frequency at the next time frame is predicted by linearly extending its transition pattern, and 2. a harmonic structure whose fundamental frequency is in the region neighboring the predicted one in the next frame is tracked as the fundamental frequency. Since musical notes have a specific fundamental frequency pattern, that is, musical scale, it can be used as a constraint on musical stream. If a sound being traced is identified as a musical stream, the primitive fundamental frequency tracer is replaced by musical ffi~da~-e~?.~i frpqnenc)j traceri First. ) fllndr.mf?nt.7%1 --_--- -_---- -- frequency is predicted by calculating its average fundamental frequency, and the search region for the fundamental frequency is restricted to a narrower region than the default. As a result of using stricter constraints, more precise and less ambiguous fundamental frequencies of musical notes can be extracted. Mediator of single sound tracers Two types of interaction modules between single sound tracers are designed in order to assign individual harmonic components to a musical or speech stream segregation system: one module in the relation class [harmonic structure, harmonic structure], and another in the relation class [musical note, musical note]. The interaction module in the relation class [harmonic structure, harmonic structure] defines defauit interaction, because [harmonic structure, harmonic structure] is the parent relation class of [musical note, musical note] and so on. This interaction module decompses overlapping frequency components into streams in the same way as Bi-HBSS (Nakatani, Okuno, and Kawabata 199513). The other module in the relation class [musical note, musical note] leaves overlapping frequency components shared between sinie sound tracers, because such decomposing is quite difficult (Kashino et al. 1995). 1008 Understanding Sound

Some additional modules such as an unvoiced consonant interpolation module and rhythm and chord extraction modules are under development to improve the performance of segregation of all parts of voice and music. (a) Input mixture of music (Auld Lang Syne) and narration (female voice) (b) Segregated harmonic stream corresponding to narration (female voice) (c) Segregated harmonic stream corresponding to music (Auld Lang Syne) Figure 4: Spectrograms of input mixture and segregated harmonic streams. The other relation classes such as [speech, speech] are not explicitly specified, but inherit the interaction modules of the default relation class [harmonic structure, harmonic structure]. Results of speech and musical segregation Evaluation is performed with a mixture of the musical piece Auld Lang Syne played on flute and piano (sound synthesized by a sampler) and a narration by a woman uttering a Japanese sentence. The spectrograms of the input and extracted two harmonic streams are shown in Fig. 4. Although the prototype system takes monaural sounds as input instead of binaural sounds, its performance of segregation is better in +^r-^ bplll.m I..A- pua.11,:+,.i. rxl J.3.-... ^..,. auu,,a -.,-.,+...I qjclblcu UI3lJ LbI II A: +,..+:,., rl,.. bllau CL-C LIIa?LI of Bi-HBSS. Of course, when fundamental frequencies of voice and music cross each other, two sounds are not segregated well. This problem is unavoidable if harmonics is the only clue for segregation. As mentioned before, Bi-HBSS overcomes the problem by using directional information. Related work Nawab et al. proposed unified terminology as universal representation of speech and sporadic environmental sounds, and developed a spoken digit recognition system using the IPUS architecture, a variant of blackboard architecture (Nawab et al 1995). The idea of combining processing modules based on unified terminology is quite similar to our ontology-based sound stream segregation. Since a module is implemented as a separate knowledge source, the processing is performed in batch and incremental processing is difficult. We think that HEARSAY-II like usage of blackboard architecture which each knowledge source has a limited capability would require a sound ontology Minsky suggested a musacal CYC project, a musical common sense database, to promote researches on understanding music (Minsky and Laske 1992). However, a collection of various musical common sense databases m~rr IUruJ ho x. mnv.a lll lx. o~c;lxr bu.xij tn Y s-nnctvmt ~ ~AU IU th3.n Yllcull u I mnnnl;th III II &I II h,.cm LAUbb database, and we expect that a music ontology will play an important role in combining musical common sense databases. Conclusion Sound ontology is presented as a new framework for integrating existing sound stream segregation systems, interfacing sound stream segregation systems with applications, and integration of top-down and bottom-up processings. That is, sound ontology specifies a common representation of sounds and a common specification of sound processing to combine individual sound stream segregation systems. We believe that sound ontology is a key to an expansible CASA system because it provides a systematic and comprehensive principle of integrating segregation technologies. Future work includes design of more universal sound ontology, full-scale implementation of speech and musical segregation systems, and attacking of the CASA challenge. Last but not least, controling bottomup processing with top-down processing along with a sound ontology is an important and exciting future work. Acknowledgments We thank Drs. Kunio Kashino, Takeshi Kawabata, Hiroshi Murase, and Ken ichiro Ishii of NTT Basic Research Labs, Dr. Masataka Goto of Electra Technology Lab, and Dr. Hiroaki Kitano of Sony CSL for their valuable discussions. Sound Understanding 1009

References Bregman, A.S. 1990. Auditory Scene Analysis - the Perceptual Organization of Sound. MIT Press. Brown, G.J., and Cooke, M.P. 1992. A computational model of auditory scene analysis. In Proceedings of Intern1 Conf. on Spoken Language Processing, 523-526. Cherry, E.C. 1953. Some experiments on the recognition of speech, with one and with two ears. Journal of Acoustic Society of America 25:975-979. Cooke, M.P., Brown, G.J., Crawford, M., and Green, P. 1993. Computational Auditory Scene Analysis: listening to several things at once. Endeawour, 17(4):186-190. &&no, K., Nakadai, K., Kinoshita, T., and Tanaka, H. 1995. Organization of Hierarchical Perceptual Sounds: Music Scene Analysis with Autonomous Processing Modules and a Quantitative Information Integration Mechanism, In Proceedings of 14th International Joint Conference on Artajicial Intelligence (IJCAI-95), vol.l:173-178, IJCAI. Kita, K., Kawabata, T., and Shikano, K. 1990. HMM continuous speech recognition using generalized LR parsing. Transactions of Information Processing Society of Japan, 31(3):472-480. Lesser, V., Nawab, S.H., Gallastegi, I., and Klassner, F. 1993. IPUS: An Architecture for Integrated Signal Processing and Signal Interpretation in Complex Environments. In Proceedings of Eleventh National Con- Jc arlce *-----.. I ^- IrruJLua A -z:ltc^:- I 1rLLecr ge rlce T-1.111 ^^_^^^ (filil-lil-yj,) /A 1 T AO, 43--LoJ, a * nrr? AAAI. Minami, Y., and Furui, S. 1995. A Maximum Likelihood Procedure for A Universal Adaptation Method based on HMM Composition. In Proceedings of 1995 International Conference on Acoustics, Speech and Signal Processing, vol.l:129-132, IEEE. Minsky, M., and Laske, 0. 1992. Forward: Conversation with Marvin Minsky, In Balaban, M.! Ebcioglu, K., and Laske, 0. eds. Understanding Music with AI: Perspectives on Music Cognition, ix-xxx, AAAI Press/MIT Press. Nakatani, T., Okuno, H.G., and Kawabata, T. 1994a. Auditory Stream Segregation in Auditory Scene Analysis with a Multi-Agent System. In Proceedings of ram 4.,-L,- lval$onal AT..:...-).-..A-... i- bonjerence on A i.i@qi InteEizgeiice (AAAI-94), 100-107, AAAI. Nakatani, T., Okuno, H.G., and Kawabata, T. 1995b. Residue-driven architecture for Computational Auditory Scene Analysis. In Proceedings of 14th International Joint Conference on Artificial Intelligence (IJCAI-95), vol.l:165-172, IJCAI. Nawab, S.H., Espy-Wilson, C.Y., Mani, R., and Bitar, N.N. 1995. Knowledge-Based analysis of speech mixed with sporadic environmental sounds. In Rosenthal and Okuno, eds. Working Notes of IJCAI-95 Workshop on Computational Auditory Scene Analysis, 76-83. Okuno, H.G., Nakatani, T., and Kawabata, T. 1996. Interfacing Sound Stream Segregation to Speech Recognition Systems - Preliminary Results of Listening to Several Things at the Same Time. In Proceedings 0s i%h lliationai Conference on Artijiciai intelligence (AAAI-96), 1082-1089. Okuno, H.G., Nakatani, T., and Kawabata, T. 1997. Understanding Three Simultaneous Speakers. In Proceedings of 15th International Joint Conference on Artificial Intelligence (IJCAI-97), Vol.l:30-35. Ramalingam, C.S., and Kumaresan, R. 1994. Voicedspeech analysis based on the residual interfering signal canceler (?_ISC) 2~m-wifhm D---Y..--*. Tn &AA* Pmrwdimnc 8 w, yv YJ nf L./VT 1491 International Conference on Acoustics, Speech, and Signal Processing, 473-476, IEEE. Rosenthal, D., and Okuno, H.G. eds. 1998. Computational Auditory Scene Analysis, NJ.:Lawrence Erlbaum Associates, (in print). Young, S., Jansen, J., Ode& J.! Ollanson, D.! and Woodland, P. 1996. the HTK Book for HTK V2.0. Entropic Combridge Reseach Lab. Inc. 16 10 Understanding Sound