AUDIO FEATURE EXTRACTION AND ANALYSIS FOR SCENE SEGMENTATION AND CLASSIFICATION

Similar documents
Reducing False Positives in Video Shot Detection

A Framework for Segmentation of Interview Videos

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

DETECTION OF SLOW-MOTION REPLAY SEGMENTS IN SPORTS VIDEO FOR HIGHLIGHTS GENERATION

VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS. O. Javed, S. Khan, Z. Rasheed, M.Shah. {ojaved, khan, zrasheed,

Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting

Browsing News and Talk Video on a Consumer Electronics Platform Using Face Detection

2. AN INTROSPECTION OF THE MORPHING PROCESS

Voice & Music Pattern Extraction: A Review

Transcription of the Singing Melody in Polyphonic Music

MUSI-6201 Computational Music Analysis

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

... A Pseudo-Statistical Approach to Commercial Boundary Detection. Prasanna V Rangarajan Dept of Electrical Engineering Columbia University

Story Tracking in Video News Broadcasts. Ph.D. Dissertation Jedrzej Miadowicz June 4, 2004

Speech and Speaker Recognition for the Command of an Industrial Robot

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

Topic 10. Multi-pitch Analysis

Automatic Laughter Detection

CS229 Project Report Polyphonic Piano Transcription

A repetition-based framework for lyric alignment in popular songs

Reconstruction of Ca 2+ dynamics from low frame rate Ca 2+ imaging data CS229 final project. Submitted by: Limor Bursztyn

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

Hidden Markov Model based dance recognition

Department of Electrical & Electronic Engineering Imperial College of Science, Technology and Medicine. Project: Real-Time Speech Enhancement

Music Radar: A Web-based Query by Humming System

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

Predicting Performance of PESQ in Case of Single Frame Losses

Improving Frame Based Automatic Laughter Detection

Automatic Rhythmic Notation from Single Voice Audio Sources

Speech To Song Classification

Music Database Retrieval Based on Spectral Similarity

GCT535- Sound Technology for Multimedia Timbre Analysis. Graduate School of Culture Technology KAIST Juhan Nam

Automatic Piano Music Transcription

Shot Transition Detection Scheme: Based on Correlation Tracking Check for MB-Based Video Sequences

EXPLORING THE USE OF ENF FOR MULTIMEDIA SYNCHRONIZATION

Automatic Music Clustering using Audio Attributes

MUSIC TONALITY FEATURES FOR SPEECH/MUSIC DISCRIMINATION. Gregory Sell and Pascal Clark

Audio-Based Video Editing with Two-Channel Microphone

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

Deep Neural Networks Scanning for patterns (aka convolutional networks) Bhiksha Raj

Automatic Construction of Synthetic Musical Instruments and Performers

Singer Traits Identification using Deep Neural Network

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video

Incorporating Domain Knowledge with Video and Voice Data Analysis in News Broadcasts

hit), and assume that longer incidental sounds (forest noise, water, wind noise) resemble a Gaussian noise distribution.

Phone-based Plosive Detection

A Parametric Autoregressive Model for the Extraction of Electric Network Frequency Fluctuations in Audio Forensic Authentication

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

Lecture 9 Source Separation

Getting Started. Connect green audio output of SpikerBox/SpikerShield using green cable to your headphones input on iphone/ipad.

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)

Automatic Laughter Detection

Evaluation of Automatic Shot Boundary Detection on a Large Video Test Suite

Tempo and Beat Analysis

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

A Parametric Autoregressive Model for the Extraction of Electric Network Frequency Fluctuations in Audio Forensic Authentication

Analysis of local and global timing and pitch change in ordinary

Wipe Scene Change Detection in Video Sequences

An Examination of Foote s Self-Similarity Method

Multi-modal Analysis for Person Type Classification in News Video

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

CSC475 Music Information Retrieval

Study of White Gaussian Noise with Varying Signal to Noise Ratio in Speech Signal using Wavelet

Enhancing Music Maps

Retrieval of textual song lyrics from sung inputs

EE373B Project Report Can we predict general public s response by studying published sales data? A Statistical and adaptive approach

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

Chord Classification of an Audio Signal using Artificial Neural Network

Measurement of overtone frequencies of a toy piano and perception of its pitch

Topics in Computer Music Instrument Identification. Ioanna Karydi

Detection and demodulation of non-cooperative burst signal Feng Yue 1, Wu Guangzhi 1, Tao Min 1

Topic 4. Single Pitch Detection

Subjective Similarity of Music: Data Collection for Individuality Analysis

Pitch Based Sound Classification

Keywords Separation of sound, percussive instruments, non-percussive instruments, flexible audio source separation toolbox

Supervised Learning in Genre Classification

MPEG has been established as an international standard

Robert Alexandru Dobre, Cristian Negrescu

THE importance of music content analysis for musical

Region Adaptive Unsharp Masking based DCT Interpolation for Efficient Video Intra Frame Up-sampling

A Music Retrieval System Using Melody and Lyric

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis

Principles of Video Segmentation Scenarios

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

An Efficient Low Bit-Rate Video-Coding Algorithm Focusing on Moving Regions

Efficient Implementation of Neural Network Deinterlacing

Automatic music transcription

Fast MBAFF/PAFF Motion Estimation and Mode Decision Scheme for H.264

Detecting Musical Key with Supervised Learning

WHAT'S HOT: LINEAR POPULARITY PREDICTION FROM TV AND SOCIAL USAGE DATA Jan Neumann, Xiaodong Yu, and Mohamad Ali Torkamani Comcast Labs

Analysis, Synthesis, and Perception of Musical Sounds

Hidden melody in music playing motion: Music recording using optical motion tracking system

Speech Recognition and Signal Processing for Broadcast News Transcription

Transcription:

AUDIO FEATURE EXTRACTION AND ANALYSIS FOR SCENE SEGMENTATION AND CLASSIFICATION Zhu Liu and Yao Wang Tsuhan Chen Polytechnic University Carnegie Mellon University Brooklyn, NY 11201 Pittsburgh, PA 15213 {zhul,yao}@vision.poly.edu tsuhan@ece.cmu.edu Abstract Understanding of the scene content of a video sequence is very important for content-based indexing and retrieval of multimedia databases. Research in this area in the past several years has focused on the use of speech recognition and image analysis techniques. As a complimentary effort to the prior work, we have focused on using the associated audio information (mainly the nonspeech portion) for video scene analysis. As an example, we consider the problem of discriminating five types of TV programs, namely commercials, basketball games, football games, news reports, and weather forecasts. A set of low-level audio features are proposed for characterizing semantic contents of short audio clips. The linear separability of different classes under the proposed feature space is examined using a clustering analysis. The effective features are identified by evaluating the intracluster and intercluster scattering matrices of the feature space. Using these features, a neural net classifier was successful in separating the above five types of TV programs. By evaluating the changes between the feature vectors of adjacent clips, we also can identify scene breaks in an audio sequence quite accurately. These results demonstrate the capability of the proposed audio features for characterizing the semantic content of an audio sequence. - 1 -

1. Introduction A video sequence is a rich multimodal information source, containing speech, audio, text (if closed caption is available), color patterns and shape of imaged objects (from individual image frames), and motion of these objects (from changes in successive frames). Although the human being can quickly interpret the semantic content by fusing the information from different modalities, computer understanding of a video sequence is still in a quite primitive stage. With the booming of the Internet and various types of multimedia resources, there is a pressing need for efficient tools that enable easier dissemination of audiovisual information by the human being. This means that multimedia resources should be indexed, stored and retrieved in a way similar to the way that a human brain processes them. This requires the computer to understand their contents before all other processing. Other applications requiring scene understanding include spotting and tracing of special events in a surveillance video, active tracking of special objects in unmanned vision systems, video editing and composition, etc. The key to understanding of the content of a video sequence is scene segmentation and classification. Research in this area in the past several years has focused on the use of speech and image information. These include the use of speech recognition and language understanding techniques to produce keywords for each video frame or a group of frames [1, 2], the use of image statistics (color histograms, texture descriptors and shape descriptors) for characterizing the image scene [3-5], detection of large differences in image intensity or color histograms for segmentation of a sequence into groups of similar content [6, 7], and finally detection and tracking of a particular object or person using image analysis and object recognition techniques [8]. Another related work is to create a summary of the scene content by creating a mosaic of the imaged scene with trajectories of moving objects overlaying on top [9], by extracting key frames in a video sequence that are - 2 -

representative frames of individual shots [10], and by creating a video poster and an associated scene transition graph [11]. Recently several researchers have started to investigate the potential of analyzing the accompanying audio signal for video scene classification [12-15]. This is feasible because, for example, the audio in a football game is very different from that in a news report. Obviously, audio information alone may not be sufficient for understanding the scene content, and in general, both audio and visual information should be analyzed. However, because audio-based analysis requires significantly less computation, it can be used in a preprocessing stage before more comprehensive analysis involving visual information. In this paper, we focus on audio analysis for scene understanding. Audio understanding can be based on features in three layers: low-level acoustic characteristics, intermediate-level audio signatures associated with different sounding objects, and high level semantic models of audio in different scene classes. In the acoustic characteristics layer, we analyze low level generic features such as loudness, pitch period and bandwidth of an audio signal. This constitutes the pre-processing stage that is required in any audio processing system. In the acoustic signature layer, we want to determine the object that produces a particular sound. The sounds produced by different objects have different signatures. For example, each music instrument has its own impulse response when struck. Basketball bouncing is different from a baseball hit by the bat. By storing these signatures in a database and matching them with an audio segment to be classified, it is possible to categorize this segment into one object class. In the high level modelbased layer, we make use of some a prior known semantic rules about the structure of audio in different scene types. For example, there is normally only speech in news report and weather forecast, but in a commercial, usually there is always a music background, and finally, in a sports program there exists a prevailing background sound that consists of human cheering, ball bouncing - 3 -

and music sometimes. Saraceno and Leonardi presented a method for separating silence, music, speech and noise clips in an audio sequence [12], and so did Pfeiffer, et al. in [13]. These can be considered as low-level classification. Based on these classification results, one can classify the underlying scene based on some semantic models that govern the composition of speech, music, noise, etc. in different scene classes. In general, when classifying an audio sequence, one can first find some low-level acoustic characteristics associated with each short audio clip, and then compare it with those pre-calculated for different classes of audio. Obviously classification based on these low-level features alone may not be accurate, but the error can be addressed in a higher layer by examining the structure underlying a sequence of continuous audio clips. This tells us that the very first and crucial step for audio-based scene analysis is to determine appropriate features that can differentiate audio clips associated with various scene classes. This is the focus of the present work. As an example, we consider the discrimination of five types of TV programs: commercials, basketball games, football games, news and weather reports. To evaluate the scene discrimination capability of these features, we analyze the intra- and inter-class scattering matrices of feature vectors. To demonstrate the effectiveness of these features, we apply them to classify audio clips extracted from above TV programs. Towards this goal, we explore the use of neural net classifiers. The results show that an OCON (One Class One network) neural network can handle this problem quite well. To further improve the scene classification accuracy, more sophisticated techniques operating at a level higher than individual clips are necessary. This problem is not addressed in this paper. We also employ the developed features for audio sequence segmentation. Saunders [16] presented a method to separate speech from music by tracking the change of the zero crossing rate, and Nam and Tewfik [14] proposed to detect sharp temporal variations in the power of the subband signals. Here, we propose to use the changes in the feature vector to detect scene transitions. - 4 -

The organization of this paper is as follows. In Section II, we describe all the audio features we have explored. Analysis of the feature space is presented in Section III. In Sections IV and V, we show the applications of the developed features for scene classification and segmentation. Experimental results are provided within each section. Finally, Section VI concludes the paper by summarizing the main results and presenting remaining research issues. II. Audio Feature Analysis There are many features that can be used to characterize audio signals. Generally they can be separated into two categories: time-domain and frequency-domain features. In this section, we describe several audio features that we have explored. In our experiment, the audio signal is sampled at 22KHz, and divided into clips of one second long. Feature analysis is conducted on each clip (i.e. a feature vector is calculated for each clip). These clip-level features are computed based on frame-level features, which are calculated over overlapping short intervals known as frames. Each frame contains 512 samples shifted by 128 samples from the previous frame. Figure 1 illustrates the relation between clips and frames. Features Derived from Volume Contour The volume distribution of an audio clip reveals the temporal variation of the signal s magnitude, which is important for scene classification. Here, we use the root mean square (RMS) of the signal magnitude within each frame to approximate the volume of that frame. Specifically, the volume of the n-th frame is calculated by: 1 2 v ( n) = s n ( i), N N 1 i= 0 where s n (i) is the i-th sample in the n-th frame audio signal and N is the frame length. - 5 -

Note that the volume of an audio signal depends on the gain of the recording and digitizing system. Therefore, the mean volume of a clip does not necessarily reflect the scene content. On the other hand, the temporal variation of the volume in a clip does. Figure 2 gives the waveforms and Figure 3 shows the volume contours of three audio clips: a commercial, a basketball game and a news report. The characteristics of the football sequence are similar to basketball and so are the weather forecasts to news reports. From these plots, we know that the volume of these three audio clips have different distributions. To measure the temporal variation of an audio clip s volume, we define two time domain features based on volume distribution. The first one is the volume standard deviation (VSTD), which is the standard deviation of the volume over a clip, normalized by the maximum volume in the clip. The second one is the volume dynamic range (VDR) defined as VDR = (max( v) min( v))/ max( v), where min(v ) and max(v ) represent the minimum and maximum volume within an audio clip. We have found that, in sports programs, there is a nearly constant level of the background sound, and therefore the volume does not change a lot. On the other hand, in news and weather reports, there are silent periods between speech, so the VSTD and VDR are much higher. Obviously these two features are correlated, but we found that they do carry some independent information about the scene content. To detect frames that are silent, we compare the volume and zero crossing rate (ZCR - the times that an audio waveform crosses the zero axis) of each frame to some preset thresholds. If both volume and ZCR are less than the thresholds, the frame is declared to be silent. The use of ZCR in addition to the volume can prevent the low energy unvoice speech frames from being classified to be silent. In our simulation, the signal waveform is digitized to a range of -127 to 128 and the maximum volume is about 40. Based on the distribution of the volume and ZCR of the audio signals in our database (see below), we found that a threshold of 3 for volume and a threshold of 50 for - 6 -

ZCR worked well. From the result of silence detection, we calculate the non-silence-ratio (NSR), which is the ratio of the non-silent interval to the entire clip. We found that this ratio varies quite significantly in different video sequences. In news reports there are regular pauses in the reporter s speech; on the other hand in commercials there are always some background music which results in a higher NSR. It is known that speech has a characteristic energy modulation peak around 4Hz syllabic rate [17]. We define the frequency component of the volume contour around 4Hz (FCVC4) as FCVC4 = 0 where C(ω) is the Fourier transform of volume contour of the current clip and W(ω) is a triangular window function centered at 4Hz. Clips composed of speech tend to have higher FCVC4 than those composed of music or noise. Features Derived from Pitch Contour Pitch is the fundamental period of an audio waveform, and is an important parameter in the analysis and synthesis of speech signals. In an audio signal, we can still use pitch as a low-level feature to characterize the periodicity of waveforms in different audio signals. Among the many available pitch determination algorithms, we choose the one that uses the short time Average Magnitude Difference Function (AMDF) to determine the pitch of each frame. The AMDF is defined as N l 1 W ( ω) C( ω) sn ( i + l) sn ( i) i= 0 γ ( l) = N l. 0 C( ω) 2 2 dω dω - 7 -

Figure 4 shows the AMDF of an audio frame. The algorithm for pitch detection from AMDF follows that in [18]. The idea is to find the first valley point in the AMDF from left to right in the search range. Here a valley is defined as a local minimum that satisfies additional constraints in terms of its value relative to the global minimum and its curvature. For example, the AMDF in Figure 4 has two valleys, and the pitch period is the distance between the origin and the first valley. Normally such valleys exist in voice or music frames and vanish in noise or unvoice frames. Here we set the search range to be from 2.3 ms to 15.9 ms, which is the pitch range of normal human speech. After computing the pitch of each frame, we obtain a pitch contour for the entire clip. For the intervals for which no pitch is found, the pitch is assumed to be zero. The pitch contour obtained by this method may not match perfectly the real pitch contour. In some short time intervals, the pitch value diverges from the real contour, sometimes twice or half of the real pitch. A median filter is applied to this contour to smooth out falsely detected pitches. Figure 5 gives the pitch tracks of the same three audio clips as used in Figure 2. In the commercial clip, there exists music background with overlapping notes and the detected pitch at a particular frame depends on which note is stronger. Therefore, the pitch track stays flat for short intervals and there exist both high and low pitch periods. In the basketball clip, since there is significant background noise, the pitch track is very rough, rarely with a smooth region. In the news clip, the pitch track is smooth and lasts relatively long. The intervals between two smooth tracks correspond to silence/unvoice period. We have found that the pitch level is primarily influenced by the speaker (male or female) in the scene rather than by the scene content. On the other hand, the dynamics of the pitch contour appears to reveal scene content more. We therefore use the pitch standard deviation (PSTD) as another audio feature. Based on the pitch estimation results, we can also detect which frame corresponds to voice or music since continuous voice or music frames usually yield a smooth pitch track. The voice-or- - 8 -

music ratio (VMR) is used as another audio feature, which is defined as the ratio of the length of the voice or music frames to the entire audio clip. A frame is considered to be voice or music if its associated pitch track is longer than 50ms. Note that some of the music frames with too high or too low pitch are missed in our pitch detection algorithm because the search range is set according to the human speech. If there is no pitch detected in a frame that is not silent, the frame is considered as noise or unvoice. The noise-or-unvoice ratio (NUR) is defined as the ratio of the length of the noise or unvoice frame to the entire audio clip. Frequency Domain Features To obtain frequency domain features, we first calculate the spectrogram of an audio clip, which is a 2D plot of the short-time Fourier transform (over each audio frames) along the time axis. Let S i (ω) represents the short-time Fourier transform of the ith frame. We define the frequency centroid, C (i), and bandwidth, B(i) of this frame as: C( i) π 0 = π ω S 0 ( ω) S ( ω) i i 2 2 dω, dω 2 B ( i) π ( ω C( i)) 0 = π 0 2 S ( ω) i S ( ω) 2 i dω 2 dω. Similar features have been proposed for audio classification in [19]. Figure 6 shows the spectrogram of the three audio clips given in Figure 3. Figure 7 and Figure 8 show the contours of the frequency centroid and bandwidth computed based on the spectrograms. The zero regions in the contour correspond to silent frames. From these figures, we can see that the basketball clip s frequency centroid is high and has bigger dynamic range, on the other hand, the news clip has low frequency centroid and bandwidth during the voice period and high centroid and bandwidth during the unvoice period. In the commercial clip, there is a continuous music background, so the frequency centroid and bandwidth contours are quite smooth. - 9 -

Since the energy distribution in different frequency bands varies quite significantly among different audio signals, we also use ratios of the energies in different subbands to the total energy as frequency domain features, which are referred to as subband energy ratios. Considering the perceptual property of human ears, we divide the entire frequency band into four subbands, each consists of the same number of critical bands which represent cochlear filters in the human auditory model [20]. Specifically, the frequency ranges for the four subbands are 0-630Hz, 630-1720Hz, 1720-4400Hz and 4400-11025Hz. Figure 9 shows the 4 subband energy ratio contours of the three audio clips given in Figure 3. The four contours in the commercial clip are rather smooth, on the other hand, the contours in basketball clip vary a lot. The energy ratio of subband 1 in the news clip is much higher than those of the other subbands. Considering that the four subband ratios sum to 1, we use the first three ratios as features. Since the frame with a high energy has more influence on the human ear, when we compute the clip-level features from the above five frame-level features, we use a weighted average of framelevel features, where the weighting for a frame is proportional to the energy of the frame. This is especially useful when there are many silent frames in a clip because the frequency features in silent frames are almost random. By using energy based weighting, their detrimental effects are removed. To summarize, we have developed twelve clip-level audio features: 1) non-silence-ratio (NSR), 2) volume standard deviation (VSTD), 3) volume dynamic range (VDR), 4) frequency component of the volume contour around 4Hz (FCVC4), 5) pitch standard deviation (PSTD), 6) voice-ormusic ratio (VMR), 7) noise-or-unvoice ratio (NUR), 8) frequency centroid (FC), 9) frequency bandwidth (BW), 10-12) energy ratios of subbands 1 3 (ERSB1-3). - 10 -

III. Feature Space Evaluation We have collected audio clips from TV programs containing the following five scene classes: news reports with one or two anchor men/women, weather forecasts with male/female reporters, TV commercials, live basketball games and live football games. For each scene class, we collected 1400 audio clips from different TV channels. These data were randomly divided into two sets: training and testing data sets. The training data set includes 400 clips from each scene class and the remaining 1000 clips in each class form the testing data set. All the analysis results reported in this section are computed from the training data. 1. Mean and Variance Analysis To see how the above features differ among separate scene classes, Table 1 lists the mean feature vectors for the five different scene classes, obtained by averaging the features extracted from all clips in the same class in the training data. We can see that for most features, news and weather reports have similar values. The football and basketball games also have similar values for most features. Features 2, 3, 8, and 12 (VSTD, VDR, FC and ERSB3) differ significantly among commercial, basketball/football, and news/weather. On the other hand features 5, 6 and 7 (PSTD, VMR and NUR) are different between news and weather report. This makes sense because the speech styles in the two programs are different: normally the speed of speech in a weather report is faster than that in news report and there are several speakers in news report while only one in weather report. Features 8 and 9 (FC and BW) differ between basketball games and football games. Although both contain a high level of background sound, their frequency structures are not the same, with one being indoor and the other being outdoor. Besides, in basketball games there exist a lot of high frequency audio components which are caused by the friction between the shoes of the players and the floor. - 11 -

Table 2 gives the standard deviation of individual features within each class. From the table we can see that the temporal variations of these features in different types of TV programs are different. For example, the range of NSR in news and weather is bigger than that in the other three programs, while the range of FC and BW in news and weather is smaller than those in the other three. 2. Clustering Analysis In order to evaluate the linear separability of different classes under the feature space generated by the twelve proposed features, we performed a clustering analysis of the training data using the Euclidean distance measure. The intention of this study is to see whether the five audio classes form linearly separable clusters in the feature space, so that they can be separated using a simple nearest neighbor classifier. Since we do not know how many clusters one audio class may have, and nor do we know whether certain audio classes overlap in the feature space, we use an automatic clustering method to find the number of clusters and their centroids in the feature space. There are two popular clustering algorithms: K-Means and Iterative Self-Organizing Data Analysis Techniques Algorithm (ISODATA) [21]. With the K-Means algorithm, we must provide the expected number of clusters, while with ISODATA, the number is determined by dynamic merging and splitting of clusters based on certain criteria. Without knowing the exact number of clusters in our training data, we use the ISODATA algorithm. By restricting the maximum number of clusters to be 40 and requiring each cluster to contain no less than 10 members, this method results in thirteen clusters. Table 3 lists the mean feature vector of each cluster. Table 4 shows the distribution of each cluster in different audio classes. We can see that some clusters belong to a single class, such as cluster 3 and 11 which belong to the commercial class, while other clusters are shared by different classes, e.g. cluster 6 and 8 are shared by news and weather report. - 12 -

The fact that the five scene classes are not linearly separable under the twelve clip-level features is not surprising. For example, in a TV commercial, there are periods of pure speech, periods of pure background music and periods of mixtures. The pure speech period also appears in news/weather report. In order to tell that a particular speech clip is in a TV commercial rather than a new report, one must look at several neighboring clips and make the classification based on highlevel semantic analysis. Because of this inherent ambiguity, it is doubtful that additional clip-level features can make all different scene classes linearly separable. 3. Intra-cluster and Inter-cluster Scattering Analysis In order to evaluate how discriminative each feature is and whether certain features are correlated with each other, we also calculated the intra-class and inter-class scattering matrices [21] of the feature vector consisting of the twelve features. The intra-class scattering matrix reveals the scattering of samples around their respective class centroids, and is defined by S T {( X M )( X M ) } N intra = P i ) E i i i= 1 ( ω ω, where P( ω i ) is the a priori probability of class ω, X is the sample feature vector, M i i is the mean feature vector (centroid) of class ω, N is the number of classes. On the other hand, the inter-class i scattering matrix is defined as: S N T inter = P i )( M i M 0 )( M i M 0 ) i= 1 (ω, where M 0 = P(ω i ) M i. The intra-class and inter-class scattering analysis is appropriate only if there exists a single cluster for each class. This is however not the case with our feature space. Therefore, we applied the above analysis to the 13 clusters found by the ISODATA algorithm. This analysis is therefore used to see how useful each feature is in differentiating among clusters. Because of the overlapping of certain clusters among several classes, a feature that cannot discriminate among clusters is likely to fail to i N i= 1-13 -

differentiate one class from the other. On the other hand, a feature that can separate different clusters may not be able to differentiate all the classes. Table 5 and Table 6 give the intra-cluster and inter-cluster scatter matrices determined from the training data. When calculating these matrices, and when performing classification (to be discussed in Section IV), all features are normalized by the maximum values of respective features in the training set. The diagonal entries in these two matrices characterize the intra- and inter-cluster separability of individual features. If the diagonal entry in the intra-cluster scattering matrix is small while that in the inter-cluster matrix is large, then the corresponding feature has good cluster separability. The off-diagonal entries in these two matrices reveal the correlation between different features. We can use these measures to eliminate highly correlated features and reduce the dimensionality of the feature space. From Table 5 we can see that there exists high correlation between some feature pairs, such as VSTD and VDR, FC and BW, ERSB1 and ERSB2, etc. Here, the correlation is measured by Si (, j)/ SiiS (,) ( j, j), where S(i,j) is the entry at the i-th row and j-th column of the intra-cluster scattering matrix. Based on this table we can reduce the dimension of the feature space by proper transformations. Table 7 compares the diagonal entries of the intra-cluster and inter-cluster scattering matrices. We can see that FCVC4, FC, BW, ERSB1 and ERSB3 have good cluster separability (with an inter-to-intra ratio greater than 2.5), while NSR, PSTD and VSTD have poor cluster separability. IV. Audio-Based Scene Classification As described in Section I, there are three layers at which we can fulfill scene classification. In the audio characteristics layer, we consider the features of individual clips independently. Classification - 14 -

based on the clip-level features alone is difficult, because the feature space is not linearly separable. If we use a simple nearest neighbor classifier using the Euclidean distance, we will not get satisfactory results. This has been proven by a preliminary experiment. In this simulation, we calculate the mean feature vector for each of the five classes based on the feature vectors obtained from the training data consisting of 400 clips from each class (results are previously given in Table 1). Then for the testing data set that consists of 1000 clips for each class, we use a nearest neighbor classifier. That is, each clip is classified by calculating the distance of its feature vector to the mean feature vectors of the five classes and identifying the class to which the distance is the shortest. The classification results using this method for the testing data are shown in Table 8, which are quite poor. The reason is that there are more than one cluster for each audio class in the feature space and not all these clusters are closer to the centroid of this class than those of other classes. Because of above reasons, the conventional nearest neighbor classifier is not effective for our classification task. Artificial neural networks have been used successfully as pattern classifiers in many applications for their ability to implement nonlinear decision boundaries and their capability to learn complicated rules from training data [22, 23]. Conventional multi-layer perceptron (MLP) use the all-class-in-one-network (ACON) structure, which is shown in Figure 10. But such a network structure has the burden of having to simultaneously satisfy all the desired outputs for all classes, so the required number of hidden units tends to be large. Besides, if one wants to adapt the network to new training data or add new classes, all the weights need to be re-computed. On the other hand, in the one-class-one-network (OCON) structure, one subnet is designated for recognizing one class only [24]. The structure is illustrated in Figure 11. Each subnet is trained individually using the back-propagation algorithm so that its output is close to 1 if the input pattern belongs to this class, otherwise the output is close to 0. Given an input audio clip, it is classified to the class whose - 15 -

subnet gives the highest score. An advantage of the OCON structure is that one can accommodate a new class easily by adding a subnet trained for that class. We have applied both ACON (1 hidden layer with 14 neurons) and OCON (5 subnets, 1 hidden layer with 7 neurons in each subnet) structures to perform audio classification. The weights are trained using the training data. The classification results for the testing data are shown in Table 9 and Table 10. Both classifiers can accurately distinguish among commercials, basketball game, football game, and news/weather reports. But the separation of the news from weather reports is less successful. This is not surprising because they contain primarily speech. To distinguish these two classes, some high level correlation information between successive clips that reflects the flow of the conversation may be necessary. As for the comparison of OCON and ACON structures, the classification accuracy using OCON is slightly higher than using ACON. Here the classification accuracy is defined as the average of the correct classification rates of all classes, i.e. the diagonal entries in Table 9 and Table 10 respectively. Using more neurons in the hidden layer of the ACON classifier may improve its performance. The classification results reported in this section are obtained by using all the twelve features described in Section II. We have also tested the performance of the OCON classifier when the three features involving pitch calculation (PSTD, VMR and UNR) are eliminated. From the analysis in Section III.3, these three features do not have good cluster separability. Also, pitch calculation is time consuming. The classification results obtained from using the remaining nine features are given in Table 11. It can be seen that the classification accuracy is about the same as that obtained using twelve features. This confirms that the three pitch-based features are not very useful and that the reduced feature set is sufficient for audio classification. Theoretically, using a reduced feature set should never give better classification results than using the original feature set. The somewhat improved performance for the classification of certain classes by the reduced feature set (compare - 16 -

Table 9 and Table 11) may be due to the better convergence performance achievable by the reduced feature space when the neural net classifier is trained. V. Scene Segmentation Using Audio Features To perform scene analysis in a video sequence, one common approach is to first segment the sequence into shots so that each shot contains the same type of scene, and then classify each shot into one scene type. Usually, scene segmentation is accomplished by detecting significant changes in the statistics of the underlying visual and audio signals. In this section, we consider scene segmentation based on audio information only. Note that an audio segment belonging to the same scene class may be segmented to different shots. For example, if in the middle of a commercial the background music is changed, then the sequence may be segmented into two shots, although they may both be classified as commercial in a later classification stage. Speech segmentation is a fundamental processing in speech recognition, where a speech signal is segmented into pieces containing voice, unvoice and silence. Segmentation of the audio signal is quite different from that of pure speech. In speech, the length of each segment is very short and the onset and offset of the segment should be precisely determined. On the other hand, in our task, we want to track the semantic content of an audio sequence. Normally, the audio signal with the same scene content will last from several seconds to several minutes. Because a scene transition usually occurs over a relatively long period so that one scene gradually changes to another, exact localization of the transition time is difficult to achieve and is usually not necessary. It is sufficient for most practical applications if a transition is detected shortly after the new scene is stabilized. There are two considerations in developing segmentation algorithms: the types of audio features based on which a change index is calculated, and the way such indices are calculated and used to locate scene changes. First, we should choose audio features that well describe the statistic - 17 -

characteristics of individual audio clips. Second, we should find a proper way to compare the features of current and previous audio clips. Saunders [16] used several features derived from the zero crossing rate to separate speech and music, and Nam and Tewfik [14] used energy ratios in five subbands to segment audio input. Here, we explore the use of the distances between feature vectors of adjacent clips for this task. From the intra-cluster and inter-cluster scattering analysis of the twelve features presented in Sec. III.3, we know that the three features involving pitch (PSTD, VMR and NUR) do not have good class discrimination capability. Since the computation of pitch is rather time-consuming, we exclude these three features when computing the feature vector difference. In [14], the difference between the feature vectors of two successive clips (80 msec. long) is computed for measuring scene change. This method is sensitive to short interference, such as the whistling in a basketball game. Here we adopt a different strategy, which compares the feature vector of the current clip to some previous clips and following clips. For a clip to be declared as a scene change, it must be similar to all the neighboring future clips, and different from all the neighboring previous clips. Based on this criterion, we propose using the following measure: Scene change index = 1 N 1 f ( i) N 1 N 1 i= N ( c + var( f ( N),..., f ( 1)) )( c + var( f (0),..., f ( N 1)) ) where f(i) is the feature of the i-th clip, with i=0 representing the current clip, i>0 a future clip, i<0 a previous clip,. is the L-2 norm, var( ) is the average of the squared Euclidean distances between each vector and the mean feature vector of the N clips considered, and c is a small constant to prevent division by zero. When the feature vectors are similar within previous N clips and following N clips, respectively, but differ significantly between the two groups, a scene break is declared. The selection of the window length N is critical: If N is too large, this strategy may fail to i= 0 f ( i) 2-18 -

detect scene changes between short audio shots. It will also add unnecessary delay to the processing. Through trials-and-errors, we have found that N=6 give satisfactory results. Figure 12(a) shows the content of one testing audio sequence used in segmentation. This sequence is digitized from a TV program that contains seven different semantic segments. The first and the last segments are both football games, between which are TV station s logo shot and four different commercials. The duration of each segment is also shown in the graph. The sequence in Figure 13(a) is obtained by manually combining 5 segments of TV programs, a commercial, a basketball game, a weather report, a football game and news report, each 25 seconds long. Figure 12(b) and Figure 13(b) show the scene-change-indices computed for these two sequences. Scene changes are detected by identifying those clips for which the scene-change-indices are higher than a threshold, D min, and are at least T min seconds away from a previous detected scene change. We have used D min =3, and N =6, which have been found to yield good results through trial-and-error. In these graphs, mark o indicates real scene changes and * detected scene changes. All the real scene changes are detected using this algorithm. Note that there are two falsely detected scene changes in the first segment of the first sequence. They correspond to the sudden appearance of the commentator s voice and the audience s cheering. Table 12 summarizes the segmentation results for ten test sequences (including the above two) using the proposed the algorithm. These sequences cover various combinations of different TV programs, including football and commercials, news and commercials, basketball and commercials, news and weather report and commercials, etc. Totally there are 44 real scene changes. The algorithm detects 42 of them. One missed scene change happens between two smoothly connected commercial programs and the other one happens between the news and weather report. There are 36 false alarms given by the algorithm, most of them happen during the sports programs and commercials, where audio discontinuities actually occurred. To remove these false detections, we - 19 -

can make use of the visual information in the video sequence For example, the color histogram in a sports segment usually remains the same. By requiring that both color histogram and audio features experience sharp changes, we can eliminate falsely detected scene changes based on the audio features alone. VII. Conclusions The primary goal of this study is to identify low-level features that are effective in characterizing the semantic content of an audio sequence. Towards this goal, we developed and analyzed twelve features. Through feature space analysis we have found that FCVC4, FC, BW, ERSB1 and ERSB3 have good scene discrimination capability while the three features involving pitch calculation (PSTD, VMR and NUR) have poor discrimination capacity. Using all twelve features, an OCON neural network classifier was successful (82.5% on average) in separating four kinds of TV programs: commercial, basketball game, football game, news report/weather forecast. The discrimination of news report from weather forecast was less accurate (less than 70%). Similar classification results were obtained with a reduced feature set containing nine features (excluding the three pitch-related features). Using the reduced feature vector consisting of the above nine features, a scene-change-index function was developed. From the experiment conducted, it promises to be an effective tool for scene segmentation. Because these nine features are easy to compute, clip-level scene classification and segmentation using these features can be achieved at quite low computation cost. Our feature space analysis has indicated that some features are less effective than others, and that several useful features are highly correlated. Use of feature space reduction techniques to deduce a more efficient set of features that can retain the discrimination capability of these twelve features is a topic of our future studies. There are also other features that we like to explore and compare with - 20 -

the ones presented here. In the present study, the effectiveness of the proposed features is evaluated by examining their capability in discriminating five common types of TV programs. However, because these are low-level features calculated over short audio-clips (one second long), we believe they are useful fundamental statistical features for any audio analysis task. Of course, analysis of a general audio sequence acquired in a noisy environment is more difficult and further study is needed to validate the effectiveness of the features presented here for other applications. The classification and segmentation results reported here are meant to show the promise of using audio features for scene analysis. Better classification and segmentation results should be obtainable with more optimized algorithms. In the present study, classification is accomplished for each clip based on the features of that clip alone. To achieve more accurate classification, such clip-wise classification results should be refined by considering high level semantic models. For example, in a news program, although there is only speech, the reporters are alternated frequently. On the other hand, during the weather forecast, usually the speaker talks for a much longer time period. This requires one to look at the classification results for several clips at a time. For both the classification and segmentation tasks, ideally, one should make use of both visual and audio information. In general, visual-based analysis involves much more computation than audio-based one. Using audio information alone can often provide a good initial solution for further examination based on visual information. A challenging problem is how to combine the results from audio and visual signal analysis for understanding the semantic content of a video sequence. These are some of the interesting and important problems for future study. - 21 -

ACKNOWLEDGMENT This material is based upon work supported by the National Science Foundation under Grant No. IRI-9619114 and by the New York State Center for Advanced Technology in Telecommunications at Polytechnic University, Brooklyn, New York. REFERENCES [1] M. A. Smith and T. Kanade, Video Skimming and Characterization through the Combination of Image and Language Understanding Techniques, Carnegie Mellon University Technical Report CMU-CS-97-111, Feb. 1997 [2] Y. Chang, W. Zeng, I. Kamel and R. Alonso, Integrated Image and Speech Analysis for Content-Based Video Indexing, Proc. of the 3rd IEEE International Conference on Multimedia Computing and Systems, pp. 306-313, 1996 [3] M. Flickner, et al., Query by image and video content: The QBIC system, IEEE Computer, Vol. 28, No. 9, pp. 23-32, September 1995. [4] S. W. Smoliar and H. Zhang, Content-Based Video Indexing and Retrieval, IEEE Multimedia Magazine, Vol. 1, No. 2, pp. 62-72, Summer 1994. [5] J. R. Smith and S. -F. Chang, SaFe: A General Framework for Integrated Spatial and Feature Image Search, Proc. IEEE 1 st Multimedia Signal Processing Workshop, pp. 301-306, June 1997. [6] H. J. Zhang, A. Kankanhalli, and S. W. Smoliar, Automatic partitioning of full-motion video, Multimedia Systems, Vol. 1, No. 1, pp. 10-28. 1993. [7] H. J. Zhang, et. al, An Integrated System for Content-Based Video Retrieval and Browsing, Pattern Recognition, Vol. 30, No. 4, pp.643-658, 1997. - 22 -

[8] J. D. Courtney, Automatic Video Indexing via Object Motion Analysis, Pattern Recognition, Vol. 30, No. 4, pp.607-625, 1997. [9] M. Irani, P. Anandan, J. Bergern, R. Kumar, and S. Hsu, Efficient Representations of Video Sequences and Their Applications, Signal Processing: Image Communication, Vol. 8, pp. 327-351, 1996. [10] B. Shahraray and D. C. Gibbon, Pictorial Transcripts: Multimedia Processing Applied to Digital Library Creation, Proc. IEEE 1 st Multimedia Signal Processing Workshop, pp. 581-586, June 1997. [11] M. M. Yeung and B. -L. Yeo, Video Visualization for Compact Presentation and Fast Browsing of Pictorial Content, IEEE Trans. Circuits and Systems for Video Technology, Vol. 7, No. 5, pp. 771-785, Oct. 1997. [12] C. Saraceno and R. Leonardi, Audio as a Support to Scene Change Detection and Characterization of Video Sequences, Proc. of ICASSP 97, Vol. 4, pp. 2597-2600,1997. [13] S. Pfeiffer, S. Fischer and W. Effelsberg, Automatic Audio Content Analysis, Proc. ACM Multimedia 96, pp. 21-30, 1996 [14] J. Nam and A. H. Tewfik, Combined Audio and Visual Streams Analysis for Video Sequence Segmentation, Proc. of ICASSP 97, Vol. 3, pp. 2665-2668, 1997. [15] Y. Wang, J. Huang, Z. Liu, and T. Chen, Multimedia Content Classification using Motion and Audio Information, Proc. of IEEE ISCAS 97, Vol. 2, pp.1488-1491, 1997. [16] J. Saunders, Real-time Discrimination of Broadcast Speech/Music, Proc. of ICASSP 96, Vol. 2, pp. 993-996, 1996. [17] E. Scheirer and M. Slaney, "Construction and Evaluation of a Robust Multifeature Speech/Music Discriminator," Proc. of ICASSP'97, Vol. 2, pp.1331-1334,1997. [18] W. Hess, Pitch Determination of Speech Signals, Springer-Verlag, 1983. - 23 -

[19] E. Wold, et al., Content-Based Classification, Search, and Retrieval of Audio, IEEE Multimedia Magazine, Vol. 3, No. 3, pp. 27-36, Fall 1996. [20] N. Jayant, J. Johnston and R. Safranek, Signal Compression Based on Models of Human Perception, Proceedings of IEEE, Vol. 81, No. 10, pp. 1385-1422 October 1993. [21] K. Fukunaga, Introduction to statistical pattern recognition, Academic Press, 1972. [22] B. Kosko, Neural Networks for Signal Processing, Englewood cliffs, NJ Prentice Hall, 1992. [23] R. P. Lippman, An Introduction to Computing with Neural Nets, IEEE ASSP Magazine, pp. 4-22, April 1987. [24] S. H. Lin, S. Y. Kung, and L. J. Lin, Face Recognition/ Detection by Probabilistic Decision- Based Neural Network, IEEE Trans. Neural Networks, Vol. 8, No. 1, pp. 114-132, Jan. 1997. - 24 -

List of Figures Figure 1 Clip and frames used in feature analysis...27 Figure 2 Waveforms of three audio clips...27 Figure 3 Volume contours of three audio clips...27 Figure 4 The ADMF of one speech frame...28 Figure 5 Pitch contours of three audio clips...28 Figure 6 Spectrograms of three audio clips...29 Figure 7 Contours of frequency centroid of three audio clips...29 Figure 8 Contours of bandwidth of three audio clips...29 Figure 9 Energy ratio in 4 subbands of three audio clips...30 Figure 10 Structure of the ACON neural net classifier...30 Figure 11 Structure of the OCON neural net classifier...31 Figure 12 Content and Scene-change-index calculated for the second sequence...32 Figure 13 Content and Scene-change-index calculated for the second sequence...33-25 -

List of Tables Table 1 Mean Feature Vectors of Different Scene Classes...34 Table 2 Standard Deviation of Features in Different Scene Classes...34 Table 3 Mean Feature Vectors of 13 Clusters...35 Table 4 Distributions of Feature Clusters in Different Scene Classes...35 Table 5 Intra-Cluster Scattering Matrix...36 Table 6 Inter-Cluster Scattering Matrix...36 Table 7 Diagonal Entries of Intra and Inter-Cluster Scattering Matrices...37 Table 8 Classification Results Using a Nearest Neighbor Classifier...37 Table 9 Classification Results Using an OCON Neural Net...38 Table 10 Classification Results Using an ACON Neural Net...38 Table 11 Classification Results Using Nine Features...39 Table 12 Segmentation Results...39-26 -

1 clip (1 second long: 22050 samples) 1st frame (512 samples) 2nd frame (512 samples, shifted by 128 samples) 3rd frame... Figure 1 Clip and frames used in feature analysis (a) commercial (b) basketball (c) news Figure 2 Waveforms of three audio clips (a) commercial (b) basketball (c) news Figure 3 Volume contours of three audio clips - 27 -

Figure 4 The ADMF of one speech frame (a) commercial (b) basketball (c) news Figure 5 Pitch contours of three audio clips - 28 -

(a) commercial (b) basketball (c) news Figure 6 Spectrograms of three audio clips (a) commercial (b) basketball (c) news Figure 7 Contours of frequency centroid of three audio clips (a) commercial (b) basketball (c) news Figure 8 Contours of bandwidth of three audio clips - 29 -

(a) commercial (b) basketball (c) news Figure 9 Energy ratio in 4 subbands of three audio clips Output Input Figure 10 Structure of the ACON neural net classifier - 30 -

Classification Results MAXNET............... ( Class 1 ) ( Class N ) Input Features Figure 11 Structure of the OCON neural net classifier - 31 -

Football TV Logo Ad. 1 Ad. 2 Ad. 3 Ad. 4 Football 0 93 105 135 151 180 207 242 Time (s) (a) Semantic contents of first testing sequences 10 9 8 Sc en e- ch an ge - in 7 6 5 4 3 2 1 0 0 50 100 150 200 Time (second) (b) Scene-change-index Figure 12 Content and Scene-change-index calculated for the first sequence - 32 -

Advertisement Basketball Forecast Football News 0 25 50 75 100 125 Time (s) (a) Semantic contents of the sequence 10 9 8 Sc en e- ch an ge - in 7 6 5 4 3 2 1 0 0 20 40 60 80 100 120 Time (second) (b) Scene-change-index Figure 13 Content and Scene-change-index calculated for the second sequence - 33 -

Table 1 Mean Feature Vectors of Different Scene Classes Feature 1. NSR 2. VSTD 3. VDR 4. FCVC4 5. PSTD 6. VMR Commercial 0.997 0.195 0.798 0.011 2.289 0.281 Basketball 0.999 0.178 0.719 0.011 2.115 0.254 Football 0.998 0.174 0.702 0.011 2.241 0.297 News 0.898 0.271 0.953 0.027 1.328 0.459 Weather 0.897 0.272 0.966 0.029 1.198 0.505 Feature 7. NUR 8. FC 9. BW 10. ERSB1 11. ERSB2 12. ERSB3 Commercial 0.540 829.4 797.8 0.630 0.245 0.090 Basketball 0.589 1432.5 1176.3 0.353 0.352 0.217 Football 0.520 1161.4 875.2 0.367 0.398 0.213 News 0.299 666.1 436.0 0.650 0.292 0.044 Weather 0.234 664.8 406.5 0.620 0.333 0.035 Table 2 Standard Deviation of Features in Different Scene Classes Feature 1. NSR 2. VSTD 3. VDR 4. FCVC4 5. PSTD 6. VMR Commercial 0.013 0.045 0.109 0.007 1.163 0.221 Basketball 0.003 0.045 0.119 0.006 1.087 0.172 Football 0.008 0.059 0.167 0.008 1.131 0.198 News 0.112 0.033 0.033 0.014 0.736 0.170 Weather 0.095 0.031 0.018 0.013 0.562 0.150 Feature 7. NUR 8. FC 9. BW 10. ERSB1 11. ERSB2 12. ERSB3 Commercial 0.272 337.8 310.0 0.205 0.178 0.080 Basketball 0.189 368.7 195.1 0.156 0.138 0.094 Football 0.196 295.5 164.9 0.165 0.166 0.089 News 0.129 219.0 146.5 0.191 0.183 0.037 Weather 0.095 125.2 98.1 0.160 0.156 0.029-34 -