Automatic Summarization of Music Videos

Similar documents
Automatic Rhythmic Notation from Single Voice Audio Sources

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis

2. AN INTROSPECTION OF THE MORPHING PROCESS

Supervised Learning in Genre Classification

Computational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)

A repetition-based framework for lyric alignment in popular songs

Content-based Music Structure Analysis with Applications to Music Semantics Understanding

Subjective Similarity of Music: Data Collection for Individuality Analysis

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

DETECTION OF SLOW-MOTION REPLAY SEGMENTS IN SPORTS VIDEO FOR HIGHLIGHTS GENERATION

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

Week 14 Query-by-Humming and Music Fingerprinting. Roger B. Dannenberg Professor of Computer Science, Art and Music Carnegie Mellon University

Automatic Replay Generation for Soccer Video Broadcasting

Repeating Pattern Discovery and Structure Analysis from Acoustic Music Data

Music Segmentation Using Markov Chain Methods

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

CS229 Project Report Polyphonic Piano Transcription

Toward Automatic Music Audio Summary Generation from Signal Analysis

VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS. O. Javed, S. Khan, Z. Rasheed, M.Shah. {ojaved, khan, zrasheed,

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors

Music Genre Classification and Variance Comparison on Number of Genres

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

Popular Song Summarization Using Chorus Section Detection from Audio Signal

hit), and assume that longer incidental sounds (forest noise, water, wind noise) resemble a Gaussian noise distribution.

AUTOMATIC ACCOMPANIMENT OF VOCAL MELODIES IN THE CONTEXT OF POPULAR MUSIC

An Examination of Foote s Self-Similarity Method

Music structure information is

MUSI-6201 Computational Music Analysis

A Framework for Segmentation of Interview Videos

THE importance of music content analysis for musical

Speech and Speaker Recognition for the Command of an Industrial Robot

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

GRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION

Reducing False Positives in Video Shot Detection

Semantic Segmentation and Summarization of Music

CS 591 S1 Computational Audio

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

Hidden Markov Model based dance recognition

... A Pseudo-Statistical Approach to Commercial Boundary Detection. Prasanna V Rangarajan Dept of Electrical Engineering Columbia University

Music Radar: A Web-based Query by Humming System

Music Source Separation

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

Browsing News and Talk Video on a Consumer Electronics Platform Using Face Detection

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

Composer Style Attribution

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

Audio-Based Video Editing with Two-Channel Microphone

Polyphonic Audio Matching for Score Following and Intelligent Audio Editors

Robert Alexandru Dobre, Cristian Negrescu

Singer Traits Identification using Deep Neural Network

Tempo Estimation and Manipulation

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

Research Article. ISSN (Print) *Corresponding author Shireen Fathima

Topic 10. Multi-pitch Analysis

A Parametric Autoregressive Model for the Extraction of Electric Network Frequency Fluctuations in Audio Forensic Authentication

Singer Identification

Automatic Laughter Detection

Tempo and Beat Analysis

Wipe Scene Change Detection in Video Sequences

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Figure 1: Feature Vector Sequence Generator block diagram.

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

Singer Recognition and Modeling Singer Error

Music Alignment and Applications. Introduction

A Parametric Autoregressive Model for the Extraction of Electric Network Frequency Fluctuations in Audio Forensic Authentication

Query By Humming: Finding Songs in a Polyphonic Database

Analysis of local and global timing and pitch change in ordinary

Acoustic Scene Classification

Music Recommendation from Song Sets

A Study of Synchronization of Audio Data with Symbolic Data. Music254 Project Report Spring 2007 SongHui Chon

EE391 Special Report (Spring 2005) Automatic Chord Recognition Using A Summary Autocorrelation Function

Story Tracking in Video News Broadcasts. Ph.D. Dissertation Jedrzej Miadowicz June 4, 2004

Automatic Construction of Synthetic Musical Instruments and Performers

MusCat: A Music Browser Featuring Abstract Pictures and Zooming User Interface

Features for Audio and Music Classification

Music Structure Analysis

Week 14 Music Understanding and Classification

Pattern Recognition in Music

Predicting Variation of Folk Songs: A Corpus Analysis Study on the Memorability of Melodies Janssen, B.D.; Burgoyne, J.A.; Honing, H.J.

Agilent PN Time-Capture Capabilities of the Agilent Series Vector Signal Analyzers Product Note

Transcription of the Singing Melody in Polyphonic Music

IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 12, NO. 7, NOVEMBER

ISSN ICIRET-2014

Automatic music transcription

Chord Classification of an Audio Signal using Artificial Neural Network

Outline. Why do we classify? Audio Classification

On Human Capability and Acoustic Cues for Discriminating Singing and Speaking Voices

UC San Diego UC San Diego Previously Published Works

A Survey of Audio-Based Music Classification and Annotation

Unit Detection in American Football TV Broadcasts Using Average Energy of Audio Track

Audio Structure Analysis

Classification of Timbre Similarity

Getting Started. Connect green audio output of SpikerBox/SpikerShield using green cable to your headphones input on iphone/ipad.

Detection of Panoramic Takes in Soccer Videos Using Phase Correlation and Boosting

Automatic Music Clustering using Audio Attributes

Topics in Computer Music Instrument Identification. Ioanna Karydi

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

Interacting with a Virtual Conductor

Transcription:

Automatic Summarization of Music Videos XI SHAO, CHANGSHENG XU, NAMUNU C. MADDAGE, and QI TIAN Institute for Infocomm Research, Singapore MOHAN S. KANKANHALLI School of Computing, National University of Singapore and JESSE S. JIN School of Design, Communication and IT, University of Newcastle, Australia In this article, we propose a novel approach for automatic music video summarization. The proposed summarization scheme is different from the current methods used for video summarization. The music video is separated into the music track and video track. For the music track, a music summary is created by analyzing the music content using music features, an adaptive clustering algorithm, and music domain knowledge. Then, shots in the video track are detected and clustered. Finally, the music video summary is created by aligning the music summary and clustered video shots. Subjective studies by experienced users have been conducted to evaluate the quality of music summaries and effectiveness of the proposed summarization approach. Experiments are performed on different genres of music videos and comparisons are made with the summaries generated based on music track, video track, and manually. The evaluation results indicate that summaries generated using the proposed method are effective in helping realize users expectations. Categories and Subject Descriptors: H.3.1 [Information Storage and Retrieval]: Content Analysis and Indexing Abstracting methods, indexing methods General Terms: Algorithms, Experimentation, Performance Additional Key Words and Phrases: Music summarization, video summarization, music video 1. INTRODUCTION The rapid development of various affordable technologies for multimedia content capture, data storage, high bandwidth/speed transmission, and the multimedia compression standards have resulted in a rapid increase in digital multimedia data content. This has led to greater availability of multimedia content for general users. How to create a concise and informative abstraction that best summarizes the original digital content is extremely important in large-scale information organization. Nowadays, many music companies are putting music videos on websites, and customers can purchase them via the Internet. However, from the customer point of view, it would be preferable to watch the highlights before making purchases. On the other hand, from the music company point of view, it is better to provoke Authors addresses: X. Shao, C. Xu, N. C. Maddage, and Q. Tian, Institute for Infocomm Research, 21 Heng Mui Keng Terrace, Singapore 119613; email: xucs@i2r.a-star.edu.sg; M. S. Kankanhalli, School of Computing, National University of Singapore, J. S. Jin, School of Design, Communication and IT, University of Newcastle, Australia. Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or direct commercial advantage and that copies show this notice on the first page or initial screen of a display along with the full citation. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, to republish, to post on servers, to redistribute to lists, or to use any component of this work in other works requires prior specific permission and/or a fee. Permissions may be requested from Publications Dept., ACM, Inc., 1515 Broadway, New York, NY 10036 USA, fax: +1 (212) 869-0481, or permissions@acm.org. c 2006 ACM 1551-6857/06/0500-0127 $5.00 ACM Transactions on Multimedia Computing, Communications and Applications, Vol. 2, No. 2, May 2006, Pages 127 148.

128 X. Shao et al. Fig. 1. A block diagram of the proposed summarization system. the buying interests of music fans by showing the highlights of a music video rather than showing it all, as there are no profits for company if they allow music fans download the whole music video freely. Music video summaries are available on some music websites, but they are generated manually, which is very labor intensive and time consuming. Therefore, it is crucial to come up with an automatic summarization approach for music videos. In this article, a novel and effective approach for automatic music video summarization is presented. Figure 1 is the block diagram of the proposed approach. The music video is separated into the music track and the video track. For the music track, a music summary is created by analyzing the music content using music features, an adaptive clustering algorithm, and music domain knowledge. For the video track, shots are detected and clustered using visual content analysis. Finally, the music video summary is created by specially aligning the music summary and clustered visual shots. This approach provides better results than audiocentric or imagecentric summarization methods alone, based on the feedback of users in our study. Our method is capable of maximizing the coverage for both audio and visual contents without sacrificing either of them. The rest of the article is organized as follows: Section 2 introduces related work. Section 3 describes the structure of music videos and the framework of music video summarization. Section 4 explains the music summarization scheme. Section 5 explains shot detection and clustering for the video track, while Section 6 illustrates how to align the music summary and video shots to create the music video summary. Section 7 discusses the user study. Finally, Section 8 presents conclusions and plans for future work. 2. RELATED WORK Several methods have been proposed for automatic music summarization in the past. A summarization system for MIDI data has been developed [Kraft et al. 2001] which utilizes the repetition nature of MIDI compositions to automatically recognize the main melody theme segment. However, MIDI compositions are synthesized and structured, and their representations are different from sampled audio format such as WAV, which is highly unstructured. Therefore, MIDI summarization methods cannot be applied to real music summarization. A real music summarization system [Logan and Chu 2000] used mel

Automatic Summarization of Music Videos 129 frequency cepstral coefficients (MFCCs) to parameterize each music song. Based on MFCCs, a crossentropy measure and hidden markov model (HMM) were used to discover the song structure. Heuristics were applied to extract the key phrase in line with this structure. This summarization method was suitable for certain music genres such as rock or folk music, but was less applicable to classical music. MFCCs were also used as features in the work of Cooper and Foote [2002]. They used a 2-D similarity matrix to represent music structures and generate a summary. Peeters et al. [2002] proposed a multipass approach to generate music summaries. The first pass used segmentation to create states and the second pass used these states to structure music content by unsupervised learning (k-means). The summary was constructed by choosing a representative example of each state. The generated summary can be further refined by an overlap-add method and a tempo detection/beat alignment algorithm. However, there were no reports on quality evaluations of the generated summaries. There are a number of approaches proposed for automatic video summarization. Existing video summarization methods can be classified into two categories: key-frame extraction and highlight creation. Using a set of key frames to create the video summary is the most common approach and many key frame extraction methods [DeMenthon et al. 1998; Gunsel and Tekalp 1998; Zhuang et al. 1998] have been proposed. Key frames can help identify the desired shots of the video, but they are insufficient to obtain a general idea of whether a created summary is relevant or not. To make the summary more relevant to and representative of the video content, video highlight methods [Sundaram et al. 2002; Assfalg et al. 2002] were proposed to reduce a long video into a short sequence. This was intended to help users determine whether a video was worth viewing in its entirety. It can provide an impression of the entire video content or contain only the most interesting video sequences. Since there is no ground truth to evaluate whether the extracted highlight is able to represent the most interesting and salient parts of a given video content, it is hard to develop an automated system for extracting video highlights that would be acceptable to different users. For example, an extracted highlight of a video content that is acceptable to one user may not be acceptable to another user. Therefore, designing standards for user-studies to evaluate video summarization is still an open question. The music video (MV) is one video genre popular among music fans today. Nowadays, most MV summaries are manually produced. This is in contrast to other video genres. Automatic video summarization has been applied to sports video [Yow et al. 1995], news video [Nakamura and Kanade 1997; Gong et al. 2002], home video [Gong et al. 2001; Foote et al. 2002], and movies [Pfeiffer et al. 1996]. Although recent work on video summarization techniques of music video has been reported [Agnihotri et al. 2003, 2004], this work used high-level information such as titles, artists, and closed captions rather than low-level audio/visual features to generate the music video summary. However, such high-level metadata is not easily obtained directly from music video content. Therefore, assuming the availability of such metadata makes the problem easier and is not feasible to automatic summarization based on music video content only. The approach we propose in this article is to generate music video summary based on low-level audio/visual features which can be directly obtained from the music video content. To our best knowledge, there is no summarization technique available for music videos using low-level audio/visual features. 3. MUSIC VIDEO STRUCTURE Video programs such as movies, dramas, talk shows, etc., have a strong synchronization between their audio and visual contents. Usually, what we hear from the audio track directly explains what we see on the screen, and vice versa. For this type of video program, since synchronization between audio and image is critical, the summarization strategy has to be either audiocentric or imagecentric. The audiocentric summarization can be accomplished by first selecting important audio segments of the original video based on certain criteria and then concatenating them together to compose an audio

130 X. Shao et al. summary. To enforce the synchronization, the visual summary has to be generated by selecting the image segments corresponding to those audio segments in the audio summary. Similarly, an imagecentric summary can be created by selecting representative image segments from the original video to generate a visual summary, and then taking the corresponding audio segments to generate the associated audio summary. For both summarization approaches, either the audio or visual content of the original video will be sacrificed in the summaries. However, the music video is a special type of video. The visual and audio content combination in the music video can be divided into two categories: the polyphonic structure and homophonic structure [Zettl 1999]. In a polyphonic structure, the visual content does not in any way parallel the lyrics of the music. The visual content seems to tell its own story and is relatively independent of the meaning of the lyrics. For example, while the music proclaims tender love, the pictures may show surprisingly violent scenes. For these music videos, due to their weak synchronization between the visual and audio content, summarizing the visual and audio track separately and then sticking them together appears to be satisfactory. In a homophonic structure, the lyrics of the music, or at least its major literal themes, are in step with the visual event with similar meanings. According to Zettl [1999], the picture and sound in these videos are organized as an aesthetic whole using some matching criteria such as historical matching, geographical matching, thematic matching, and structure matching. For the music videos in this category, on the one hand, we can summarize them using the same methods as in audiocentric and imagecentric summarization, which enforces the synchronization but has to sacrifice either the audio or visual content of the original video. On the other hand, we can also use the same summarization approach as for the polyphonic structure music video, which enforces the maximum coverage both for video and audio content but has to sacrifice the synchronization thereof. In other words, we have to trade off between the maximum coverage and synchronization. Considering human perception, there is an asymmetrical effect of audio-visual temporal asynchrony on the auditory and visual attention [Sugana and Iwamiya 2000]. Auditory attention is sensitive to audio-visual asynchrony, while visual attention is insensitive to audio-visual asynchrony. Therefore, the minor deviation from the visual content of the music is within the range of human perceptional acceptance. Based on the aforementioned analysis, we use the same summarization approach for the music video in homophonic structure as the one used in polyphonic structure, which can maximize the coverage for both audio and visual contents without having to sacrifice either of them, at the cost of some potential asynchrony between the audio and video track. However, we have realized that the ideal summarization scheme for music video in homophonic structure should have both maximum coverage of and strict synchronization for the visual and auditory content. This can be achieved by semantic structure analysis for both the visual and music content and will be addressed in the future work mentioned in Section 8. 4. MUSIC SUMMARIZATION Compared with text, speech, and video summarization techniques, music summarization provides a special challenge because raw digital music data is a featureless collection and is only available in the form of highly unstructured monolithic sound files. Music summarization refers to determining the most common and salient themes of a given piece of music that can be used as a representative of the music and readily recognized by a listener. Music structure analysis is important for music summarization. We found that a song (such as Top of the World, by Carpenter) is normally composed of three parts: the intro, principal, and outro, as shown in Figure 2. The vertical axis represents the normalized frequency and the horizontal axis represents the sample index. In Figure 2, V represents the pure singing voice and I represents pure

Automatic Summarization of Music Videos 131 Fig. 2. Typical music structure embedded in the spectrogram. instrumental music. The combination I + V refers to vocal music, which is defined as music containing both singing voice and instrumental music. The intro and outro parts usually contain pure instrumental music without vocal components, while the principal part contains a mixture of vocal and instrumental music as well as some pure music portions. Pure music is here defined as music that contains only instrumentals lasting for at least 3 s because the pure music used to bridge different parts is normally more than 3 s, while the music between the music phrases within the verse or chorus is less than 3 s and cannot be treated as the pure music. Because these three parts play different roles in conveying musical information to listeners, we treat them separately when creating the music summary (see Section 4.2 for a detailed description). For each part of the music, the content is segmented into fixed-length and overlapping frames. Feature extraction is performed in each frame. Based on the calculated features, an adaptive clustering algorithm is applied to group these frames so as to obtain the structure of the musical content. Finally, a music summary is created, based on the clustered results and music domain knowledge. 4.1 Feature Extraction Feature extraction is very important for music content analysis. The extracted features should reflect the significant characteristics of the musical content. Commonly extracted features include linear prediction coefficient derived cepstrum (LPCC), zero-crossing rates (ZCR), and mel frequency cepstral coefficients (MFCCs). 4.1.1 Linear Prediction Coefficients (LPC) and LPC Derived Cepstrum Coefficients. LPC and LPCC are two linear prediction methods [Rabiner and Juang 1993] and they are highly correlated to each other. The basic idea behind linear predictive analysis is that a music sample can be approximated as a linear combination of past music samples. By minimizing the sum of the squared differences (over a finite interval) between the actual music samples and the linear predictive ones, a unique set of predictor coefficients can be determined. Our experiment shows that LPCC is much better than LPC in identifying vocal music [Gao et al. 2003]. Generally speaking, the performance of LPC and LPCC can be improved by approximately 20 25% by filtering the full band music signal (0 22.05 KHz with 44.1 KHz sampling rate) into subfrequency bands and then down-sampling the sub-bands before calculating the coefficients. The sub-bands

132 X. Shao et al. Fig. 3. Block diagram for calculating LPC & LPCC. Fig. 4. Zero-crossing rates (0 276 s is vocal music and 276 592 s is pure music). are defined according to the lower, middle, and higher music scales [Deller et al. 1999], as shown in Figure 3. Frequency ranges for designed filter banks are [0 220.5], [220.5 441], [441 661.5], [661.5 882], [882 1103], [1103 2205], [2205 4410], [4410 8820], [8810 17640], and [17640 22050] Hz. Therefore, calculating the LPC for different frequency bands can represent the dynamic behavior of the spectrums of the selective frequency bands (i.e., different octaves of the music). 4.1.2 Zero-Crossing Rates (ZCR). In the context of discrete-time signals, a zero-crossing refers to two successive samples having different algebraic signs. The rate at which zero-crossings occur is a simple measure of the frequency content of a signal. This average zero-crossing rate gives a reasonable way to estimate the frequency content of a signal. While ZCR values of instrumental music are normally within a relatively small range, vocal music is often indicated by high-amplitude ZCR peaks resulting from pronunciations of consonants [Zhang 2003]. Therefore, ZCR values are useful for distinguishing between vocal and pure music. Figure 4 is an example of zero-crossing rates for vocal music and pure music. It can be seen that the vocal music has higher zero-crossing rates than pure music. This feature is also quite sensitive to vocals and percussion instruments. Mean values are 188.247 and 47.023 for vocal music and pure music, respectively. 4.1.3 Mel Frequency Cepstral Coefficients. Mel-Cepstral features have proven to be highly effective in automatic speech recognition and in modeling the subjective pitch and frequency content of audio signals. Mel-Cepstral features can be illustrated by the mel-frequency cepstral coefficients (MFCCs), which are computed from the FFT power coefficients. The power coefficients are filtered by a triangular

Automatic Summarization of Music Videos 133 Fig. 5. The 3rd mel-frequency cepstral coefficient (0 276 s is vocal music and 276 573 s is pure music). band-pass filter bank. The filter bank consists of K = 19 triangular filters. Denoting the output of the k-th filter bank by S k (k = 1, 2,...,K), the MFCCs can be calculated as: 2 K c n = (log S k ) cos[n(k 0.5)π/K] n = 1, 2,...,L, (1) K k=1 where L is the number of cepstral coefficients. In our experiment, we let L be equal to 19. MFCCs are good features for analyzing music because of the significant spectral differences between human vocalization and musical instruments [Maddage et al. 2003]. Figure 5 is an example of the MFCCs for vocal and instrumental music. It can be seen that the mean value is 1.3704 for vocal music and 0.9288 for instrumental ( pure ) music. The variance is very high for vocal music, while it is considerably low for pure music. 4.2 Music Classification The purpose of music classification is to analyze a given musical, sequence to identify the pure music and the vocal music segments. According to music theory, the most distinctive or representative musical themes should repetitively occur in the vocal part of an entire musical work [Eugene 1990] and the summary should focus on the mixture portion (instrumental-only music is not considered in this article). Therefore, the pure music in the principal part is not the key component of a song (mostly, the pure music in the principal part is the bridge between the chorus and verse) and can be discarded. But pure music in the intro and outro contains information indicating the beginning and the end of the musical work and cannot be ignored. Therefore, if the pure music segment is detected at the beginning and the end of the musical sequence, it will be identified as the intro and outro part, separately. We will keep these two parts in the music summary. As for the pure music in the principal part, we discard it and only create a summary of mixed music in the principal part. Based on the calculated features (LPCC, ZCR, and MFCCs) of each frame, we employ a nonlinear support vector classifier to discriminate the vocal from the pure music. The support vector machine (SVM) technique is a useful statistical machine learning technique that has been successfully applied in the area of pattern recognition [Joachims 1998; Papageorgiou et al. 1998]. Figure 6 illustrates a conceptual block diagram of the training process to produce classification parameters of the classifier.

134 X. Shao et al. Fig. 6. Diagram of the SVM training process. The training process analyses music training data to find an optimal way to classify music frames into either the pure or vocal class. The training data is segmented into fixed-length and overlapping frames (in our experiment, we used 20 ms frames overlapping 50%). Features such as LPCC, ZCR, and MFCCs are calculated from each frame. The SVM methodology is applied to produce the classification parameters according to the calculated features. The training process needs to be performed only once. The derived classification parameters are used to classify pure and vocal music. After training, the derived classification parameters are used to identify pure music and vocal music. For a given music track: (1) segment it into fixed-length frames; (2) for every frame, extract features such as LPCC, ZCR, and MFCCs to construct the feature vector; (3) input each feature vector to a trained SVM and the SVM will label the corresponding frame as either pure or vocal music; (4) for those frames labeled pure music, if the continuous frames last for more than 3 s, identify them as a pure music portion. Pure music portions located at the head and tail of a music piece are retained for the next processing step, while the other pure music portions are discarded. 4.3 Clustering The aim of music summarization is to analyze a given musical sequence and extract the important frames to reflect the salient theme of the music. The critical issue here is how to measure the similarity for the different resulting frames and how to group these frames. Current approaches for performing such tasks can be classified into two main categories: machine learning approaches and pattern matching approaches. Machine learning approaches [Logan and Chu 2000; Dannenberg and Hu 2002; Lu and Zhang 2003] attempt to use clustering methods to categorize each frame of a song into a certain cluster based on the similarity distance between this frame and other frames in the same song. Then, the frame number of each cluster is used to measure the occurrence frequency. The final summary is generated based on the cluster that contains the largest number of frames. Pattern matching approaches [Bartsch and Wakefield 2001; Cooper and Foote 2002; Chai and Vercoe 2003] aim at matching the underlying candidate excerpts, which include a fixed number of continuous frames, with the whole song. The final summary is generated based on the best matching excerpt.

Automatic Summarization of Music Videos 135 All of the aforementioned methods use the fixed overlapping rate segmentation scheme to segment the music frames. However, in the initial stage it is difficult to determine the exact proper length of the overlapping frames. As a result, fixed overlapping rate segmentation cannot guarantee ideal results for frame grouping. In our proposed method, based on the calculated features of each frame, we use an adaptive clustering method to group the music frames and obtain the structure of the music. Two of the issues associated with music segmentation are the length and degree of overlap of the segmentation window. An inappropriate choice of either one of these two will affect the final clustering result. For speech signals, a typical segmentation window size is 20 ms, as the speech signal is generally treated as stationary over such time intervals. Considering popular music, the tempo of a song is constrained between 30 150 M.M (Mälzel s metronome: the number of quarter notes per minute) and is almost constant [Scheirer 1998], and the signals between two notes can be thought of as being stationary. Therefore, the time interval between two quarter-notes can range from 400 ms 2000 ms (the time intervals for quaver and semiquaver notes are multiples of the time interval of quarter-notes). We choose the smaller one as our segmentation window size. As mentioned, the overlapping length of adjacent frames is another issue associated with music segmentation. If the overlapping length is too long, the redundancy of the two adjacent frames will be high, while on the other hand, if it is too short, the time resolution of the signals will be low. In the initial stage, it is difficult to exactly determine the proper length for the overlapping frames. But we can adaptively adjust the overlapping length if the clustering result is not ideal for frame grouping. This is the key point in our algorithm, which differs from the nonadaptive clustering algorithm proposed in [Logan and Chu 2000]. The clustering algorithm is described as follows. (1) Segment the music signal (vocal or pure music) into w (400 ms, in this case) fixed-length and λ p % overlapping frames, and label each frame with a number i (i = 1, 2,...,n), where the overlapping rate is λ p = 10*p, (p = 1, 2, 3, 4, 5, 6). Here, we vary λ p at a step of 10 (empirically derived) because a smaller step (i.e., 1 or 2) will make our algorithm too computationally complex. (2) For each frame, calculate the music features to form a feature vector: V i = (LPCC i, ZCR i, MFCC i ) i = 1, 2,..., n. (2) (3) Calculate the distance between every pair of the music frames i and j using the Mahalanobis distance [Sun et al. 2001]: D M ( V i, V j ) = [ V i V j ]R 1 [ V i V j ] i j, (3) where R is the covariance matrix of the feature vector. The reason we use the Mahalanobis distance is that it is very sensitive to intervariable changes in all dimensions of the data. Since R 1 is symmetric, it is a semimatrix or positive matrix. It can be diagonalized as R 1 = P T P, where is a diagonal matrix and P is an orthogonal matrix. Thus, Equation (3) can be simplified in terms of Euclidean distance as follows: D M ( V i, V j ) = D E ( P V i, P V j ). (4) Since and P can be computed directly fromr 1, the computational complexity of the vector distance can be reduced from O(n 2 )too(n). (4) Embed the calculated distances into a 2-D matrix which contains the similarity metric calculated for all frame combinations, hence, the frame indexes i and j such that the (i, j)th element of is D(i, j). (5) Normalize the matrix according to the greatest distance between frames, that is, 0 D(i, j) 1.

136 X. Shao et al. (6) For a given overlapping rate λ p, calculate the summation of the total distance between all frames, denoted as S d, which is defined as follows: S d = n 1 n i=1 j=i+1 D(i, j). (5) (7) Repeat steps (1) (6) by varying the overlapping rate λ p, an optimal λ p can be found which can give the maximum value for S d. In our experiment, we found that about 80% of the songs had the optimal λ p = 30, about 18% of the songs had the optimal λ p = 20 and 40, and less than 2% of the songs had the optimal λ p taking the other values, that is, 10, 50, and 60. (8) Perform agglomerative hierarchical clustering [Duda et al. 2000]. Here, we consider putting n music frames into C clusters. At the initial stage, we start with n singleton clusters and form C clusters by successive merging in a bottom-up manner. Here, C is the optimal desired number of clusters, which can be defined as follows: C = k Lsum where L sum is the time length of the music summary (in seconds) and T c T c, (6) is the minimum time length of the subsummary generated in a cluster (for subsummary generation, see Section 4.4 for details). Factor k is the magnification constant selected in the experiment and it is better to select the number of clusters k times more than the required number of clusters to guarantee that enough clusters are selected in the summary. Our human study experiment has shown that the ideal time length of a subsummary is between 3 and 5 s. A playback time which is shorter than 3 s will be nonsmooth and have an unacceptable music quality, while a playback time which is longer than 5 s will be a lengthy and slow-paced. Thus, T c = 3 has been selected for our experiment. The detailed procedure for agglomerative hierarchical clustering can be described as following: Procedure (1) Let Ĉ = n, V i H i, i = 1,...,n, where Ĉ is the initial number of clusters and H i denotes the ith cluster. Initially, one cluster contains one frame. (2) If Ĉ = C, stop. C is the desired number of clusters. (3) Find the nearest pair of distinct clusters, H i and H j, where i and j are cluster indexes. (4) Merge H i and H j, delete H j, and Ĉ Ĉ 1. (5) Go to step (2) At any level, the distance between the nearest clusters can be used as the dissimilarity value for that level. Dissimilarity measures can be calculated by where m i and m j are mean values of the cluster H i and H j. d mean (H i, H j ) = m i m j, (7) 4.4 Summary Generation After clustering, the structure of the music content can be obtained. Each cluster contains frames with similar features. The summary can be generated both in terms of this structure and with domainspecific music knowledge. According to music theory, the most distinctive or representative musical themes should repetitively occur over the duration of the entire piece [Eugene 1990]. Based on this music

Automatic Summarization of Music Videos 137 knowledge and the clustering results, the summary of a music piece can be generated as follows: Assume the summary length is 1000 L sum ms, the number of clusters is C, and the music frame length is w ms. (1) The total number of music frames in the summary can be calculated as: n total = 1000 L sum w λ p %, (8) (1 λ p %) w where λ p is the overlapping rate defined in Section 4.3. The equation can be derived from the fact that the final summary (with the length of 1000 L sum ms) is padded by n total overlapped music frames with a w ms frame length and a λ p % overlapping rate. (2) According to the cluster mean distance matrix, we arrange the distance between cluster pairs in descending order and the higher distance clusters are selected for generating the summary so as to maximize the coverage of musical contents in the final summary. (3) Subsummaries are generated within the cluster. Selected frames in the cluster must be as continuous as possible and the length of the combined frames within the cluster should be approximately 3s 5s or the number of frames should be between n s frames and n e frames, where: n s = 3000 w λ p% (9) (1 λ p %) w and n e = 5000 w λ p% (1 λ p %) w. (10) Assume F i and F j are the first frame and last frame, respectively, in the time domain of a selected cluster such that ( j > i) and n c = ( j i) > 1. From music theory and our user study experiment, a piece of music with discontinuous frames is not acceptable to human ears. Based on this, we should generate continuous subsummaries. If frames are discontinuous between frame F i and frame F j, we first add frames between F i and F j, make the frames in this cluster continuous, and simultaneously delete these added frames from other clusters; we then follow conditions (1), (2), or (3) to adjust the subsummary length within the cluster to meet the subsummary length requirement defined in Equation (9) and Equation (10). Condition (1). n c < n s, as Figure 7(a) shows, we add frames before the head (F i ) and after the tail (F j ) until the subsummary length is equal to n s. Assume x represents the required number of added frames before F i (head frame), and y represents the required number of the added frames after F j (tail frame). Initially, x should be approximate to y, which means the added frames before F i and after F j are distributed in a balanced manner. Therefore, x and y can be calculated as: x = (n s n c )/2 (11) y = n s x. (12) However, if the added frames exceed the first or last frame of the original music, then the exceeding frames will be added to the tail or the head, respectively. After adjustment, the actual number of added frames before F i and after F j, denoted as x and y, respectively, can be calculated as in following: x = i 1; y = y + (x x ) (13) y = (n j) + 1; x = x + (y y ), (14) where n is the total number of frames in the music.

138 X. Shao et al. Fig. 7. Subsummaries generation. Equation (13) calculates the actual number of added frames before F i and after F j when the required number of added frames before head frame F i exceeds the first frame of the original music. The actual number of added frames before F i is (i 1) and the rest of the frames of x will be added to the tail. Therefore, the actual number of the added frames after F j is y+ (x x ). A similar analysis can also be applied to Equation (14), which calculates the actual number of added frames before F i and after F j, when the required number of added frames after the tail frame F j exceeds the last frame of the original music. Condition (2). n s n c n e, as Figure 7(b) shows, no change is made in the subsummary length and it is equal to n c. Condition (3). (n c >n e ), as Figure 7(c) shows, we delete frames both from the head frame and the tail frame until the subsummary length is equal to n e. (4) Repeat step (3) to generate individual subsummaries for another selected cluster and stop the process when the summation of the subsummary length is equal to, or slightly greater than, the required summary length. (5) If the summation of the subsummary length exceeds the required summary length, we find the last subsummary added to the music summary and adjust its length to fit the final summary length. (6) Merge these subsummaries according to their positions in the original music to generate the final summary. 5. SHOT DETECTION AND CLUSTERING After music summarization, we need to turn the raw video sequence into a structured data set W (named the clustered shot set here), where the boundaries of all camera shots are identified and visually similar shots are grouped together. In the clustered shot set W, every pair of clusters in W must be visually different, and all the shots belonging to the same cluster must be visually similar. The total number of clusters varies depending on the internal structure of the original video. It has been shown [Gong et al. 2001] that video programs with more than one shot cluster, where each has an equal time length, will have minimum redundancy. It has been also mentioned that for

Automatic Summarization of Music Videos 139 the purpose of reviewing the visual content, the ideal playback length for each shot cluster is between 1.5 and 2.0 s. A playback time which is shorter than 1.5 s will result in a nonsmooth and choppy video, while a playback time which is longer than 2.0 s will yield a lengthy and slow-paced one. Therefore, given a clustered shot set W, the video sequence with the minimum redundancy measure is the one in which all the shot clusters have a uniform occurrence probability and an equal time length of 1.5 s. Based on these criteria, our video summarization method creates video summaries using the following steps: (1) Segment the video into individual camera shots using the method in Gong et al. [2001]. The output of this step is a shot set S ={s 1, s 2,...,s i,...,s n }, where s i represents the ith shot detected and n is the total number of shots detected. (2) Group the camera shots into a clustered shot set W based on their visual similarities. The similarity between two detected shots can be represented by their key frames. For each shot s i. ES, we choose the key frame f i as the representative frame of that shot. We choose the middle frame of a shot as the key frame, rather than the two ends of a shot, because the shot boundaries commonly contain transition frames. When comparing the visual similarities of two different shots, we calculate the difference between the two key frames related to these two shots using color histograms: D v (i, j) = k=1..n e=y,u,v h e i (k) he j (k), (15) where hi e and he j are the histograms of the key frames i and j, respectively. The main difficulty here is that the optimal number of clusters needs to be determined automatically. To solve this problem, we use the adaptive shot clustering method described in Gong et al. [2001]. After this step, the original video sequence can be described by the clustered shot set W ={w 1,w 2,...,w k }. (3) For each cluster, find the shot with the longest length and use it as the representative shot for the cluster. (4) Discard the clusters whose representative shots are shorter than 1.5 s. For those clusters whose representative shots are longer than 1.5 s, we curtail these shots to 1.5 s by truncating the first 1.5 s visual content from these shots. (5) Sort the representative shots of all the clusters by the time code. Now, we have the representative shot set U ={u 1, u 2,...,u m }, where m n, and n is the total number of the shots in set S. 6. MUSIC/VIDEO ALIGNMENT The final task towards creating a music video summary is to align the image segments in the video summary with the associated music segments in the music summary. According to Zettl [1999], the visual and audio content combination in the music video can be divided into two categories: the polyphonic structure and the homophonic structure. Based on the analysis in Section 3, we currently use the same alignment scheme for these two music video structures. As mentioned in that section, the goal for alignment is to make the summary smooth and natural, and to generate the summary so as to maximize coverage for both the music and the visual content of the original music video without sacrificing either music or visual parts. Assume that the whole time span L sum of the video summary is divided by alignment into P partitions (required clusters), and that the time length of the partition i is T i. Because each image segment in

140 X. Shao et al. Fig. 8. Alignment operations on image and music. the video summary must be at least L min s long (a time slot equals one L min duration), partition i will provide N i time slots, as shown in Figure 8: N i = T i /L min, (16) and hence, the total number of available time slots becomes : N total = P N 1 i. (17) Recall that for each partition, the time length of the music subsummary lasts for 3 5 s approximately and the time length of a shot is 1.5 s. The situation that the sum of the visual shots exceeds the subsummary of the music in the partition will appear. We handle this situation by constraining the last shot of that partition to fit the subsummaries of the music. As shown in Figure 8, T P is the time length of partition P and lasts for 5 s. Four shots are found to fill in this partition, each of which lasts for 1.5 s. The total length of the video subsummary is 6 s, which is longer than the music subsummary. Thus, we curtail the last shot (4) to fit the video subsummary to the music subsummary. Therefore, the alignment problem can be formally described as follows. Given: (1) An ordered set of representative shots U ={u 1, u 2,...,u m }, where m n, and n is the total number of shots in the shot set S. (2) P partitions and N total time slots. To extract: P sets of output shots R = {R 1, R 2,...,R P } which best match between the shot set U and N total time slots, where P = the number of partitions and R i ={r i1,...,r ij,...,r ini } U, i = 1, 2,...,P; N i = T i /L min. Shots r i1,...,r ij,...,r ini are optimal shots selected from the shot set U for the i-th partition. By a proper reformulation, this problem can be converted into a minimum spanning tree (MST) problem [Cormen et al. 2001]. Let G = (V, E) represent an undirected graph with a weighted edge set Vand a finite set of vertices E. The MST of a graph defines the lowest-weight subset of edges that spans the graph in one connected component. To apply the MST to our alignment problem, we use each vertex to represent a representative shots u i, and an edge e ij = (u i, u j ) to represent the similarity between the shots u i and u j.the similarity here is defined as the combination of time similarity and visual similarity. The similarity function is defined as follows: e ij = (1 α)t(i, j) + α ˆD(i, j), (18) where α is a weight coefficient which is set in advance according to the priority given to the visual similarity and the time similarity. The lower α is, the lower the priority for visual similarity and the higher the priority for time similarity, and vice versa. In our experiment, since the time similarity, which indicates time synchronization information, is much more important than the visual similarity, we give the time similarity a higher priority. We set α = 0.2 for all testing samples. ˆD(i, j) and T(i, j) in Equation (18) represent the normalized visual similarity and time similarity, respectively.

Automatic Summarization of Music Videos 141 ˆD(i, j) is defined as follows: ˆD(i, j) = D v (i, j) / max(d v (i, j)), (19) where D v (i, j) is the visual similarity calculated from Equation (15). After being normalized, ˆD(i, j) has a value range from zero to one. T(i, j) is defined as follows: { 1/(Fj L i ) L i < F j, T(i, j) = (20) 0 otherwise where L i is the index of the last frame in the ith shot, and F j is the index of the first frame in the jth shot. Using this equation, the closer two shots are in the time domain, the higher time similarity value they have. Value T(i, j) varies from 0 to 1, and the biggest value of T(i, j) is achieved when shot j just follows shot i and there no other frames between these two shot. Thus, we can create the similarity matrix for all shots in the representative shot set U, and the i, jth element of is e ij. For every partition R i, we generate an MST based on the similarity matrix. To create content rich audio-visual summary, we propose the following alignment operations: (1) Summarize the music track of the music video using the method described in Section 4. The music summary consists of several partitions, each of which lasts for 3 to 5 s. The total duration of the summary is about 30 s. We can get the music summary by adjusting the parameters of the algorithm described in the previous section. (2) Divide each music partition into several time slots, each of which lasts for 1.5 s. (3) For each music partition, we find the corresponding image segment as follows: In the first time slot of the partition, find the corresponding image segment in the time domain. If it exists in the representative shot set U, assign it to the first slot and delete it from the shot set U; if not, identify it in the shot set S, and find the most similar shot in shot set U using the similarity measure defined in Equation (15). We then take this shot as the root, apply the MST algorithm to it, find other shots in the shot set U, and fill them in the subsequent time slots in this partition. Figure 9 illustrates the alignment process, where A(t i, τ i ) and I(t i, τ i ) denote audio and visual segments that start at time instant t i and last for τ i s, respectively. The length of the original video program is 40 s. Assume that the music summarization has selected three partitions A(0, 3), A(13, 5) and A(23, 4), and the shot clustering process has generated the twelve shot clusters shown in Figure 9. As the music summary is generated by A(0, 3), A(13, 5), and A(23, 4), we divide this 12 s summary into 9 time slots. For each slot, we assign a corresponding shot. For the first partition, we assign shot (1) to time slot a and shot (2) to time slot b, respectively. When we assign a shot to the time slot c, there is no corresponding image segment in the time domain. According to our alignment algorithm, we choose shot (4), which is the most similar shot in line with the time index in the shot set S. Then, based on shot (4), we apply the MST algorithm to find other shots for the second partition. For the third partition, in the first time slot g, because the corresponding visual segment (7) has been used by other slots, we have to find the most similar shot to shot (7) in the shot cluster set U. Based on the algorithm described previously, we find shot (8). We then apply the MST algorithm to find the other two shots from this partition. In this way, our proposed summarization scheme can maximize the coverage for both the musical and visual content of the original music video without sacrificing either of them. In the created summary, visual content may not be strictly synchronized with the music. As we mentioned before, an experiment on human perception shows that visual attention is not sensitive to audio-visual asynchrony [Sugana and Iwamiya 2000]. Therefore, within the range of human perception acceptance, the minor deviation of the visual content from the music is allowed. In our method, by giving the time similarity between

142 X. Shao et al. Fig. 9. An example of the audio-visual alignment. the shots a high priority (adjust weight α) we can control the visual deviation from the music so as to keep it in the range of human perception acceptance. 7. EXPERIMENTAL RESULTS AND EVALUATION Our experiment consists of two parts. In the first part, we investigate the accuracy of the SVM classifier which is used to classify the pure music and the vocal music. In the second part, we evaluate the performance of our proposed music video summarization approach. 7.1 Pure Music and Vocal Music Classification In order for the training results to be statistically significant, the training data should be sufficient and should cover various music genres. The training data in our experiment contains four music genres: pop, classical, jazz, and rock. Each genre contains fifty 20 s music excerpts collected from music CDs and the Internet, and each excerpt is hand-labeled to indicate where the singing (if any) begins and ends. All data has a 44.1 khz sample rate, stereo channels, and 16 bits per sample. The training data is segmented into fixed-length and overlapping frames (in our experiment, we used 20 ms frames overlapping 50%). Features such as LPCC (10 dimensions), MFCCs (19 dimensions), and ZCR (1 dimension) are calculated from each frame. The training data of the vocal and pure music frames is assigned to classes +1 and 1, respectively, according to the labels. We use SVM-Light, which is available in http://svmlight.joachims.org/.

Automatic Summarization of Music Videos 143 Table I. SVM Classification for Pure Music and Vocal Music Test Set 1 Test Set 2 Test Set 3 Error rate 0.17% 6.66% 3.98% Table II. The Content of Top of the World Section Range (Frame Number) Content 1 0 20 Instrumental music as intro 2 21 176 Verse by the female singer 3 177 227 Chorus by male and female singer 4 228 248 Instrumental music as bridge 5 249 450 Verse by the female singer 6 451 504 Chorus by male and female singer 7 505 513 Instrumental music as outro We employ the radial basic function (RBF) with a Gaussian kernel as the kernel function in SVM training. The radial basic function (RBF) with a Gaussian kernel can be defined as follows: K(x, x i ) = exp( x x i 2 /c), (21) where x denotes the vector drawn from the input space, x i represents training vectors (i = 1...n), and c is the width of a Gaussian kernel. In our experiment, we set c = 2. After creating the training SVM, we use it as the classifier to classify the vocal and pure music on the test set. Since we perform system testing on the held-out data that is not used in tuning the system parameters, we evaluate our SVM classifier on the new data set. The test set is divided into three parts. The first part contains 20 pure music excerpts, each lasting 15 s. The second part contains 20 vocal music excerpts, each lasting 15 s. The third part contains 20 music songs (whole song). All excerpts and songs in each part are selected from the four music genres. For each excerpt or song, the vocal portions are also hand-labeled. We need to calculate the same features and labels for each excerpt or song, run the classifiers, and compare the results. Table I shows the SVM classification result for pure music and vocal music. In the frame-level, the classifier achieves higher accuracy on the pure music test set (0.17% error rate) than on the vocal music test set (6.66% error rate), while for the third test set, which contains 20 complete songs, the average error rate is 3.98%. Since our purpose is to identify the intro/outro part of a song and filter out the music bridge in the principal part, such a small error rate can be further absorbed by some heuristic rules, that is, if the vocal portion in a continuous pure music segment is less than 4%, this segment still can be identified as pure music. 7.2 Music Video Summarization Evaluation 7.2.1 Objective Evaluation. Our aim for the music video summarization is to maximize the coverage of both musical and visual content without having to sacrifice either of them. For this purpose, in the music track, we need to extract the most common and salient themes of a given piece of music. Ideally, a music summary lasting for long time should fully contain the music summary lasting for a short time. Table II shows the musical content of our testing music video Top of the World (by Carpenter). Sections 1 and 7 are the intro and outro, respectively, of the whole music track, while Sections 2 6 compose the principal part. Sections 2 and 5 are verses by the female singer and Sections 3 and 6 are the chorus by male and female singers, and Section 4 is the bridge portion. In this example, Sections 5 and 6 are refrains of Sections 2 and 3. For Sections 1, 4, and 7, our method will filter the instrumental music out, and it will perform a music summarization process on the vocal music parts in Sections 2, 3, 5, and 6.

144 X. Shao et al. Fig. 10. Results of experiment on the music video Top of the World. Music summaries are extracted with respect to changes of the summary length, as shown in Figure 10. The vertical axis represents the summary length, and the horizontal axis represents the frame number. The bar in the figure corresponds to the frames extracted from the original music. The result shows that the music summary is located at the beginning of the first verse portion and the later part of the two chorus portions. This excerpt is selected because the most salient themes of the music occurred commonly in the memorable introductory theme and the later part of the chorus. Therefore, our proposed music summarization method is able to capture the main themes of the musical work. 7.2.2 Subjective Evaluation. Since there is no absolute measure available to evaluate the quality of the music summary/music video summary, we employed a subjective user study to evaluate the performance of our music summarization method and music video summarization method. The study is borrowed from the idea of the questionnaire for user interaction satisfaction (QUIS) formulated by the Department of Psychology at the University of Maryland [Chin et al. 1988]. We use the following attributes to evaluate the music summary/music video summary: (a) Clarity. This pertains to the clearness and comprehensibility of the music video summary. (b) Conciseness. This pertains to the terseness of the music summary/music video summary and to how much of it captures the essence of the music/music video. (c) Coherence. This pertains to the consistency and natural drift of the segments in the music summary/music video summary. (d) Overall Quality. This pertains to the general perception or reaction of the users to the music summaries/music video summaries. For the dataset, four genres of the music video are used in the test. They are pop, classical, rock and jazz. Each genre contains five music video samples. The aim of providing different music videos from the different genres is to determine the effectiveness of the proposed method in creating summaries of different genres. The length of music video testing samples ranges from 2 m 52 s to 3m33s.The length of the summary for each sample is 30 s. In this experiment, there are 20 participants with music experience; 12 males and 8 females, with most of the participants being graduate students. Their ages range from 18 to 30 years old. Before the tests, the participants were asked to spend at least half an hour watching each testing sample for as many times as needed till they grasped the theme of this sample.