Extracting Alfred Hitchcock s Know-How by Applying Data Mining Technique

Size: px
Start display at page:

Download "Extracting Alfred Hitchcock s Know-How by Applying Data Mining Technique"

Transcription

1 Extracting Alfred Hitchcock s Know-How by Applying Data Mining Technique Kimiaki Shirahama 1, Yuya Matsuo 1 and Kuniaki Uehara 1 1 Graduate School of Science and Technology, Kobe University, Nada, Kobe, , Japan {kimi, yuya, uehara}@ai.cs.scitec.kobe-u.ac.jp Abstract. Video editing process is the work to produce final video sequences with certain duration by finding and selecting appropriate shots from the video material and connecting them. Then, according to the editor s skills, the video editing differs in qualities from various characteristic patterns. In order to produce excellent videos, this process is generally conducted according to the cinematic rule. Furthermore, in movies, there are several kinds of rhythms associated with contents. We define such rhythms based on the character s appearance. In this paper, we propose the methods of extracting cinematic rules and interesting patterns of rhythms by introducing data mining techniques. Data mining is the technique to discover useful patterns of interest from a vast quantity of data. Finally, we can edit new video by applying the extracted rules to our video editing support system. The resultant video must be produced in similar quality to the video that is used to extract the rules. 1 Introduction The video editing is a work to produce the final video sequences with certain duration by finding and selecting appropriate shots from the original video material and connecting them. In order to produce excellent videos, the selection process is generally conducted according to the specific rules called cinematic rules. The set of cinematic rules is a group of rules to define how the cut connection is made. For example, two cuts cannot be connected to each other, where their shot sizes are greatly different, such as tight shot (TS) and loose shot (LS). Most of the professional movie editors have their own cinematic rules. It is beneficial to have a way to mine their movies and videos for important information or patterns that may be contained within [6]. We will call this task video data mining. The most common way to complete video data mining is to index many different aspects of metadata from the raw (signal) level to the semantic level. Video data mining can be achieved by performing the data mining techniques on the metadata directly. On the other hand, in movies, there are several kinds of rhythms associated with contents. It is easily conceivable that the rhythm of a romantic scene is very slow and that of a suspense scene is very fast. We consider such rhythms as rates of character s appearances. We also extract interesting patterns of rhythms by using the data mining

2 technique. These patterns represent the experience of a professional movie editor and embody heuristics about shot duration and transitions. In the following sections, we discuss designs of video editing systems that use different levels of video data mining techniques. 2 Video Editing System We present our video editing system developed in [3]. In this section, we will explain our video editing system in the context that it can automatically edit the video material according to the given cinematic rules. First of all, this system chooses the shot from the video material according to the cinematic rules stored in the production memory. Each cut of video material is stored in the MySQL database indexed by the metadata shown in Table 1. The engine of the video editing system is forwardchaining production system, implemented in Prolog. The overview of the system is illustrated in Fig. 1. The interface between production system and MySQL database is implemented in Java. The working memory is a collection of items. The production memory is a collection of production rules. The reasoning control unit determines the conflict set, resolves the conflict and fires the appropriate rule. By repeatedly applying the cinematic rule, this system can automatically find and select the appropriate raw video shots. Fig. 1. Video editing system. Fig. 2 shows the snapshot of the editing process. The shot whose ID is 19 could not be a master shot, therefore the value of Master is 0. The value of Shot is TS. This cut is a zoom cut and the shot size changes into middle shot at the end of the cut. So, Shot, End Shot, and Camerawork become TS (Tight Shot), MS (Medium Shot) and Zoom, respectively. The shot with ID 20 in Fig. 2 represents a restaurant exterior of Scene 1. The shot could be regarded as a master shot. Therefore, the value of Master is 1. The head shot size (Shot) of the cut is LS and this cut is a fixed shot, therefore the shot size at the end (Shot End) of the cut is the same as Shot value, LS.

3 Fig. 2. Snapshot of the editing process. The first column of the Table 1 shows the name of the metadata. The next column indicates the explanation of the metadata. Examples of the metadata are shown in the last column. Table 1. Table of metadata. Metadata Explanation Example Id Identification number of the cut 19, 20 Scene A stage name where the same action or location is presented Scene1 Master The cut can be a master shot, or not 0, 1 Shot Head shot size of the cut TS, LS Shot End Last shot size of the cut MS, - Camerawork Camera movement fix, zoom, pan Start Frame The head frame number of the cut 1000, 1501 End Frame The last frame number of the cut 1500, 2000 Camera Start Frame number where camera work has started 1250, - Camera End Frame number where camera work has done 1350, - Use The cut had already been used or not 0, 0 Object The objects what is displayed on the screen character, object Direction The direction of character's gaze front, right Motion The motion type of the character displayed on the screen gaze, motion Sound The sound type of the scene speech, cheep 3 Mining Cinematic Rules from Multi-streams of data In movies, an event consists of a sequence of actions. The video editing is the work to make the video more meaningful and effective. So there must be some relations included among a sequence of actions within an event. For example, in the event where a person is looking at some objects from moderate distance (MS), the close-up of the objects (TS) must be selected as the subsequent cut. We regard such relation as having the association (i.e. MS TS) between these 2 cuts. This is so called a cinematic rule. In order to extract cinematic rules, we propose the mining method that focuses on the correlation between the nearby cuts (i.e. sequential multiple cuts) and the correlation of the metadata of each cut. As illustrated in Fig. 3, we formulate the video stream that is indexed by the raw level metadata Shot, Camerawork, Duration (= (EndFrame StartFrame) / 30), Ob-

4 ject, Direction, Motion and Sound from Table 1. We propose the mining method that can extract the frequent cinematic rules from the multi-streams of data in Fig. 3. Fig. 3. Video stream indexed by the metadata. Our method accepts a set of multi-stream time-series data as input. Each element of the pattern x n = (v 1, v 2,, v 7 ) represents the metadata that can be identified by using our mining method. In order to express the pattern over multiple cuts, it is required to express it as a combination of x n chronological patterns. That is, we denote the pattern over multiple cuts as the form p n = {x 1,, x n } (n > 1). In order to search patterns, we set up a searching window w as shown in Fig. 4. The size of this window is denoted as w s. The searching process starts at position 0 (i.e. Cut No. 0). w s is initialized to 0 and is increased by 1 while the searching process is under way. Increasing the value of w s is equivalent to extending the window size. Whenever extending the window size, it is required to confirm whether the current pattern p i ={x i, y i } satisfies two parameter conditions that determine the importance of the pattern shown in the equation (1) and (2). ap(p i ) = cnt(x i y i ) (1) prob(p i ) = cnt(x i y i ) / cnt(x i ) (2) For the pattern p i, cnt(p i ) represents the number of times the current pattern p i is present in the video stream. ap(p i ) is equal to cnt(p i ). ap(p i ) is essential to assess the frequency of the pattern. We call ap(p i ) appearance. prob(p i ) represents the value of conditional probability that shows the appearance y i when x i appears. prob(p i ) is essential to assess the appearance confidence of the pattern. We call prob(p i ) probability. When confirming the two threshold conditions, the elements that do not satisfy both of them should be disregarded and consequently denoted as -. If the efficient patterns that satisfy the threshold conditions remain, we extend the window size, and continue the searching process. We repeat this process until all elements in x i (i 1) are replaced by -. If the algorithm cannot extend the window size any more, at first w s is initialized to 0. Next, the starting position of searching is shifted to the right, and then the same process is started again. From Fig. 4, the extracted pattern at w s = 3 is denoted as p x ={(LS, fix, t < 6, A, R, G), (MS, -, t < 4, B, L, -), (TS, fix, -, A, R, M)}. Such a pattern is often found in the scene when the first cut is LS, MS must be inserted as the next cut to avoid the rapid transition. This is one of the cinematic rules of zoom-in, illustrated in Fig. 5.

5 Fig. 4. The process of extending patterns. Fig. 5. The cinematic rule of zoom in over the sequential multiple cuts. As stated above, we can extract the cinematic rules that frequently occur in multistreams of data. In related work [5], the searching method doesn t have the flexibility in many respects. For example, while the method in [5] searches the pattern that completely matches in multi-streams of data, our method has the flexibility to search by setting two threshold conditions so as to extract frequent patterns, and can complete the search quickly. Furthermore, our method has the flexibility to search within a window that can extend its size. So, according to the searching position, we can effectively look for frequently occurring patterns. We have implemented our method to extract cinematic rules. The videos used for this purpose are the movies Madadayo directed by Akira Kurosawa, Star Wars Episode I directed by George Lucas and PSYCHO directed by Alfred Hitchcock. First, we show the results of cinematic rule extraction in Table 2 as applied in Madadayo. From Table 2, we can find that his editing technique doesn t allow connecting to TS (LS TS, MS TS), so the extracted cinematic rules represent the continuously connected pattern (LS LS, LS MS) between LS and MS. Furthermore, the duration of each cut is very long compared with the duration written in video grammar. In this movie, durations of LS and MS are more than 10 and 6 seconds respectively. Table 2. Results of cinematic rule extraction from Madadayo. cinematic rule ap(p i )prob(pi) 1 {(LS, fix, t>10 ),(LS, fix, t>10 ),(LS, -,- ),(LS,-,- )} {(LS, fix, t>10 ),(MS, fix, t>6 ),(LS, fix, t>10 ),(MS,-,- )} {(MS, fix, t>6 ),(LS, fix, t>10 ),(MS,-,t>6 ),(LS, fix, - )} Next, in Table 3, we show the results obtained by applying our method to Star Wars Episode I. From Table 3, we can find that this movie is well connected be-

6 tween the same shot sizes (MS MS, TS TS), and is edited in quick rhythm compared with the previous movie Madadayo. In this movie, durations of LS, MS and TS are about 4, 3 and 2.5 seconds respectively. These cinematic rules characterize the speedy development of this movie. Furthermore, since the rapid transition of the shot size such as LS TS, TS LS makes it hard for the viewer to understand what is going on, these cinematic rules are not allowed especially in such a speedy development. In fact, they are not extracted in Table 3. Table 3. Results of cinematic rule extraction from Star Wars Episode I. cinematic rule ap(p i) prob(pi) 1 {(MS,- ), (MS, 2<t<4 ), (MS,- )} {(MS,- ), (MS,- ), (MS, t<2 )} {(TS, 2<t<4 ), (MS,- ), (TS,- )} {(TS,- ), (TS,- ), (TS, 2<t <4 )} Finally, in Table 4, we show the results obtained by applying our method to PSYCHO. We can say that this video is normally edited like video grammar considering the following factors: Firstly, from Table 4, we can find that this movie is edited efficiently without rapid transition of shot size such as LS MS, MS MS and TS LS. Secondly, durations of LS, MS and TS are about 6, 4 and 3 seconds respectively. Table 4. Results of cinematic rule extraction from PSYCHO. cinematic rule ap(p i ) prob(p i ) 1 {(LS, fix, t < 6 ),(MS, fix, t < 4 ),(LS, fix, t < 6 )} {(LS, fix, t < 6 ),(MS, fix, t > 4 ), (LS, fix, - )} {(MS, fix, t < 6 ),(MS, fix, t < 4 ),(MS, fix, - )} {(TS, fix, t < 3, speech ),(TS, fix, -, speech ),(TS, fix, t > 3, speech )} {(TS, fix, t > 3, speech ),(TS, fix, -, speech ),(TS, fix, -, speech )} {(TS, fix, t < 6, Marion, F, Gaze ),(LS, fix, t < 6 ),(TS, fix, t < 6, Marion, F, Gaze )} {(TS, fix, -, R, Gaze, speech ),(TS, fix, -, L, Gaze, speech ),(TS, fix, t > 3, R, Gaze, speech )} Topic Mining Based on Burst Streams of Character Appearance Usually in movies, there are some editing rhythms associated with contents. Two examples of such rhythms in Alfred Hitchcock s movie PSYCHO, are shown in the two tables of Fig. 6. Each table represents several cuts in a talking scene. In SCENE 1, the woman (Marion) and the man (Sam) are talking to each other while hugging in a hotel room. In SCENE 2, the policeman finds the woman (Marion) who stole a lot of money, sleeping in a car. She is surprised by being found and tries to escape from him. However, the policeman stops her, and the interrogation begins. As we can conclude, SCENE 1 is a romantic scene, while SCENE 2 is a tense scene.

7 The numeric values in the bottom row of each table represent the length of intervals of Marion s appearance between consecutive cuts. For example, the rate of Marion s appearance in SCENE 2 is clearly more frequent than that in SCENE 1. That is in SCENE 1, the intervals of Marion s appearance are 20.3, 65.3, 45.1 seconds, etc, whereas in SCENE 2, they are 3.7, 10.0, 1.7 seconds, etc. Thus, depending on the type of the scene, rates of character s appearance differ. We call the rate of character s appearance a rhythm. For a character, we define a burst as a rate of his/her appearance. An approach for modeling bursts in streams has been introduced by [2]. According to this approach, the stream consisting of messages relevant to a single topic of interest is divided into sub-topics by rates of message arrivals (i.e. bursts). The method in Section 4.1 for modeling the character s burst stream is based on a similar approach. We define a topic as the interval where a character appears at the same rate, that is, bursts maintain constant intensity. A rhythm of topics is defined as a transition of that character s topics in a scene. In Section 4.2, we will classify of talking scenes into groups of similar rhythms. Then, we investigate how a talking scene is characterized by rhythms of topics specific for the characters. As a result, we can use the pattern of rhythms for the creation of a new video material. We refer to this approach for extraction of useful patterns as topic mining. 4.1 Character s Burst Stream We index the following metadata for every cut as shown in each table of Fig. 6. Here the information about a cut is represented in a column. StartTime and EndTime (the third entry) are raw level metadata representing the starting point and the ending point of a cut, respectively. SkipTime (the fourth entry) is semantic level metadata that roughly represents a span of time to be found between the events presented in consecutive cuts. In this paper, SkipTime is classified into four levels, that is, continuous, a little distant, distant and very distant. In Fig. 6, all SkipTime values are continuous, because the conversation of Marion and Sam (or the policeman) does not have any gaps between consecutive cuts. Characters names (the fifth entry) represents their appearances occurring in the cut. There are two kinds of time categories in a movie. One is the time devoted to the actual movie as it is presented, time1. Another, time2, is the time semantically spanned by the events presented in a movie. Therefore, the current StartTime and EndTime represent the starting point and the ending point of each cut in time1 (the third entry in Fig. 6), whereas the character actions are reflected in time2. So we must model his/her burst stream in time2. As a rough translation of time1 into time2 for a cut, the amount of SkipTime values from the beginning to the previous cuts of the movie is added to StartTime and EndTime of the current cut (the sixth entry in Fig. 6). For example, in SCENE 2 of Fig. 6, when Marion is caught by the policeman (cut 92), 741 seconds (12 minutes 21 seconds) have passed since PSYCHO started as indicated in time1, but in the overall story seconds have passed, as reflected in time2.

8 Fig. 6. Examples of rhythms in a romantic and a tense scene. Based on [2], for cuts where Marion appears, we can obtain her burst stream from time2. Two segments of Marion s burst stream corresponding to SCENE 1 and 2 in Fig. 6 are shown in Fig. 7. A dot represents a value of Marion s burst intensity associated with the interval of her burst. The association of Marion s burst intensity values with the intervals of her appearance is shown in Fig. 7. In contrast to scattered intervals of her appearance like in Fig.6, her burst intensity graph is smoothly plotted. The reason for this is that our method prevents frequent changes of burst intensity values triggered by insignificant small changes in intervals of her appearance. That is, this method assigns a burst intensity to a cut in which Marion appears, only if she appears at a sufficient rate to be associated with the burst intensity. Fig. 7. Two segments of Marion s burst stream (i.e. two rhythms of topics). Since Marion s topic is defined as an interval where burst intensity value keep constant, in SCENE 1, [cut 5, cut 8], [cut 10, cut 11], [cut 12, cut 21] and [cut 23, -]

9 represent four Marion s topics. Note that, her actions also tend to change at the cut where Marion s topic changes. For example, Marion and Sam talk to each other while hugging, and then they step away for a while, between cut 5 and 8. Then they come close to each other and hug again between cut 9 and 11. Being out of the scene, Marion doesn t appear in cut 9. She appears for the first time again in cut 10. Hence, her topic changes at cut 10. Therefore, regarding her appearance, SCENE 1 can be divided into four topics, each topic corresponding to one semantically relevant action. This method is applicable to any other scene where Marion appears. The two segments in Fig. 7 correspond to two rhythms of topics for Marion. In SCENE 1, Marion s burst stream gradually grows from intensity 3 corresponding to long interval of her appearance, 49.1 seconds, to intensity 7 associated with 12.8 seconds. Her actions turn from hugging Sam into talking to him, resulting in a higher rate of appearance. On the other hand, at the beginning of SCENE 2, Marion s burst stream suddenly grows from intensity 0 to intensity 9 associated with a relatively short interval of her appearance, 6.5 seconds. Consequently, her burst stream remains at high intensity in SCENE 2. In this scene, the quick switch of cuts between Marion and the policeman characterizes the tense conversation. One of the attributes of character s burst stream is namely growth. The above investigation suggests that character s actions in a scene affect on the way of growth of his/her burst stream. In the case of a burst stream for a different character (e.g. Sam), intervals of his appearance associated with burst intensities differ from those of Marion. Hence, we must consider intervals of character s appearance in close relation with burst intensity of the character. Another attribute of character s burst stream is intervals. Both attributes are represented as time sequence. For character s burst stream, the attribute growth is a sequence of burst intensities, and the interval is a sequence of intervals of his/her appearance associated with burst intensities. Two segments of Marion s burst stream in SCENE 1 and 2 (i.e. two rhythms of topics for Marion) are also shown in Fig Clustering Rhythms of Topics We can extract various rhythms of topics for characters that appear in talking scenes in PSYCHO. For each pair of rhythms, we compute similarities on each attribute by using Dynamic Time Warping (DTW) [1]. DTW computes the similarity between a pair of time series by aligning them to achieve a reasonable fit, where each time axis for the time series is stretched (or compressed). A smaller value of similarity means that a pair of time series is more similar to each other. DTW is so efficient to compute similarities on intervals, although it is not effective to compute growth values. A pair of rhythms may grow from different starting points of burst intensity, and we need to know how subsequent growth is similar. For a pair of rhythms, we apply DTW by incrementing the intensity values of either one of the rhythms. When the minimum difference is obtained, we have achieved maximum similarity between the two rhythms. Similarities in growth and intervals for three rhythms are shown in Fig. 8. Two of them correspond to the ones in Fig. 7, which are extracted for Marion in the romantic

10 and the tense scene. The remaining rhythm for Sam is extracted in the tense scene, where he is talking to Norman who killed Marion, while carefully choosing his words to get some information about her. As shown in Fig. 8, the two rhythms extracted from the tense scenes are similar to each other on both growth and intervals, whereas the rhythm extracted from the romantic scene isn t similar to either rhythm. Fig. 8. (Dis) Similarities for three rhythms of topics. We employ Similarity-Based Agglomerative Clustering (SBAC) [4] to compute clusters consisting of similar rhythms. SBAC provides a unified framework for clustering objects having nominal and numeric attribute values, without making any assumptions of the underlying distributions of the attribute values. We extend SBAC to handle rhythms of topics having two time series attributes (i.e. growth and interval) based on similarities computed by DTW. Here, we describe only the extended part of SBAC (See [4] for detail). For a pair of rhythms, SBAC computes the similarity contribution on each attribute, taking into account the degree of the similarity computed by DTW and the uniqueness of the pair of time series. The uniqueness value is defined as the number of time series, which poses a greater similarity to both time series than the original time series themselves. That is, the uniqueness can be considered as the density of time series encompassed by the pair of time series. For a pair of rhythms, dissimilarity score on each attribute is measured based on the similarity contribution of the pair of time series. The subsequent procedure follows [4]. For the three rhythms above in Fig. 8, dissimilarity values on growth and intervals are shown in Fig. 8. On each attribute, a dissimilarity score is a probabilistic value, and a pair having the value closer to 1 is considered to be more dissimilar. Fig. 8 shows that, the pair of two Marion s rhythms in the romantic and the tense scene is one of the most dissimilar pair regarding growth, because the dissimilarity score is For these rhythms, overall dissimilarity scores computed by aggregating the dissimilarity scores on growth and intervals are also shown in Fig. 8. Based on overall dissimilarity scores for all rhythms extracted for characters in talking scenes of PSYCHO, a nested sequence of partitions of rhythms in the form of a dendrogram is obtained. As a result of investigating the rhythms of topics, some interesting patterns that characterize a talking scene are discovered. Fig. 9 shows the outlines of those patterns of rhythms.

11 Fig. 9. Extracted patterns by applying our topics mining method to PSYCHO. At first, we compare the leftmost pattern with the rest of the patterns. The main difference between them is that a rhythm in the leftmost pattern doesn t contain any topics associated with short intervals of character s appearance, but a rhythm in the remaining certainly contains some topics associated with short intervals. Then it is revealed that the character and rhythm in the leftmost pattern doesn t contain active interaction between his/her talking partners. For example, a character in the leftmost pattern just listens to the conversation of other characters that actively interact with each other. In contrast, a character in the remaining patterns interacts actively with his/her partners in topics associated with short intervals. This annotation can be applied to a character s topic where one semantically relevant action is presented as described in Section 4.1. Furthermore, for the characters that are classified as activity interacting with each other, their topics are associated with short intervals. In that way we can know the transition of characters interactions during a talking scene. In the topics associated with short intervals of character s appearance, the conversation is heated. In such a case, tension grows, and we consider it as the climax of a talking scene. The right-hand three patterns in Fig. 9 are three types of patterns of rhythms to reach the climax in a talking scene. The left pattern of the three has a rhythm represented by a burst stream that hardly grows, and is associated with short intervals. A scene where a character appears in such a rhythm, the climax is in the very beginning. The middle pattern has a rhythm represented by a burst stream that suddenly grows from low burst intensity to a high intensity. A scene where a character appears in such a suddenly intensified rhythm, the climax is quickly reached. We consider that these two types of rhythms are characteristic of the tense talking scenes. The right one has a rhythm represented by a burst stream that gradually grows into high burst intensity. A scene where all characters appear in such gradually intensifying rhythms, reaches its climax without having any strong impact on audience. We consider this scene as calm. In PSYCHO, there is only one calm talking scene (i.e. SCENE 1 in Fig. 8). Accordingly, in order to raise the spectator s tension, in most talking scenes of PSYCHO, some characters must have rhythms similar to the two middle patterns in Fig. 9.

12 5 Conclusion and Future Work In this paper, we proposed the methods of pattern extraction by using data mining techniques from movies. But some problems and future work still remain. They are the following: In Section 3, we proposed effective searching method. The size of multi-streams of data illustrated in Fig. 4 is very large, so the efficient searching method is needed. The problem of our method is at first to enumerate the number of times of all the possible patterns, so it requires much time to complete it. Then, the Knuth - Morris - Pratt (KMP) and Boyer - Moore (BM) algorithms are the efficient string-matching ones. BM algorithm scans the characters of the pattern from right to left beginning with the rightmost one. In case of a mismatch, it uses two functions to shift the window to the right. In our method, if we intend to search the pattern of length m from the strings of length n, it takes O(mn) time complexity. KMP and BM can search in O(n) and O(n/m) time complexity, respectively. So applying these methods to our video streams, we can complete the search more quickly. In Section 4, we proposed a topic mining method. But at present, it is not robust enough in terms of video segmentation. In order to extract a rhythm of topics for a character, we manually divide a movie into scenes. That is, the quality of segmentation depends on the person who manually segments the movie. Furthermore, we described that, by using character s burst stream, the movie can be divided into topics based on that character s perspective, each topic corresponding to one semantically relevant action. However, as it is often the case, there appears several characters in a scene that interact with each other. Therefore, analyzing burst streams for all characters may lead to discovering an effective method for segmentation of the movie into scenes. References 1. D. J. Berndt and J. Clifford: Finding Patterns in Time Series: A Dynamic Programming Approach. U. M. Fayyad, G. Piatetsky-Shapiro, P. Smyth and R. Uthurusamy eds. Advance in Knowledge Discovery and Data Mining, AAAI Press, pages , J. Kleinberg: Bursty and Hierarchical Structure in Streams. In Proc. of the 8th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages , M. Kumano, Y. Ariki, M. Amano, K. Uehara, K. Shunto and K. Tsukada: Video Editing Support System Based on Video Grammar and Content Analysis. In Proc. of 16th International Conference on Pattern Recognition (ICPR), pages , C. Li and G. Biswas: Unsupervised Learning with Mixed Numeric and Nominal Data. IEEE Transactions on Knowledge and Data Engineering, Vol. 14, No. 4, pages , T. Oates and P. R. Cohen: Searching for Structure in Multiple Streams of Data. In Proc. of 13th International Conference on Machine Learning, pages , D. Wijesekera and D. Barbara: Mining Cinematic Knowledge: Work in Progress. In Proc. of the International Workshop on Multimedia Data Mining, pages , 2000.

Reducing False Positives in Video Shot Detection

Reducing False Positives in Video Shot Detection Reducing False Positives in Video Shot Detection Nithya Manickam Computer Science & Engineering Department Indian Institute of Technology, Bombay Powai, India - 400076 mnitya@cse.iitb.ac.in Sharat Chandran

More information

Browsing News and Talk Video on a Consumer Electronics Platform Using Face Detection

Browsing News and Talk Video on a Consumer Electronics Platform Using Face Detection Browsing News and Talk Video on a Consumer Electronics Platform Using Face Detection Kadir A. Peker, Ajay Divakaran, Tom Lanning Mitsubishi Electric Research Laboratories, Cambridge, MA, USA {peker,ajayd,}@merl.com

More information

Automated extraction of motivic patterns and application to the analysis of Debussy s Syrinx

Automated extraction of motivic patterns and application to the analysis of Debussy s Syrinx Automated extraction of motivic patterns and application to the analysis of Debussy s Syrinx Olivier Lartillot University of Jyväskylä, Finland lartillo@campus.jyu.fi 1. General Framework 1.1. Motivic

More information

Principles of Video Segmentation Scenarios

Principles of Video Segmentation Scenarios Principles of Video Segmentation Scenarios M. R. KHAMMAR 1, YUNUSA ALI SAI D 1, M. H. MARHABAN 1, F. ZOLFAGHARI 2, 1 Electrical and Electronic Department, Faculty of Engineering University Putra Malaysia,

More information

Wipe Scene Change Detection in Video Sequences

Wipe Scene Change Detection in Video Sequences Wipe Scene Change Detection in Video Sequences W.A.C. Fernando, C.N. Canagarajah, D. R. Bull Image Communications Group, Centre for Communications Research, University of Bristol, Merchant Ventures Building,

More information

Audio-Based Video Editing with Two-Channel Microphone

Audio-Based Video Editing with Two-Channel Microphone Audio-Based Video Editing with Two-Channel Microphone Tetsuya Takiguchi Organization of Advanced Science and Technology Kobe University, Japan takigu@kobe-u.ac.jp Yasuo Ariki Organization of Advanced Science

More information

A Framework for Segmentation of Interview Videos

A Framework for Segmentation of Interview Videos A Framework for Segmentation of Interview Videos Omar Javed, Sohaib Khan, Zeeshan Rasheed, Mubarak Shah Computer Vision Lab School of Electrical Engineering and Computer Science University of Central Florida

More information

Name Identification of People in News Video by Face Matching

Name Identification of People in News Video by Face Matching Name Identification of People in by Face Matching Ichiro IDE ide@is.nagoya-u.ac.jp, ide@nii.ac.jp Takashi OGASAWARA toga@murase.m.is.nagoya-u.ac.jp Graduate School of Information Science, Nagoya University;

More information

VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS. O. Javed, S. Khan, Z. Rasheed, M.Shah. {ojaved, khan, zrasheed,

VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS. O. Javed, S. Khan, Z. Rasheed, M.Shah. {ojaved, khan, zrasheed, VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS O. Javed, S. Khan, Z. Rasheed, M.Shah {ojaved, khan, zrasheed, shah}@cs.ucf.edu Computer Vision Lab School of Electrical Engineering and Computer

More information

Temporal data mining for root-cause analysis of machine faults in automotive assembly lines

Temporal data mining for root-cause analysis of machine faults in automotive assembly lines 1 Temporal data mining for root-cause analysis of machine faults in automotive assembly lines Srivatsan Laxman, Basel Shadid, P. S. Sastry and K. P. Unnikrishnan Abstract arxiv:0904.4608v2 [cs.lg] 30 Apr

More information

Enhancing Music Maps

Enhancing Music Maps Enhancing Music Maps Jakob Frank Vienna University of Technology, Vienna, Austria http://www.ifs.tuwien.ac.at/mir frank@ifs.tuwien.ac.at Abstract. Private as well as commercial music collections keep growing

More information

Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting

Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting Dalwon Jang 1, Seungjae Lee 2, Jun Seok Lee 2, Minho Jin 1, Jin S. Seo 2, Sunil Lee 1 and Chang D. Yoo 1 1 Korea Advanced

More information

Generating Cinematic Camera Shots for Narratives

Generating Cinematic Camera Shots for Narratives Generating Cinematic Camera Shots for Narratives Introduction John Mason CSC725: Intelligent Multimedia Systems Spring 2005 Arnav Jhala We have built a system that automatically generates camera actions

More information

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM A QUER B EAMPLE MUSIC RETRIEVAL ALGORITHM H. HARB AND L. CHEN Maths-Info department, Ecole Centrale de Lyon. 36, av. Guy de Collongue, 69134, Ecully, France, EUROPE E-mail: {hadi.harb, liming.chen}@ec-lyon.fr

More information

FULL-AUTOMATIC DJ MIXING SYSTEM WITH OPTIMAL TEMPO ADJUSTMENT BASED ON MEASUREMENT FUNCTION OF USER DISCOMFORT

FULL-AUTOMATIC DJ MIXING SYSTEM WITH OPTIMAL TEMPO ADJUSTMENT BASED ON MEASUREMENT FUNCTION OF USER DISCOMFORT 10th International Society for Music Information Retrieval Conference (ISMIR 2009) FULL-AUTOMATIC DJ MIXING SYSTEM WITH OPTIMAL TEMPO ADJUSTMENT BASED ON MEASUREMENT FUNCTION OF USER DISCOMFORT Hiromi

More information

Hidden Markov Model based dance recognition

Hidden Markov Model based dance recognition Hidden Markov Model based dance recognition Dragutin Hrenek, Nenad Mikša, Robert Perica, Pavle Prentašić and Boris Trubić University of Zagreb, Faculty of Electrical Engineering and Computing Unska 3,

More information

AUTOMATIC MAPPING OF SCANNED SHEET MUSIC TO AUDIO RECORDINGS

AUTOMATIC MAPPING OF SCANNED SHEET MUSIC TO AUDIO RECORDINGS AUTOMATIC MAPPING OF SCANNED SHEET MUSIC TO AUDIO RECORDINGS Christian Fremerey, Meinard Müller,Frank Kurth, Michael Clausen Computer Science III University of Bonn Bonn, Germany Max-Planck-Institut (MPI)

More information

THE importance of music content analysis for musical

THE importance of music content analysis for musical IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 15, NO. 1, JANUARY 2007 333 Drum Sound Recognition for Polyphonic Audio Signals by Adaptation and Matching of Spectrogram Templates With

More information

Generation of Video Documentaries from Discourse Structures

Generation of Video Documentaries from Discourse Structures Generation of Video Documentaries from Discourse Structures Cesare Rocchi ITC-Irst Trento Italy rocchi@itc.it Massimo Zancanaro ITC-Irst Trento Italy zancana@itc.it Abstract Recent interests in the use

More information

SHOT DETECTION METHOD FOR LOW BIT-RATE VIDEO CODING

SHOT DETECTION METHOD FOR LOW BIT-RATE VIDEO CODING SHOT DETECTION METHOD FOR LOW BIT-RATE VIDEO CODING J. Sastre*, G. Castelló, V. Naranjo Communications Department Polytechnic Univ. of Valencia Valencia, Spain email: Jorsasma@dcom.upv.es J.M. López, A.

More information

METHOD TO DETECT GTTM LOCAL GROUPING BOUNDARIES BASED ON CLUSTERING AND STATISTICAL LEARNING

METHOD TO DETECT GTTM LOCAL GROUPING BOUNDARIES BASED ON CLUSTERING AND STATISTICAL LEARNING Proceedings ICMC SMC 24 4-2 September 24, Athens, Greece METHOD TO DETECT GTTM LOCAL GROUPING BOUNDARIES BASED ON CLUSTERING AND STATISTICAL LEARNING Kouhei Kanamori Masatoshi Hamanaka Junichi Hoshino

More information

Using DICTION. Some Basics. Importing Files. Analyzing Texts

Using DICTION. Some Basics. Importing Files. Analyzing Texts Some Basics 1. DICTION organizes its work units by Projects. Each Project contains three folders: Project Dictionaries, Input, and Output. 2. DICTION has three distinct windows: the Project Explorer window

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

Music Radar: A Web-based Query by Humming System

Music Radar: A Web-based Query by Humming System Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,

More information

Subjective Similarity of Music: Data Collection for Individuality Analysis

Subjective Similarity of Music: Data Collection for Individuality Analysis Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp

More information

CS229 Project Report Polyphonic Piano Transcription

CS229 Project Report Polyphonic Piano Transcription CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project

More information

... A Pseudo-Statistical Approach to Commercial Boundary Detection. Prasanna V Rangarajan Dept of Electrical Engineering Columbia University

... A Pseudo-Statistical Approach to Commercial Boundary Detection. Prasanna V Rangarajan Dept of Electrical Engineering Columbia University A Pseudo-Statistical Approach to Commercial Boundary Detection........ Prasanna V Rangarajan Dept of Electrical Engineering Columbia University pvr2001@columbia.edu 1. Introduction Searching and browsing

More information

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? NICHOLAS BORG AND GEORGE HOKKANEN Abstract. The possibility of a hit song prediction algorithm is both academically interesting and industry motivated.

More information

Automatic Piano Music Transcription

Automatic Piano Music Transcription Automatic Piano Music Transcription Jianyu Fan Qiuhan Wang Xin Li Jianyu.Fan.Gr@dartmouth.edu Qiuhan.Wang.Gr@dartmouth.edu Xi.Li.Gr@dartmouth.edu 1. Introduction Writing down the score while listening

More information

Computer Coordination With Popular Music: A New Research Agenda 1

Computer Coordination With Popular Music: A New Research Agenda 1 Computer Coordination With Popular Music: A New Research Agenda 1 Roger B. Dannenberg roger.dannenberg@cs.cmu.edu http://www.cs.cmu.edu/~rbd School of Computer Science Carnegie Mellon University Pittsburgh,

More information

Smart Traffic Control System Using Image Processing

Smart Traffic Control System Using Image Processing Smart Traffic Control System Using Image Processing Prashant Jadhav 1, Pratiksha Kelkar 2, Kunal Patil 3, Snehal Thorat 4 1234Bachelor of IT, Department of IT, Theem College Of Engineering, Maharashtra,

More information

Adaptive Key Frame Selection for Efficient Video Coding

Adaptive Key Frame Selection for Efficient Video Coding Adaptive Key Frame Selection for Efficient Video Coding Jaebum Jun, Sunyoung Lee, Zanming He, Myungjung Lee, and Euee S. Jang Digital Media Lab., Hanyang University 17 Haengdang-dong, Seongdong-gu, Seoul,

More information

Keywords: Edible fungus, music, production encouragement, synchronization

Keywords: Edible fungus, music, production encouragement, synchronization Advance Journal of Food Science and Technology 6(8): 968-972, 2014 DOI:10.19026/ajfst.6.141 ISSN: 2042-4868; e-issn: 2042-4876 2014 Maxwell Scientific Publication Corp. Submitted: March 14, 2014 Accepted:

More information

EDDY CURRENT IMAGE PROCESSING FOR CRACK SIZE CHARACTERIZATION

EDDY CURRENT IMAGE PROCESSING FOR CRACK SIZE CHARACTERIZATION EDDY CURRENT MAGE PROCESSNG FOR CRACK SZE CHARACTERZATON R.O. McCary General Electric Co., Corporate Research and Development P. 0. Box 8 Schenectady, N. Y. 12309 NTRODUCTON Estimation of crack length

More information

Getting Started. Connect green audio output of SpikerBox/SpikerShield using green cable to your headphones input on iphone/ipad.

Getting Started. Connect green audio output of SpikerBox/SpikerShield using green cable to your headphones input on iphone/ipad. Getting Started First thing you should do is to connect your iphone or ipad to SpikerBox with a green smartphone cable. Green cable comes with designators on each end of the cable ( Smartphone and SpikerBox

More information

Evaluation of Automatic Shot Boundary Detection on a Large Video Test Suite

Evaluation of Automatic Shot Boundary Detection on a Large Video Test Suite Evaluation of Automatic Shot Boundary Detection on a Large Video Test Suite Colin O Toole 1, Alan Smeaton 1, Noel Murphy 2 and Sean Marlow 2 School of Computer Applications 1 & School of Electronic Engineering

More information

Melodic Pattern Segmentation of Polyphonic Music as a Set Partitioning Problem

Melodic Pattern Segmentation of Polyphonic Music as a Set Partitioning Problem Melodic Pattern Segmentation of Polyphonic Music as a Set Partitioning Problem Tsubasa Tanaka and Koichi Fujii Abstract In polyphonic music, melodic patterns (motifs) are frequently imitated or repeated,

More information

Automatically Creating Biomedical Bibliographic Records from Printed Volumes of Old Indexes

Automatically Creating Biomedical Bibliographic Records from Printed Volumes of Old Indexes Automatically Creating Biomedical Bibliographic Records from Printed Volumes of Old Indexes Daniel X. Le and George R. Thoma National Library of Medicine Bethesda, MD 20894 ABSTRACT To provide online access

More information

Topics in Computer Music Instrument Identification. Ioanna Karydi

Topics in Computer Music Instrument Identification. Ioanna Karydi Topics in Computer Music Instrument Identification Ioanna Karydi Presentation overview What is instrument identification? Sound attributes & Timbre Human performance The ideal algorithm Selected approaches

More information

Detection of Panoramic Takes in Soccer Videos Using Phase Correlation and Boosting

Detection of Panoramic Takes in Soccer Videos Using Phase Correlation and Boosting Detection of Panoramic Takes in Soccer Videos Using Phase Correlation and Boosting Luiz G. L. B. M. de Vasconcelos Research & Development Department Globo TV Network Email: luiz.vasconcelos@tvglobo.com.br

More information

GCSE FILM STUDIES PAPER 1 EXPLORING FILM SUPERHERO GENRE. 1 hour 30 minutes (20 minutes for DVD screening) 1.4 minutes per mark

GCSE FILM STUDIES PAPER 1 EXPLORING FILM SUPERHERO GENRE. 1 hour 30 minutes (20 minutes for DVD screening) 1.4 minutes per mark GCSE FILM STUDIES PAPER EXPLORING FILM SUPERHERO GENRE hour 0 minutes (0 minutes for DVD screening). minutes per mark 0 marks = minutes 0 marks = 8 minutes Question One AO Explore, respond to and evaluate

More information

Unit Detection in American Football TV Broadcasts Using Average Energy of Audio Track

Unit Detection in American Football TV Broadcasts Using Average Energy of Audio Track Unit Detection in American Football TV Broadcasts Using Average Energy of Audio Track Mei-Ling Shyu, Guy Ravitz Department of Electrical & Computer Engineering University of Miami Coral Gables, FL 33124,

More information

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval DAY 1 Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval Jay LeBoeuf Imagine Research jay{at}imagine-research.com Rebecca

More information

On the Characterization of Distributed Virtual Environment Systems

On the Characterization of Distributed Virtual Environment Systems On the Characterization of Distributed Virtual Environment Systems P. Morillo, J. M. Orduña, M. Fernández and J. Duato Departamento de Informática. Universidad de Valencia. SPAIN DISCA. Universidad Politécnica

More information

Retrieval of textual song lyrics from sung inputs

Retrieval of textual song lyrics from sung inputs INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Retrieval of textual song lyrics from sung inputs Anna M. Kruspe Fraunhofer IDMT, Ilmenau, Germany kpe@idmt.fraunhofer.de Abstract Retrieving the

More information

Embedding Multilevel Image Encryption in the LAR Codec

Embedding Multilevel Image Encryption in the LAR Codec Embedding Multilevel Image Encryption in the LAR Codec Jean Motsch, Olivier Déforges, Marie Babel To cite this version: Jean Motsch, Olivier Déforges, Marie Babel. Embedding Multilevel Image Encryption

More information

Composer Style Attribution

Composer Style Attribution Composer Style Attribution Jacqueline Speiser, Vishesh Gupta Introduction Josquin des Prez (1450 1521) is one of the most famous composers of the Renaissance. Despite his fame, there exists a significant

More information

2. AN INTROSPECTION OF THE MORPHING PROCESS

2. AN INTROSPECTION OF THE MORPHING PROCESS 1. INTRODUCTION Voice morphing means the transition of one speech signal into another. Like image morphing, speech morphing aims to preserve the shared characteristics of the starting and final signals,

More information

WATERMARKING USING DECIMAL SEQUENCES. Navneet Mandhani and Subhash Kak

WATERMARKING USING DECIMAL SEQUENCES. Navneet Mandhani and Subhash Kak Cryptologia, volume 29, January 2005 WATERMARKING USING DECIMAL SEQUENCES Navneet Mandhani and Subhash Kak ADDRESS: Department of Electrical and Computer Engineering, Louisiana State University, Baton

More information

Singer Traits Identification using Deep Neural Network

Singer Traits Identification using Deep Neural Network Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic

More information

Color Quantization of Compressed Video Sequences. Wan-Fung Cheung, and Yuk-Hee Chan, Member, IEEE 1 CSVT

Color Quantization of Compressed Video Sequences. Wan-Fung Cheung, and Yuk-Hee Chan, Member, IEEE 1 CSVT CSVT -02-05-09 1 Color Quantization of Compressed Video Sequences Wan-Fung Cheung, and Yuk-Hee Chan, Member, IEEE 1 Abstract This paper presents a novel color quantization algorithm for compressed video

More information

Powerful Software Tools and Methods to Accelerate Test Program Development A Test Systems Strategies, Inc. (TSSI) White Paper.

Powerful Software Tools and Methods to Accelerate Test Program Development A Test Systems Strategies, Inc. (TSSI) White Paper. Powerful Software Tools and Methods to Accelerate Test Program Development A Test Systems Strategies, Inc. (TSSI) White Paper Abstract Test costs have now risen to as much as 50 percent of the total manufacturing

More information

PERCEPTUAL QUALITY COMPARISON BETWEEN SINGLE-LAYER AND SCALABLE VIDEOS AT THE SAME SPATIAL, TEMPORAL AND AMPLITUDE RESOLUTIONS. Yuanyi Xue, Yao Wang

PERCEPTUAL QUALITY COMPARISON BETWEEN SINGLE-LAYER AND SCALABLE VIDEOS AT THE SAME SPATIAL, TEMPORAL AND AMPLITUDE RESOLUTIONS. Yuanyi Xue, Yao Wang PERCEPTUAL QUALITY COMPARISON BETWEEN SINGLE-LAYER AND SCALABLE VIDEOS AT THE SAME SPATIAL, TEMPORAL AND AMPLITUDE RESOLUTIONS Yuanyi Xue, Yao Wang Department of Electrical and Computer Engineering Polytechnic

More information

Ligeti. Continuum for Harpsichord (1968) F.P. Sharma and Glen Halls All Rights Reserved

Ligeti. Continuum for Harpsichord (1968) F.P. Sharma and Glen Halls All Rights Reserved Ligeti. Continuum for Harpsichord (1968) F.P. Sharma and Glen Halls All Rights Reserved Continuum is one of the most balanced and self contained works in the twentieth century repertory. All of the parameters

More information

A HIGHLY INTERACTIVE SYSTEM FOR PROCESSING LARGE VOLUMES OF ULTRASONIC TESTING DATA. H. L. Grothues, R. H. Peterson, D. R. Hamlin, K. s.

A HIGHLY INTERACTIVE SYSTEM FOR PROCESSING LARGE VOLUMES OF ULTRASONIC TESTING DATA. H. L. Grothues, R. H. Peterson, D. R. Hamlin, K. s. A HIGHLY INTERACTIVE SYSTEM FOR PROCESSING LARGE VOLUMES OF ULTRASONIC TESTING DATA H. L. Grothues, R. H. Peterson, D. R. Hamlin, K. s. Pickens Southwest Research Institute San Antonio, Texas INTRODUCTION

More information

Comparison of Dictionary-Based Approaches to Automatic Repeating Melody Extraction

Comparison of Dictionary-Based Approaches to Automatic Repeating Melody Extraction Comparison of Dictionary-Based Approaches to Automatic Repeating Melody Extraction Hsuan-Huei Shih, Shrikanth S. Narayanan and C.-C. Jay Kuo Integrated Media Systems Center and Department of Electrical

More information

BEGINNING VIDEO PRODUCTION. Total Classroom Laboratory/CC/CVE

BEGINNING VIDEO PRODUCTION. Total Classroom Laboratory/CC/CVE Career Education BEGINNING VIDEO PRODUCTION DATE: 2016-2017 INDUSTRY SECTOR: PATHWAY: CBEDS TITLE: Arts, Media and Entertainment Sector Design, Visual and Media Arts Introduction to Media Arts CBEDS CODE:

More information

FREE TV AUSTRALIA OPERATIONAL PRACTICE OP- 59 Measurement and Management of Loudness in Soundtracks for Television Broadcasting

FREE TV AUSTRALIA OPERATIONAL PRACTICE OP- 59 Measurement and Management of Loudness in Soundtracks for Television Broadcasting Page 1 of 10 1. SCOPE This Operational Practice is recommended by Free TV Australia and refers to the measurement of audio loudness as distinct from audio level. It sets out guidelines for measuring and

More information

DETECTION OF SLOW-MOTION REPLAY SEGMENTS IN SPORTS VIDEO FOR HIGHLIGHTS GENERATION

DETECTION OF SLOW-MOTION REPLAY SEGMENTS IN SPORTS VIDEO FOR HIGHLIGHTS GENERATION DETECTION OF SLOW-MOTION REPLAY SEGMENTS IN SPORTS VIDEO FOR HIGHLIGHTS GENERATION H. Pan P. van Beek M. I. Sezan Electrical & Computer Engineering University of Illinois Urbana, IL 6182 Sharp Laboratories

More information

Music Segmentation Using Markov Chain Methods

Music Segmentation Using Markov Chain Methods Music Segmentation Using Markov Chain Methods Paul Finkelstein March 8, 2011 Abstract This paper will present just how far the use of Markov Chains has spread in the 21 st century. We will explain some

More information

A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS

A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS Mutian Fu 1 Guangyu Xia 2 Roger Dannenberg 2 Larry Wasserman 2 1 School of Music, Carnegie Mellon University, USA 2 School of Computer

More information

A New "Duration-Adapted TR" Waveform Capture Method Eliminates Severe Limitations

A New Duration-Adapted TR Waveform Capture Method Eliminates Severe Limitations 31 st Conference of the European Working Group on Acoustic Emission (EWGAE) Th.3.B.4 More Info at Open Access Database www.ndt.net/?id=17567 A New "Duration-Adapted TR" Waveform Capture Method Eliminates

More information

Processes for the Intersection

Processes for the Intersection 7 Timing Processes for the Intersection In Chapter 6, you studied the operation of one intersection approach and determined the value of the vehicle extension time that would extend the green for as long

More information

A Parametric Autoregressive Model for the Extraction of Electric Network Frequency Fluctuations in Audio Forensic Authentication

A Parametric Autoregressive Model for the Extraction of Electric Network Frequency Fluctuations in Audio Forensic Authentication Journal of Energy and Power Engineering 10 (2016) 504-512 doi: 10.17265/1934-8975/2016.08.007 D DAVID PUBLISHING A Parametric Autoregressive Model for the Extraction of Electric Network Frequency Fluctuations

More information

Wise Mining Method through Ant Colony Optimization

Wise Mining Method through Ant Colony Optimization Proceedings of the 2009 IEEE International Conference on Systems, Man, and Cybernetics San Antonio, TX, USA - October 2009 Wise Mining Method through Ant Colony Optimization Yang Jianxiong and Junzo Watada

More information

Sensor-Based Analysis of User Generated Video for Multi-camera Video Remixing

Sensor-Based Analysis of User Generated Video for Multi-camera Video Remixing Sensor-Based Analysis of User Generated Video for Multi-camera Video Remixing Francesco Cricri 1, Igor D.D. Curcio 2, Sujeet Mate 2, Kostadin Dabov 1, and Moncef Gabbouj 1 1 Department of Signal Processing,

More information

Gaining Musical Insights: Visualizing Multiple. Listening Histories

Gaining Musical Insights: Visualizing Multiple. Listening Histories Gaining Musical Insights: Visualizing Multiple Ya-Xi Chen yaxi.chen@ifi.lmu.de Listening Histories Dominikus Baur dominikus.baur@ifi.lmu.de Andreas Butz andreas.butz@ifi.lmu.de ABSTRACT Listening histories

More information

Detecting the Moment of Snap in Real-World Football Videos

Detecting the Moment of Snap in Real-World Football Videos Detecting the Moment of Snap in Real-World Football Videos Behrooz Mahasseni and Sheng Chen and Alan Fern and Sinisa Todorovic School of Electrical Engineering and Computer Science Oregon State University

More information

Audio Structure Analysis

Audio Structure Analysis Lecture Music Processing Audio Structure Analysis Meinard Müller International Audio Laboratories Erlangen meinard.mueller@audiolabs-erlangen.de Music Structure Analysis Music segmentation pitch content

More information

Analysis of Video Transmission over Lossy Channels

Analysis of Video Transmission over Lossy Channels 1012 IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, VOL. 18, NO. 6, JUNE 2000 Analysis of Video Transmission over Lossy Channels Klaus Stuhlmüller, Niko Färber, Member, IEEE, Michael Link, and Bernd

More information

A Fast Alignment Scheme for Automatic OCR Evaluation of Books

A Fast Alignment Scheme for Automatic OCR Evaluation of Books A Fast Alignment Scheme for Automatic OCR Evaluation of Books Ismet Zeki Yalniz, R. Manmatha Multimedia Indexing and Retrieval Group Dept. of Computer Science, University of Massachusetts Amherst, MA,

More information

TRAFFIC SURVEILLANCE VIDEO MANAGEMENT SYSTEM

TRAFFIC SURVEILLANCE VIDEO MANAGEMENT SYSTEM TRAFFIC SURVEILLANCE VIDEO MANAGEMENT SYSTEM K.Ganesan*, Kavitha.C, Kriti Tandon, Lakshmipriya.R TIFAC-Centre of Relevance and Excellence in Automotive Infotronics*, School of Information Technology and

More information

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods Kazuyoshi Yoshii, Masataka Goto and Hiroshi G. Okuno Department of Intelligence Science and Technology National

More information

System Level Simulation of Scheduling Schemes for C-V2X Mode-3

System Level Simulation of Scheduling Schemes for C-V2X Mode-3 1 System Level Simulation of Scheduling Schemes for C-V2X Mode-3 Luis F. Abanto-Leon, Arie Koppelaar, Chetan B. Math, Sonia Heemstra de Groot arxiv:1807.04822v1 [eess.sp] 12 Jul 2018 Eindhoven University

More information

Machine Learning Term Project Write-up Creating Models of Performers of Chopin Mazurkas

Machine Learning Term Project Write-up Creating Models of Performers of Chopin Mazurkas Machine Learning Term Project Write-up Creating Models of Performers of Chopin Mazurkas Marcello Herreshoff In collaboration with Craig Sapp (craig@ccrma.stanford.edu) 1 Motivation We want to generative

More information

MusCat: A Music Browser Featuring Abstract Pictures and Zooming User Interface

MusCat: A Music Browser Featuring Abstract Pictures and Zooming User Interface MusCat: A Music Browser Featuring Abstract Pictures and Zooming User Interface 1st Author 1st author's affiliation 1st line of address 2nd line of address Telephone number, incl. country code 1st author's

More information

A Visualization of Relationships Among Papers Using Citation and Co-citation Information

A Visualization of Relationships Among Papers Using Citation and Co-citation Information A Visualization of Relationships Among Papers Using Citation and Co-citation Information Yu Nakano, Toshiyuki Shimizu, and Masatoshi Yoshikawa Graduate School of Informatics, Kyoto University, Kyoto 606-8501,

More information

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION ULAŞ BAĞCI AND ENGIN ERZIN arxiv:0907.3220v1 [cs.sd] 18 Jul 2009 ABSTRACT. Music genre classification is an essential tool for

More information

NETFLIX MOVIE RATING ANALYSIS

NETFLIX MOVIE RATING ANALYSIS NETFLIX MOVIE RATING ANALYSIS Danny Dean EXECUTIVE SUMMARY Perhaps only a few us have wondered whether or not the number words in a movie s title could be linked to its success. You may question the relevance

More information

Gyorgi Ligeti. Chamber Concerto, Movement III (1970) Glen Halls All Rights Reserved

Gyorgi Ligeti. Chamber Concerto, Movement III (1970) Glen Halls All Rights Reserved Gyorgi Ligeti. Chamber Concerto, Movement III (1970) Glen Halls All Rights Reserved Ligeti once said, " In working out a notational compositional structure the decisive factor is the extent to which it

More information

Color Image Compression Using Colorization Based On Coding Technique

Color Image Compression Using Colorization Based On Coding Technique Color Image Compression Using Colorization Based On Coding Technique D.P.Kawade 1, Prof. S.N.Rawat 2 1,2 Department of Electronics and Telecommunication, Bhivarabai Sawant Institute of Technology and Research

More information

Pattern Smoothing for Compressed Video Transmission

Pattern Smoothing for Compressed Video Transmission Pattern for Compressed Transmission Hugh M. Smith and Matt W. Mutka Department of Computer Science Michigan State University East Lansing, MI 48824-1027 {smithh,mutka}@cps.msu.edu Abstract: In this paper

More information

ELEN Electronique numérique

ELEN Electronique numérique ELEN0040 - Electronique numérique Patricia ROUSSEAUX Année académique 2014-2015 CHAPITRE 5 Sequential circuits design - Timing issues ELEN0040 5-228 1 Sequential circuits design 1.1 General procedure 1.2

More information

Measurement of overtone frequencies of a toy piano and perception of its pitch

Measurement of overtone frequencies of a toy piano and perception of its pitch Measurement of overtone frequencies of a toy piano and perception of its pitch PACS: 43.75.Mn ABSTRACT Akira Nishimura Department of Media and Cultural Studies, Tokyo University of Information Sciences,

More information

Condensed tips based on Brad Bird on How to Compose Shots and Storyboarding the Simpson s Way

Condensed tips based on Brad Bird on How to Compose Shots and Storyboarding the Simpson s Way Storyboard Week 3 Condensed tips based on Brad Bird on How to Compose Shots and Storyboarding the Simpson s Way 1. Adjust down on the action. Avoid empty space above heads Lower the horizon 2. Make the

More information

Analysis of Schubert's "Auf dem Flusse" Seth Horvitz

Analysis of Schubert's Auf dem Flusse Seth Horvitz Analysis of Schubert's "Auf dem Flusse" Seth Horvitz shorvitz@mills.edu Mills College Tonal Analysis - Music 256 David Bernstein November 25, 2008 This essay will attempt to provide a detailed analysis

More information

An Integrated Music Chromaticism Model

An Integrated Music Chromaticism Model An Integrated Music Chromaticism Model DIONYSIOS POLITIS and DIMITRIOS MARGOUNAKIS Dept. of Informatics, School of Sciences Aristotle University of Thessaloniki University Campus, Thessaloniki, GR-541

More information

UC San Diego UC San Diego Previously Published Works

UC San Diego UC San Diego Previously Published Works UC San Diego UC San Diego Previously Published Works Title Classification of MPEG-2 Transport Stream Packet Loss Visibility Permalink https://escholarship.org/uc/item/9wk791h Authors Shin, J Cosman, P

More information

Video summarization based on camera motion and a subjective evaluation method

Video summarization based on camera motion and a subjective evaluation method Video summarization based on camera motion and a subjective evaluation method Mickaël Guironnet, Denis Pellerin, Nathalie Guyader, Patricia Ladret To cite this version: Mickaël Guironnet, Denis Pellerin,

More information

Design of Fault Coverage Test Pattern Generator Using LFSR

Design of Fault Coverage Test Pattern Generator Using LFSR Design of Fault Coverage Test Pattern Generator Using LFSR B.Saritha M.Tech Student, Department of ECE, Dhruva Institue of Engineering & Technology. Abstract: A new fault coverage test pattern generator

More information

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video Mohamed Hassan, Taha Landolsi, Husameldin Mukhtar, and Tamer Shanableh College of Engineering American

More information

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene Beat Extraction from Expressive Musical Performances Simon Dixon, Werner Goebl and Emilios Cambouropoulos Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria.

More information

DIGITAL NARRATIVE ASSIGNMENT 2

DIGITAL NARRATIVE ASSIGNMENT 2 DIGITAL NARRATIVE ASSIGNMENT 2 Essay STEVEN HALL KC42245 YEAR 2 SEMESTER 2 DATE SUBMITTED: 20 MAY 2016 WORD COUNT: 1000 Table of Contents Planning: Changing from A to B Model... 2 Production: Filming and

More information

Contents. Welcome to LCAST. System Requirements. Compatibility. Installation and Authorization. Loudness Metering. True-Peak Metering

Contents. Welcome to LCAST. System Requirements. Compatibility. Installation and Authorization. Loudness Metering. True-Peak Metering LCAST User Manual Contents Welcome to LCAST System Requirements Compatibility Installation and Authorization Loudness Metering True-Peak Metering LCAST User Interface Your First Loudness Measurement Presets

More information

University of Bristol - Explore Bristol Research. Peer reviewed version. Link to published version (if available): /ISCAS.2005.

University of Bristol - Explore Bristol Research. Peer reviewed version. Link to published version (if available): /ISCAS.2005. Wang, D., Canagarajah, CN., & Bull, DR. (2005). S frame design for multiple description video coding. In IEEE International Symposium on Circuits and Systems (ISCAS) Kobe, Japan (Vol. 3, pp. 19 - ). Institute

More information

AUTOMATIC LICENSE PLATE RECOGNITION(ALPR) ON EMBEDDED SYSTEM

AUTOMATIC LICENSE PLATE RECOGNITION(ALPR) ON EMBEDDED SYSTEM AUTOMATIC LICENSE PLATE RECOGNITION(ALPR) ON EMBEDDED SYSTEM Presented by Guanghan APPLICATIONS 1. Automatic toll collection 2. Traffic law enforcement 3. Parking lot access control 4. Road traffic monitoring

More information

Correlated to: Massachusetts English Language Arts Curriculum Framework with May 2004 Supplement (Grades 5-8)

Correlated to: Massachusetts English Language Arts Curriculum Framework with May 2004 Supplement (Grades 5-8) General STANDARD 1: Discussion* Students will use agreed-upon rules for informal and formal discussions in small and large groups. Grades 7 8 1.4 : Know and apply rules for formal discussions (classroom,

More information

6.5 Percussion scalograms and musical rhythm

6.5 Percussion scalograms and musical rhythm 6.5 Percussion scalograms and musical rhythm 237 1600 566 (a) (b) 200 FIGURE 6.8 Time-frequency analysis of a passage from the song Buenos Aires. (a) Spectrogram. (b) Zooming in on three octaves of the

More information

ITU-T Y Specific requirements and capabilities of the Internet of things for big data

ITU-T Y Specific requirements and capabilities of the Internet of things for big data I n t e r n a t i o n a l T e l e c o m m u n i c a t i o n U n i o n ITU-T Y.4114 TELECOMMUNICATION STANDARDIZATION SECTOR OF ITU (07/2017) SERIES Y: GLOBAL INFORMATION INFRASTRUCTURE, INTERNET PROTOCOL

More information

2. Materials Development. 1) Desktop Video Production

2. Materials Development. 1) Desktop Video Production 2. Materials Development 1) Desktop Video Production Dr. Merza Abbas Acting Deputy Director Chairman of Graduate Studies Centre for Instructional Technology and Multimedia University of Science, Malaysia

More information