Universität Bamberg Angewandte Informatik. Seminar KI: gestern, heute, morgen. We are Humor Beings. Understanding and Predicting visual Humor

Size: px
Start display at page:

Download "Universität Bamberg Angewandte Informatik. Seminar KI: gestern, heute, morgen. We are Humor Beings. Understanding and Predicting visual Humor"

Transcription

1 Universität Bamberg Angewandte Informatik Seminar KI: gestern, heute, morgen We are Humor Beings. Understanding and Predicting visual Humor by Daniel Tremmel 18. Februar 2017 advised by Professor Dr. Ute Schmid

2 Abbildung 1: An Example Scene of the Dataset. Although there already is some research done in the field of humor in artificial intelligence, there are only a few works dealing with the issue of visual humor in artificial intelligence. The reason is that the problem of object recognition, which is important for a study on visual humor, isn t solved yet. The technical paper We are Humor Beings. Understanding and Predicting visual Humor is one of the few research papers on visual humor in the field of artificial intelligence. This paper will outline the most important points of the study, including the research approach, the features that have been used and the results of the study. 1 Introduction Until now, there has not been much research in the field of visual humor in artificial intelligence. The technical paper We are humor Beings. Understanding and Predicting Visual Humor is one of the few approaches to deal with the question wether it is possible to model visual humor in artificial intelligence. The problem so far was, that for an understanding of visual humor it is vital to recognize all objects within a scene, to distinguish them from each other and to learn in which ways they are interacting with each other. All of these challenges, however, haven t been solved by artificial intelligence so far. Nevertheless the topic could become relevant in lots of areas. One could think of smart cameras, picking up the right funny moment, recommendation tools for rating funny pictures higher than others or video summarization tools, able to recognize and extract funny scenes out of longer film sequences. These are just a few examples of how artificial intelligence could make an impact in the field of visual humor. Therefore, the researchers behind We are humor Beings. Understanding and Predicting visual Humor searched a way to bypass the object recognition problem of artificial intelligence by using clipart scenes. The objects in the clipart scenes were densely annotated, so that the computational model always knew which objects were in the scene and where in the scene they were located. This made it possible to conduct a study on visual humor, using scene level features as well as object level features. The technical paper not only 1

3 provided a way to model and evaluate visual humor through artificial intelligence, it also made a contribution to how visual humor can be defined. In order to conduct the study, the researchers used a support vector regressor, that was being trained with the scene level and the instance level features on two different abstract datasets, that were crafted by Amazon Mechanical Turk Workers. As a success criteria the researchers formulated two tasks, that served as guideline for detecting humor. In this paper I will discuss the technological background of the project in order to provide the knowledge that is needed to understand the constraints of the project. In the main part of the paper I will discuss the research approach in detail as well as the outcomes of the project. The last chapters will serve as a summarization and a conclusion of the project. 2 Background The researchers modelled two tasks, which they considered to be appropriate guidelines for measuring humor: Rating the funniness of a scene and altering the funniness of a scene. The tasks served as success criteria to see wether the experiment was successful or not. In order to carry out those tasks they had to apply several features on the technical level, instance level features as well as scene level features. Instance level features were needed mainly for altering the funniness, whereas the scene level features were needed for measuring the funniness of a scene. One important term with regard to scene level features is support vector regression (SVR). A Support Vector Regressor was trained on the scene level features and an ablation study was performed subsequently. SVR has its origin in the statistical learning theory, also called VC theory, that was developed to enable learning machines to generalize to unseen data. Initially, the SVR algorithm was a generalized portrait algorithm developed in the sixties in Russia by Vapnik, Chervonenkis and others. SVR is widely used not only in artificial intelligence, but also in high risk financial predictions and approximation of complex engineering analyses, for example. In artificial intelligence SVR is mainly used in order to train machine learning models to predict the most probable outcomes of regularly carried out tasks.(smola and Schölköpf, 2004) The idea is that recurring patterns are extracted out of bigger data sets and therefore a certain behaviour can be modelled based upon those patterns. It achieves this by minimizing the generalization error and thus reaching a generalized behaviour. A linear regression function is computed in a high dimensional feature space, in which the input data is mapped by a nonlinear function.(debasish Basak and Patranabis, 2004) The scene level features on which a support vector regressor was trained on are made up of cardinality, location and scene embeddings. Location stands for the location of an object within the scene. The clipart scenes are subdivided by numerous vertical and horizontal lines, which are forming lots of sectors. Those lines are internal features that aren t seen by the viewer of the scenes, but they serve as orientation for the computational modell, so it recognizes where in the scene a certain object occurs. Cardinality refers to the number of instances of every instance category within the scene, whereas scene embbeddings refers to the total number of all instances in a certain scene. The picture 2

4 below demonstrates very well the different features: Abbildung 2: Funny Scene of the Data Set. As we can see in the picture above, there are two oaks, three trees, two benches, and three people in the scene, two of them male and one female, all being of a middle age. Also, there is the sun in the background. Therefore, the cardinality of people is three, of trees three, of oaks two, of benches two and of suns one. As the scene embeddings are the sum of all objects in the scene, its value is eleven. The whole scene is inherently covered with a grid through which the computational modell gets an understandng of location. The modell can for example see that the two persons in the right of the picture are in a sector, which is further away with regard to the y-axis whereas for the person in the left of the picture, the modell can recognize that the person is in a sector which has a lower value on the y-axis. It can further recognize, that the sun has relatively high x-value, plus it is relatively small with respect to the people in the foreground, which is an indicator that the sun is in the background of the picture. Given the cardinality, the computational modell can conclude that, due to the presence of a certain object category, and due to the number of its appearances, a scene might be funny or not. The scene embeddings can help to further recognize wether a scene is funny or not, as a scene with a higher number of objects, is more likely to have interactions between objects. If a scene has more interactions, it is more likely being funny as there must be some form of interaction between objects, so a scene can be funny. If there is no interaction, there is also no possibility that a funny situation arises. So, location, cardinality, and scene embeddings make up the scene level features, based upon which the evaluation of the funniness of the scene is made. Apart from the scene level features the researchers also used instance level features, with which they altered the funniness of the scene. The instance level features are object embedding and local embedding. Object embedding describes the distributed representation of every object in the scene through which the context of each object category can be retrieved. Basically every object in the scene has some pointers to other objects that typically appear to be around the object under consideration. For that a neural network is trained with a so called bag-of-words modell. The bag-of-words modell is the 3

5 representation through which the neural network learns the context of a certain object. Is is a representation method for object categorization. It is often used in natural language processing and computer vision. The idea behind it is that the occurence of each word, or each object is counted and represented in a histogram. In relation to computer vision it is helpful to get an idea of the context in which an object normally occurs. Every object has its histogramm of objects that are normally around the object. This is how the object embedding feature works in We are humor Beings. Annother instance level feature, that was used in the project was the local embedding feature. The local embedding feature is a representation of the distance of each object in the scene to the object under consideration. The picture below helps to demonstrate the instance level features. In the scene we see an old lady sitting on the couch watching TV. The scene is Abbildung 3: Funny Scene of the Data Set. annotated with black arrows and blue lines. The black arrows are pointing to the couch, the pillow and the TV. These things are objects that normally occur to be around an old lady. So the bag-of-words modell of the old lady is containing a couch, a TV and a pillow. Training the neural network with those unfunny, common scenes, the modell learns the normal context in which the objects normally appear. The blue lines represent the local embedding features. The blue lines are measuring the distance between the objects in the scene to the lady. If the distance is lower between the object and the old lady, it is more likely that there is an interaction between the two than between an object that has a higher distance to the lady. The grid, that is covering the scene also shows, how location and distance is calculated. At first, the location of an object is determined by the sector in which it occurs to be. Then the distance between the object and the other objects is determined by calculating the number of sectors that are between an object and the object under consideration. However, the scene also shows potential weeknesses of the modell. As we can see the lady is sitting on the couch and is watching TV (the fact that the TV is not showing a picture can be neglected in this context). Thus, we can see that there is an interaction between the lady and the TV. According to the local 4

6 embedding features, however, it is rather unlikely, that there is an interaction, as the distance between the TV and the lady is very big. So in this situation, the local embedding feature could lead to a misperceiption, that could have an impact on the modell, as the neural network is trained with those kind of scenes. Although this misperceiption could be balanced out by the object embedding feature, as it represents the normal context of objects. So, if the neural network would be trained with several scenes, in which an old lady appears together with a couch, the modell would still learn, that there might be a potential connection between an old lady and a couch So, all the features described above were used in the course of the project. They helped to create a computational modell, which should be capable of recognizing and altering visual humor. In the next chapter the research approach will be described in detail, as well as how the scene level features and instance level features were used. 3 Research Approach In order to train a neural network, the researchers created two datasets, the Abstract Visual Humor Dataset (AVH) and the Funny Object Replaced (FOR) Dataset. The datasets were created by Amazon Mechanical Turk Workers (AMT). The scenes were developed with a clipart interface, consisting of 20 deformable human models, 31 animal models and about 100 indoor and outdoor objects. The human models contain different genders, races and ages and different face expressions. The abstract visual humor dataset consisted of about 3200 funny scenes and about 3200 unfunny scenes. Out of the funny scenes dataset the workers also created the FOR dataset consisting out of about scenes. For the FOR dataset, the workers were asked to change some objects of the funny scenes in order to make them unfunny. In the next paragraph, the AVH dataset will be explained. 3.1 Abstract Visual Humor Dataset For the AVH dataset the AMT were told to create realistic funny scenes, that could happen in a daily context. Through this condition they wanted to prevent the AMT from creating scenes that require insider s knowledge. After creating the scenes, the workers were told to write a short description of the scene and why they consider it to be funny. The intention here was, that the workers care more about the humor of the scene they created and are even more careful not to create scenes with insider s humor. Then, the workers were told to create annother unfunny dataset of every day scenes. In the next step the scenes had to be rated. As humor is a highly subjective phenomenon, the funniness of the scenes had to be rated by other workers. This helped to create objective measurements as the worker who created a scene could t rate it on his own. The guideline was that ten workers had to give a rating between one and five, where five was extremely funny and one was not funny at all. If the average rating of the scene was above the threshold, the scene was getting into the funny scenes dataset, if it was below, it was put into the unfunny scenes dataset. In the end there were 522 unintentionally 5

7 funny scenes and 682 unintentionally unfunny scenes. This shows, that the approach to let other workers do the rating of the scene actually has an impact on the datasets. Annother technique that should help to give an better understanding was annotating the scenes with different humor techniques. The researchers created a list of different humor techniques that are based upon existing humor theories. The humor theories were made partly by personal observation and partly by the known humor theories of Bujizen and Vandenburg (Buijzen and Valkenburg, 2004). The typologies used were for example person doing something unusual, animal doing something unusual, clownish behaviour, etc. The workers were instructed to label the scenes with those typologies.the scenes below show the top voted techniques that were applied during the process. An Abbildung 4: Top voted scenes by humor technique. From left to right: animal doing something unusual, person doing something unusual, somebody getting hurt, and somebody getting scared. interesting insight was that, all techniques that involved animate objects were voted higher in terms of funniness than pictures including inanimate objects. In about 75% of all scenes the workers picked either animal doing something unusual, or person doing something unusual. In this case we can speak of incongruity that was applied to the scene. Incongruity takes place, when objects appear in a context that is unusual for them. An example would be an old man playing football. Therefore incongruity can be used very well for altering the humor of a scene. For that, the FOR dataset was used. 3.2 Funny Object Replaced Dataset The FOR dataset was created for the study of humor on the object level. For the FOR dataset the researchers asked the workers to alter the funny scenes of the AVH dataset by replacing as few objects as possible, in order to make the scenes unfunny. The intention was, by changing as few things as possible, the researchers would gain a very fine-grained understanding of what objects cause a scene to be funny and why. The workers were also told not to alter the underlying structure of the scene, so they shouldn t change the relation between the objects to each other, or the context of the scene. The altering of the scene should take place exclusively on the object level. The FOR dataset consists of approximately scenes, all of which were created out of the AVH dataset. For each scene of the funny part of the AVH dataset five counterparts for the FOR dataset were created. 6

8 3.3 Predicting the Funniness Score In order to predict the funniness score of the scene, the researchers used a Support Vector Regressor, that was trained to regress to the Funniness Score F i from the ratings given by the workers. Based upon the ratings the scene level features are applied and an ablation study is conducted. In order to measure the success of the experiment, the researchers needed a guideline after which they could measure the success. Given the following formula the success was measured by relative error: 1 N n i=1 = P redictedf i GroundT ruthf i GroundT ruthf i (1) In the formula N stands for the number of test scenes and F i stands for the predicted funniness score for a given test scene i. The ground truth is the funniness score given by the workers for the given test scene. The experiment is measured against the baseline model, which always predicts the average funniness score of the training scenes. We will discuss the results of the experiment in the result section. 3.4 Altering the Funniness of a scene The model should be capable of altering the funniness of a scene in both directions. It should be able to make a funny scene unfunny and vice versa. The researchers considered this as an appropriate guideline for measuring the model s understanding of humor. In order to be able to alter the funniness of the scene two requirements came up: In the first step, the model should propose which objects in the scene should be replaced. In the second step a potential replacer object should be proposed Predicting objects to be replaced For each scene object the model has to make a binary prediction on wether the object should be replaced or not. For that task a multi-layer perceptron was trained to make the prediction on each object. The predictions happen as a naive prediction, measuring the overall accuracy, measured against the human predictions as well as as a class wise prediction. In order to see if the model succeeds in recognizing the objects to be replaced, the model needs to be succesful both in a naive measurement as well as in a class-wise measurement. The researchers had two baselines. Priors and anormaly detection. The priors were that an instance should not be replaced. There is also one baseline computed that an object should only be replaced if it is replaced in T% of the training data. Based on the validation set T was set to 20. The anormaly detection works in a way, that from the object under consideration, the scene embedding is subtracted from the object embedding. The objects that have the least similarity are considered the anomalous objects. All objects that have a cosine similarity that is less than the threshold T are considered anomalous objects and are therefore replaced. 7

9 3.4.2 Proposing a replacer object After the model decided which objects should be replaced, it should propose an appropriate replacer object. When it comes to altering the funniness of a scene, the model is trained with the ground truth annotations of objects that were replaced by humans in the corresponding scene. This is necessary, so that the performance of the model in proposing the object to be replaced and the performance of the model proposing a replacer object can be measured seperately from each other. For making the scene unfunny the researchers used a so called top-5 metric. This means if any of the best predictions is matching with the ground truth, it is considered correct. As baselines there are again priors and anomaly detection. The priors is that every object is replaced by one of the most frequent replacers of the training set. For the anomaly detection the object that is to be replaced is subtracted from the scene embedding.the five objects that are most similar with respect to the object embedding feature are the ones that are proposed as replacers. 3.5 Results The aim of the study was to find out wether it is possible to create a computational model that was able to recognize humor, and to alter humor. The concrete tasks that the model was approached with were to rate the funniness of a scene and to alter the funniness of a scene. In this section the results of the study will be presented Results: Rating the funniness of a scene In section 3.3 the metric was presented, with which the success of the model was measured. The formula presented there represents the success of the model measured by its relative error. The results of the measurement are presented in the table below: As we Abbildung 5: The results of the measurement of rating the funniness of a scene, presented by its relative error. can see, all scene level features perform better than the average baseline model. We can further see, that the location feature shows the best performance of all features. The location has a relative error of , which is considerably better than the baseline model against which the measurements were taken. All three features combined also have a relative error of , which is however due to redundancy. So, if all three features are combined it always would end up having the measurement of the best single feature. 8

10 What can clearly be said according to the results is, that the first task that the model was tested with could be achieved, by the model. According to the table, the model was able to recognize humor and rate the funniness of a scene Results: Altering the funniness of a scene When it came to altering the funniness of a scene, the challenge was, that the model should be able to alter the funniness in both directions. In the first sight, it appears easier for the model to make a funny scene unfunny, than vice versa. However the model shows good results in both tasks. Below we can see a picture that the model altered in a way to make it unfunny. In both scenes there are some objects that were changed in order to make it unfunny. In the upper scene the model exchanged an eagle stealing a steak with a butterfly and a ball. Also the old man was exchanged with s little boy. Additionally the little boy was put into the background. Through this alternation, the context of the scene was changed completely. In the first scene the funniness was made up through the fact that there are two people having a barbeque and suddenly an eagle appears and steals a steak from the grill. By exchanging the eagle and the steak with a butterfly and a ball and by exchanging the old man with a young boy, the context, that created the funniness of the scene was altered completely. In the scene below it is Abbildung 6: The scenes on the left were the original funny scenes. The scenes on the right were the altered unfunny scenes. basically the same, although it isn t really clear why the scene is actually funny. The results of making funny scenes unfunny confirm the impression that the scenes above give. The model achieved a very good overall result altering the funniness Score F i to 1.64, which is clearly below the funniness score of the input scenes, which was For altering the scenes from unfunny to funny, the researchers used the original FOR dataset with the unfunny counterparts of the original funny scenes. The funniness score of the scenes made funny by the model was 2.14, which is actually lower than the funniness score of the original funny scenes, but still a very decent score. An interesting fact is 9

11 also, that the model s funny scenes were considered to be more funny than the original funny scenes, created by the workers in about 28% of the scenes. 4 Conclusion The aim of the study was to create a model, that was able to create and recognize humor. So far there hasn t been much research in the area of artificial intelligence and visual humor. One reason was that the problem of object recognition, which is vital for a study on visual humor, isn t solved, yet. The researchers of We are Humor Beings bypassed the problem by using clipart scenes. In order to measure the success of their model the researchers modelled the tasks of evaluating the funniness of a scene and altering the funniness of a scene. The results for both tasks show, that the model was successful. Given the fact that this research was quite new and that the total amount of research done in the field is quite rare compared to other areas, the research can be considered a milestone on the road to creating a humorous computer model. The major achievement of the research was, that it managed to teach a computational model to recognize context and interactions within a scene, which is absolutely necessary for the understanding of visual humor. However, this research can only be one of the first steps that have to be taken for the understanding of visual humor and for creating a computational model for it. The crux on every research on humor and artificial intelligence will be that humor itself is so complex, that there isn t really a single definition of humor and what humor is. Keeping that in mind, further research should be done in this field. Although this research managed to teach a computational model some sort of understanding of context and interaction, further research in this direction is vital for a better understanding of visual humor at all and for visual humor and artificial intelligence. 10

12 Literatur M. Buijzen and P. M. Valkenburg Developing a typology of humor. Media Psychology 2, 4 (2004). Srimanta Pal Debasish Basak and Dipak Chandra Patranabis Support Vector Regression. Neural Information Processing. Letters and Reviews 11, 10 (2004), Alex J. Smola and Bernhard Schölköpf A tutorial on support vector regression. Statistics and Computing 14 (2004),

We Are Humor Beings: Understanding and Predicting Visual Humor

We Are Humor Beings: Understanding and Predicting Visual Humor We Are Humor Beings: Understanding and Predicting Visual Humor Arjun Chandrasekaran 1 Ashwin K. Vijayakumar 1 Stanislaw Antol 1 Mohit Bansal 2 Dhruv Batra 1 C. Lawrence Zitnick 3 Devi Parikh 1 1 Virginia

More information

Singer Traits Identification using Deep Neural Network

Singer Traits Identification using Deep Neural Network Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic

More information

Outline. Why do we classify? Audio Classification

Outline. Why do we classify? Audio Classification Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification Implementation Future Work Why do we classify

More information

Sarcasm Detection in Text: Design Document

Sarcasm Detection in Text: Design Document CSC 59866 Senior Design Project Specification Professor Jie Wei Wednesday, November 23, 2016 Sarcasm Detection in Text: Design Document Jesse Feinman, James Kasakyan, Jeff Stolzenberg 1 Table of contents

More information

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

Music Emotion Recognition. Jaesung Lee. Chung-Ang University Music Emotion Recognition Jaesung Lee Chung-Ang University Introduction Searching Music in Music Information Retrieval Some information about target music is available Query by Text: Title, Artist, or

More information

An Introduction to Deep Image Aesthetics

An Introduction to Deep Image Aesthetics Seminar in Laboratory of Visual Intelligence and Pattern Analysis (VIPA) An Introduction to Deep Image Aesthetics Yongcheng Jing College of Computer Science and Technology Zhejiang University Zhenchuan

More information

2. Problem formulation

2. Problem formulation Artificial Neural Networks in the Automatic License Plate Recognition. Ascencio López José Ignacio, Ramírez Martínez José María Facultad de Ciencias Universidad Autónoma de Baja California Km. 103 Carretera

More information

ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC

ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC Vaiva Imbrasaitė, Peter Robinson Computer Laboratory, University of Cambridge, UK Vaiva.Imbrasaite@cl.cam.ac.uk

More information

Deep Neural Networks Scanning for patterns (aka convolutional networks) Bhiksha Raj

Deep Neural Networks Scanning for patterns (aka convolutional networks) Bhiksha Raj Deep Neural Networks Scanning for patterns (aka convolutional networks) Bhiksha Raj 1 Story so far MLPs are universal function approximators Boolean functions, classifiers, and regressions MLPs can be

More information

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception LEARNING AUDIO SHEET MUSIC CORRESPONDENCES Matthias Dorfer Department of Computational Perception Short Introduction... I am a PhD Candidate in the Department of Computational Perception at Johannes Kepler

More information

Neural Network for Music Instrument Identi cation

Neural Network for Music Instrument Identi cation Neural Network for Music Instrument Identi cation Zhiwen Zhang(MSE), Hanze Tu(CCRMA), Yuan Li(CCRMA) SUN ID: zhiwen, hanze, yuanli92 Abstract - In the context of music, instrument identi cation would contribute

More information

Experiments on musical instrument separation using multiplecause

Experiments on musical instrument separation using multiplecause Experiments on musical instrument separation using multiplecause models J Klingseisen and M D Plumbley* Department of Electronic Engineering King's College London * - Corresponding Author - mark.plumbley@kcl.ac.uk

More information

Automatic Music Clustering using Audio Attributes

Automatic Music Clustering using Audio Attributes Automatic Music Clustering using Audio Attributes Abhishek Sen BTech (Electronics) Veermata Jijabai Technological Institute (VJTI), Mumbai, India abhishekpsen@gmail.com Abstract Music brings people together,

More information

Large scale Visual Sentiment Ontology and Detectors Using Adjective Noun Pairs

Large scale Visual Sentiment Ontology and Detectors Using Adjective Noun Pairs Large scale Visual Sentiment Ontology and Detectors Using Adjective Noun Pairs Damian Borth 1,2, Rongrong Ji 1, Tao Chen 1, Thomas Breuel 2, Shih-Fu Chang 1 1 Columbia University, New York, USA 2 University

More information

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset Ricardo Malheiro, Renato Panda, Paulo Gomes, Rui Paiva CISUC Centre for Informatics and Systems of the University of Coimbra {rsmal,

More information

Music Similarity and Cover Song Identification: The Case of Jazz

Music Similarity and Cover Song Identification: The Case of Jazz Music Similarity and Cover Song Identification: The Case of Jazz Simon Dixon and Peter Foster s.e.dixon@qmul.ac.uk Centre for Digital Music School of Electronic Engineering and Computer Science Queen Mary

More information

YOUR NAME ALL CAPITAL LETTERS

YOUR NAME ALL CAPITAL LETTERS THE TITLE OF THE THESIS IN 12-POINT CAPITAL LETTERS, CENTERED, SINGLE SPACED, 2-INCH FORM TOP MARGIN by YOUR NAME ALL CAPITAL LETTERS A THESIS Submitted to the Graduate Faculty of Pacific University Vision

More information

C. PCT 1434 December 10, Report on Characteristics of International Search Reports

C. PCT 1434 December 10, Report on Characteristics of International Search Reports C. PCT 1434 December 10, 2014 Madam, Sir, Report on Characteristics of International Search Reports./. 1. This Circular is addressed to your Office in its capacity as an International Searching Authority

More information

Music Composition with RNN

Music Composition with RNN Music Composition with RNN Jason Wang Department of Statistics Stanford University zwang01@stanford.edu Abstract Music composition is an interesting problem that tests the creativity capacities of artificial

More information

DV: Liking Cartoon Comedy

DV: Liking Cartoon Comedy 1 Stepwise Multiple Regression Model Rikki Price Com 631/731 March 24, 2016 I. MODEL Block 1 Block 2 DV: Liking Cartoon Comedy 2 Block Stepwise Block 1 = Demographics: Item: Age (G2) Item: Political Philosophy

More information

An Impact Analysis of Features in a Classification Approach to Irony Detection in Product Reviews

An Impact Analysis of Features in a Classification Approach to Irony Detection in Product Reviews Universität Bielefeld June 27, 2014 An Impact Analysis of Features in a Classification Approach to Irony Detection in Product Reviews Konstantin Buschmeier, Philipp Cimiano, Roman Klinger Semantic Computing

More information

Chord Classification of an Audio Signal using Artificial Neural Network

Chord Classification of an Audio Signal using Artificial Neural Network Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------

More information

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Aric Bartle (abartle@stanford.edu) December 14, 2012 1 Background The field of composer recognition has

More information

Algebra I Module 2 Lessons 1 19

Algebra I Module 2 Lessons 1 19 Eureka Math 2015 2016 Algebra I Module 2 Lessons 1 19 Eureka Math, Published by the non-profit Great Minds. Copyright 2015 Great Minds. No part of this work may be reproduced, distributed, modified, sold,

More information

Detecting Musical Key with Supervised Learning

Detecting Musical Key with Supervised Learning Detecting Musical Key with Supervised Learning Robert Mahieu Department of Electrical Engineering Stanford University rmahieu@stanford.edu Abstract This paper proposes and tests performance of two different

More information

Detecting the Moment of Snap in Real-World Football Videos

Detecting the Moment of Snap in Real-World Football Videos Detecting the Moment of Snap in Real-World Football Videos Behrooz Mahasseni and Sheng Chen and Alan Fern and Sinisa Todorovic School of Electrical Engineering and Computer Science Oregon State University

More information

Subjective Similarity of Music: Data Collection for Individuality Analysis

Subjective Similarity of Music: Data Collection for Individuality Analysis Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp

More information

Automatic Piano Music Transcription

Automatic Piano Music Transcription Automatic Piano Music Transcription Jianyu Fan Qiuhan Wang Xin Li Jianyu.Fan.Gr@dartmouth.edu Qiuhan.Wang.Gr@dartmouth.edu Xi.Li.Gr@dartmouth.edu 1. Introduction Writing down the score while listening

More information

Introduction to Natural Language Processing This week & next week: Classification Sentiment Lexicons

Introduction to Natural Language Processing This week & next week: Classification Sentiment Lexicons Introduction to Natural Language Processing This week & next week: Classification Sentiment Lexicons Center for Games and Playable Media http://games.soe.ucsc.edu Kendall review of HW 2 Next two weeks

More information

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? NICHOLAS BORG AND GEORGE HOKKANEN Abstract. The possibility of a hit song prediction algorithm is both academically interesting and industry motivated.

More information

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC Vishweshwara Rao, Sachin Pant, Madhumita Bhaskar and Preeti Rao Department of Electrical Engineering, IIT Bombay {vishu, sachinp,

More information

The Human Features of Music.

The Human Features of Music. The Human Features of Music. Bachelor Thesis Artificial Intelligence, Social Studies, Radboud University Nijmegen Chris Kemper, s4359410 Supervisor: Makiko Sadakata Artificial Intelligence, Social Studies,

More information

Enabling editors through machine learning

Enabling editors through machine learning Meta Follow Meta is an AI company that provides academics & innovation-driven companies with powerful views of t Dec 9, 2016 9 min read Enabling editors through machine learning Examining the data science

More information

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video Mohamed Hassan, Taha Landolsi, Husameldin Mukhtar, and Tamer Shanableh College of Engineering American

More information

Indexing local features. Wed March 30 Prof. Kristen Grauman UT-Austin

Indexing local features. Wed March 30 Prof. Kristen Grauman UT-Austin Indexing local features Wed March 30 Prof. Kristen Grauman UT-Austin Matching local features Kristen Grauman Matching local features? Image 1 Image 2 To generate candidate matches, find patches that have

More information

VBM683 Machine Learning

VBM683 Machine Learning VBM683 Machine Learning Pinar Duygulu Slides are adapted from Dhruv Batra, David Sontag, Aykut Erdem Quotes If you were a current computer science student what area would you start studying heavily? Answer:

More information

Retrieval of textual song lyrics from sung inputs

Retrieval of textual song lyrics from sung inputs INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Retrieval of textual song lyrics from sung inputs Anna M. Kruspe Fraunhofer IDMT, Ilmenau, Germany kpe@idmt.fraunhofer.de Abstract Retrieving the

More information

Neural Network Predicating Movie Box Office Performance

Neural Network Predicating Movie Box Office Performance Neural Network Predicating Movie Box Office Performance Alex Larson ECE 539 Fall 2013 Abstract The movie industry is a large part of modern day culture. With the rise of websites like Netflix, where people

More information

Motion Video Compression

Motion Video Compression 7 Motion Video Compression 7.1 Motion video Motion video contains massive amounts of redundant information. This is because each image has redundant information and also because there are very few changes

More information

Automatic Laughter Detection

Automatic Laughter Detection Automatic Laughter Detection Mary Knox Final Project (EECS 94) knoxm@eecs.berkeley.edu December 1, 006 1 Introduction Laughter is a powerful cue in communication. It communicates to listeners the emotional

More information

BBM 413 Fundamentals of Image Processing Dec. 11, Erkut Erdem Dept. of Computer Engineering Hacettepe University. Segmentation Part 1

BBM 413 Fundamentals of Image Processing Dec. 11, Erkut Erdem Dept. of Computer Engineering Hacettepe University. Segmentation Part 1 BBM 413 Fundamentals of Image Processing Dec. 11, 2012 Erkut Erdem Dept. of Computer Engineering Hacettepe University Segmentation Part 1 Image segmentation Goal: identify groups of pixels that go together

More information

CS 1674: Intro to Computer Vision. Intro to Recognition. Prof. Adriana Kovashka University of Pittsburgh October 24, 2016

CS 1674: Intro to Computer Vision. Intro to Recognition. Prof. Adriana Kovashka University of Pittsburgh October 24, 2016 CS 1674: Intro to Computer Vision Intro to Recognition Prof. Adriana Kovashka University of Pittsburgh October 24, 2016 Plan for today Examples of visual recognition problems What should we recognize?

More information

Table of Contents. 2 Select camera-lens configuration Select camera and lens type Listbox: Select source image... 8

Table of Contents. 2 Select camera-lens configuration Select camera and lens type Listbox: Select source image... 8 Table of Contents 1 Starting the program 3 1.1 Installation of the program.......................... 3 1.2 Starting the program.............................. 3 1.3 Control button: Load source image......................

More information

Case Study: Can Video Quality Testing be Scripted?

Case Study: Can Video Quality Testing be Scripted? 1566 La Pradera Dr Campbell, CA 95008 www.videoclarity.com 408-379-6952 Case Study: Can Video Quality Testing be Scripted? Bill Reckwerdt, CTO Video Clarity, Inc. Version 1.0 A Video Clarity Case Study

More information

PICK THE RIGHT TEAM AND MAKE A BLOCKBUSTER A SOCIAL ANALYSIS THROUGH MOVIE HISTORY

PICK THE RIGHT TEAM AND MAKE A BLOCKBUSTER A SOCIAL ANALYSIS THROUGH MOVIE HISTORY PICK THE RIGHT TEAM AND MAKE A BLOCKBUSTER A SOCIAL ANALYSIS THROUGH MOVIE HISTORY THE CHALLENGE: TO UNDERSTAND HOW TEAMS CAN WORK BETTER SOCIAL NETWORK + MACHINE LEARNING TO THE RESCUE Previous research:

More information

Improving Frame Based Automatic Laughter Detection

Improving Frame Based Automatic Laughter Detection Improving Frame Based Automatic Laughter Detection Mary Knox EE225D Class Project knoxm@eecs.berkeley.edu December 13, 2007 Abstract Laughter recognition is an underexplored area of research. My goal for

More information

1/8. Axioms of Intuition

1/8. Axioms of Intuition 1/8 Axioms of Intuition Kant now turns to working out in detail the schematization of the categories, demonstrating how this supplies us with the principles that govern experience. Prior to doing so he

More information

APPLICATIONS OF DIGITAL IMAGE ENHANCEMENT TECHNIQUES FOR IMPROVED

APPLICATIONS OF DIGITAL IMAGE ENHANCEMENT TECHNIQUES FOR IMPROVED APPLICATIONS OF DIGITAL IMAGE ENHANCEMENT TECHNIQUES FOR IMPROVED ULTRASONIC IMAGING OF DEFECTS IN COMPOSITE MATERIALS Brian G. Frock and Richard W. Martin University of Dayton Research Institute Dayton,

More information

Automated Accompaniment

Automated Accompaniment Automated Tyler Seacrest University of Nebraska, Lincoln April 20, 2007 Artificial Intelligence Professor Surkan The problem as originally stated: The problem as originally stated: ˆ Proposed Input The

More information

Implementation and performance analysis of convolution error correcting codes with code rate=1/2.

Implementation and performance analysis of convolution error correcting codes with code rate=1/2. 2016 International Conference on Micro-Electronics and Telecommunication Engineering Implementation and performance analysis of convolution error correcting codes with code rate=1/2. Neha Faculty of engineering

More information

BIBLIOMETRIC REPORT. Bibliometric analysis of Mälardalen University. Final Report - updated. April 28 th, 2014

BIBLIOMETRIC REPORT. Bibliometric analysis of Mälardalen University. Final Report - updated. April 28 th, 2014 BIBLIOMETRIC REPORT Bibliometric analysis of Mälardalen University Final Report - updated April 28 th, 2014 Bibliometric analysis of Mälardalen University Report for Mälardalen University Per Nyström PhD,

More information

Wipe Scene Change Detection in Video Sequences

Wipe Scene Change Detection in Video Sequences Wipe Scene Change Detection in Video Sequences W.A.C. Fernando, C.N. Canagarajah, D. R. Bull Image Communications Group, Centre for Communications Research, University of Bristol, Merchant Ventures Building,

More information

Relationships Between Quantitative Variables

Relationships Between Quantitative Variables Chapter 5 Relationships Between Quantitative Variables Three Tools we will use Scatterplot, a two-dimensional graph of data values Correlation, a statistic that measures the strength and direction of a

More information

gresearch Focus Cognitive Sciences

gresearch Focus Cognitive Sciences Learning about Music Cognition by Asking MIR Questions Sebastian Stober August 12, 2016 CogMIR, New York City sstober@uni-potsdam.de http://www.uni-potsdam.de/mlcog/ MLC g Machine Learning in Cognitive

More information

Analytic Comparison of Audio Feature Sets using Self-Organising Maps

Analytic Comparison of Audio Feature Sets using Self-Organising Maps Analytic Comparison of Audio Feature Sets using Self-Organising Maps Rudolf Mayer, Jakob Frank, Andreas Rauber Institute of Software Technology and Interactive Systems Vienna University of Technology,

More information

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur Module 8 VIDEO CODING STANDARDS Lesson 27 H.264 standard Lesson Objectives At the end of this lesson, the students should be able to: 1. State the broad objectives of the H.264 standard. 2. List the improved

More information

Perceptual dimensions of short audio clips and corresponding timbre features

Perceptual dimensions of short audio clips and corresponding timbre features Perceptual dimensions of short audio clips and corresponding timbre features Jason Musil, Budr El-Nusairi, Daniel Müllensiefen Department of Psychology, Goldsmiths, University of London Question How do

More information

Large Scale Concepts and Classifiers for Describing Visual Sentiment in Social Multimedia

Large Scale Concepts and Classifiers for Describing Visual Sentiment in Social Multimedia Large Scale Concepts and Classifiers for Describing Visual Sentiment in Social Multimedia Shih Fu Chang Columbia University http://www.ee.columbia.edu/dvmm June 2013 Damian Borth Tao Chen Rongrong Ji Yan

More information

Noise (Music) Composition Using Classification Algorithms Peter Wang (pwang01) December 15, 2017

Noise (Music) Composition Using Classification Algorithms Peter Wang (pwang01) December 15, 2017 Noise (Music) Composition Using Classification Algorithms Peter Wang (pwang01) December 15, 2017 Background Abstract I attempted a solution at using machine learning to compose music given a large corpus

More information

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular Music Mood Sheng Xu, Albert Peyton, Ryan Bhular What is Music Mood A psychological & musical topic Human emotions conveyed in music can be comprehended from two aspects: Lyrics Music Factors that affect

More information

Brain-Computer Interface (BCI)

Brain-Computer Interface (BCI) Brain-Computer Interface (BCI) Christoph Guger, Günter Edlinger, g.tec Guger Technologies OEG Herbersteinstr. 60, 8020 Graz, Austria, guger@gtec.at This tutorial shows HOW-TO find and extract proper signal

More information

Some Experiments in Humour Recognition Using the Italian Wikiquote Collection

Some Experiments in Humour Recognition Using the Italian Wikiquote Collection Some Experiments in Humour Recognition Using the Italian Wikiquote Collection Davide Buscaldi and Paolo Rosso Dpto. de Sistemas Informáticos y Computación (DSIC), Universidad Politécnica de Valencia, Spain

More information

DETECTION OF SLOW-MOTION REPLAY SEGMENTS IN SPORTS VIDEO FOR HIGHLIGHTS GENERATION

DETECTION OF SLOW-MOTION REPLAY SEGMENTS IN SPORTS VIDEO FOR HIGHLIGHTS GENERATION DETECTION OF SLOW-MOTION REPLAY SEGMENTS IN SPORTS VIDEO FOR HIGHLIGHTS GENERATION H. Pan P. van Beek M. I. Sezan Electrical & Computer Engineering University of Illinois Urbana, IL 6182 Sharp Laboratories

More information

Supplemental Material: Color Compatibility From Large Datasets

Supplemental Material: Color Compatibility From Large Datasets Supplemental Material: Color Compatibility From Large Datasets Peter O Donovan, Aseem Agarwala, and Aaron Hertzmann Project URL: www.dgp.toronto.edu/ donovan/color/ 1 Unmixing color preferences In the

More information

SINGING is a popular social activity and a good way of expressing

SINGING is a popular social activity and a good way of expressing 396 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 17, NO. 3, MARCH 2015 Competence-Based Song Recommendation: Matching Songs to One s Singing Skill Kuang Mao, Lidan Shou, Ju Fan, Gang Chen, and Mohan S. Kankanhalli,

More information

Here s a question for you: What happens if we try to go the other way? For instance:

Here s a question for you: What happens if we try to go the other way? For instance: Prime Numbers It s pretty simple to multiply two numbers and get another number. Here s a question for you: What happens if we try to go the other way? For instance: With a little thinking remembering

More information

Speech To Song Classification

Speech To Song Classification Speech To Song Classification Emily Graber Center for Computer Research in Music and Acoustics, Department of Music, Stanford University Abstract The speech to song illusion is a perceptual phenomenon

More information

A repetition-based framework for lyric alignment in popular songs

A repetition-based framework for lyric alignment in popular songs A repetition-based framework for lyric alignment in popular songs ABSTRACT LUONG Minh Thang and KAN Min Yen Department of Computer Science, School of Computing, National University of Singapore We examine

More information

Speech Recognition and Signal Processing for Broadcast News Transcription

Speech Recognition and Signal Processing for Broadcast News Transcription 2.2.1 Speech Recognition and Signal Processing for Broadcast News Transcription Continued research and development of a broadcast news speech transcription system has been promoted. Universities and researchers

More information

Identifying Table Tennis Balls From Real Match Scenes Using Image Processing And Artificial Intelligence Techniques

Identifying Table Tennis Balls From Real Match Scenes Using Image Processing And Artificial Intelligence Techniques Identifying Table Tennis Balls From Real Match Scenes Using Image Processing And Artificial Intelligence Techniques K. C. P. Wong Department of Communication and Systems Open University Milton Keynes,

More information

Agilent DSO5014A Oscilloscope Tutorial

Agilent DSO5014A Oscilloscope Tutorial Contents UNIVERSITY OF CALIFORNIA AT BERKELEY College of Engineering Department of Electrical Engineering and Computer Sciences EE105 Lab Experiments Agilent DSO5014A Oscilloscope Tutorial 1 Introduction

More information

Laughbot: Detecting Humor in Spoken Language with Language and Audio Cues

Laughbot: Detecting Humor in Spoken Language with Language and Audio Cues Laughbot: Detecting Humor in Spoken Language with Language and Audio Cues Kate Park, Annie Hu, Natalie Muenster Email: katepark@stanford.edu, anniehu@stanford.edu, ncm000@stanford.edu Abstract We propose

More information

Jazz Melody Generation and Recognition

Jazz Melody Generation and Recognition Jazz Melody Generation and Recognition Joseph Victor December 14, 2012 Introduction In this project, we attempt to use machine learning methods to study jazz solos. The reason we study jazz in particular

More information

Relationships. Between Quantitative Variables. Chapter 5. Copyright 2006 Brooks/Cole, a division of Thomson Learning, Inc.

Relationships. Between Quantitative Variables. Chapter 5. Copyright 2006 Brooks/Cole, a division of Thomson Learning, Inc. Relationships Chapter 5 Between Quantitative Variables Copyright 2006 Brooks/Cole, a division of Thomson Learning, Inc. Three Tools we will use Scatterplot, a two-dimensional graph of data values Correlation,

More information

Chapter 5. Describing Distributions Numerically. Finding the Center: The Median. Spread: Home on the Range. Finding the Center: The Median (cont.

Chapter 5. Describing Distributions Numerically. Finding the Center: The Median. Spread: Home on the Range. Finding the Center: The Median (cont. Chapter 5 Describing Distributions Numerically Copyright 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Copyright 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide

More information

IDENTIFYING TABLE TENNIS BALLS FROM REAL MATCH SCENES USING IMAGE PROCESSING AND ARTIFICIAL INTELLIGENCE TECHNIQUES

IDENTIFYING TABLE TENNIS BALLS FROM REAL MATCH SCENES USING IMAGE PROCESSING AND ARTIFICIAL INTELLIGENCE TECHNIQUES IDENTIFYING TABLE TENNIS BALLS FROM REAL MATCH SCENES USING IMAGE PROCESSING AND ARTIFICIAL INTELLIGENCE TECHNIQUES Dr. K. C. P. WONG Department of Communication and Systems Open University, Walton Hall

More information

Computational Modelling of Harmony

Computational Modelling of Harmony Computational Modelling of Harmony Simon Dixon Centre for Digital Music, Queen Mary University of London, Mile End Rd, London E1 4NS, UK simon.dixon@elec.qmul.ac.uk http://www.elec.qmul.ac.uk/people/simond

More information

18-551, Spring Group #4 Final Report. Get in the Game. Nick Lahr (nlahr) Bryan Murawski (bmurawsk) Chris Schnieder (cschneid)

18-551, Spring Group #4 Final Report. Get in the Game. Nick Lahr (nlahr) Bryan Murawski (bmurawsk) Chris Schnieder (cschneid) 18-551, Spring 2005 Group #4 Final Report Get in the Game Nick Lahr (nlahr) Bryan Murawski (bmurawsk) Chris Schnieder (cschneid) Group #4, Get in the Game Page 1 18-551, Spring 2005 Table of Contents 1.

More information

Audio-Based Video Editing with Two-Channel Microphone

Audio-Based Video Editing with Two-Channel Microphone Audio-Based Video Editing with Two-Channel Microphone Tetsuya Takiguchi Organization of Advanced Science and Technology Kobe University, Japan takigu@kobe-u.ac.jp Yasuo Ariki Organization of Advanced Science

More information

Mapping Interdisciplinarity at the Interfaces between the Science Citation Index and the Social Science Citation Index

Mapping Interdisciplinarity at the Interfaces between the Science Citation Index and the Social Science Citation Index Mapping Interdisciplinarity at the Interfaces between the Science Citation Index and the Social Science Citation Index Loet Leydesdorff University of Amsterdam, Amsterdam School of Communications Research

More information

arxiv: v1 [cs.lg] 15 Jun 2016

arxiv: v1 [cs.lg] 15 Jun 2016 Deep Learning for Music arxiv:1606.04930v1 [cs.lg] 15 Jun 2016 Allen Huang Department of Management Science and Engineering Stanford University allenh@cs.stanford.edu Abstract Raymond Wu Department of

More information

Cover Page. The handle holds various files of this Leiden University dissertation.

Cover Page. The handle   holds various files of this Leiden University dissertation. Cover Page The handle http://hdl.handle.net/1887/62348 holds various files of this Leiden University dissertation. Author: Crucq, A.K.C. Title: Abstract patterns and representation: the re-cognition of

More information

AudioRadar. A metaphorical visualization for the navigation of large music collections

AudioRadar. A metaphorical visualization for the navigation of large music collections AudioRadar A metaphorical visualization for the navigation of large music collections Otmar Hilliges, Phillip Holzer, René Klüber, Andreas Butz Ludwig-Maximilians-Universität München AudioRadar An Introduction

More information

Research & Development. White Paper WHP 232. A Large Scale Experiment for Mood-based Classification of TV Programmes BRITISH BROADCASTING CORPORATION

Research & Development. White Paper WHP 232. A Large Scale Experiment for Mood-based Classification of TV Programmes BRITISH BROADCASTING CORPORATION Research & Development White Paper WHP 232 September 2012 A Large Scale Experiment for Mood-based Classification of TV Programmes Jana Eggink, Denise Bland BRITISH BROADCASTING CORPORATION White Paper

More information

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

Finding Sarcasm in Reddit Postings: A Deep Learning Approach

Finding Sarcasm in Reddit Postings: A Deep Learning Approach Finding Sarcasm in Reddit Postings: A Deep Learning Approach Nick Guo, Ruchir Shah {nickguo, ruchirfs}@stanford.edu Abstract We use the recently published Self-Annotated Reddit Corpus (SARC) with a recurrent

More information

LEARNING SPECTRAL FILTERS FOR SINGLE- AND MULTI-LABEL CLASSIFICATION OF MUSICAL INSTRUMENTS. Patrick Joseph Donnelly

LEARNING SPECTRAL FILTERS FOR SINGLE- AND MULTI-LABEL CLASSIFICATION OF MUSICAL INSTRUMENTS. Patrick Joseph Donnelly LEARNING SPECTRAL FILTERS FOR SINGLE- AND MULTI-LABEL CLASSIFICATION OF MUSICAL INSTRUMENTS by Patrick Joseph Donnelly A dissertation submitted in partial fulfillment of the requirements for the degree

More information

Piotr KLECZKOWSKI, Magdalena PLEWA, Grzegorz PYDA

Piotr KLECZKOWSKI, Magdalena PLEWA, Grzegorz PYDA ARCHIVES OF ACOUSTICS 33, 4 (Supplement), 147 152 (2008) LOCALIZATION OF A SOUND SOURCE IN DOUBLE MS RECORDINGS Piotr KLECZKOWSKI, Magdalena PLEWA, Grzegorz PYDA AGH University od Science and Technology

More information

Laughbot: Detecting Humor in Spoken Language with Language and Audio Cues

Laughbot: Detecting Humor in Spoken Language with Language and Audio Cues Laughbot: Detecting Humor in Spoken Language with Language and Audio Cues Kate Park katepark@stanford.edu Annie Hu anniehu@stanford.edu Natalie Muenster ncm000@stanford.edu Abstract We propose detecting

More information

Make Me Laugh: Recommending Humoristic Content on the WWW

Make Me Laugh: Recommending Humoristic Content on the WWW S. Diefenbach, N. Henze & M. Pielot (Hrsg.): Mensch und Computer 2015 Tagungsband, Stuttgart: Oldenbourg Wissenschaftsverlag, 2015, S. 193-201. Make Me Laugh: Recommending Humoristic Content on the WWW

More information

Getting Started. Connect green audio output of SpikerBox/SpikerShield using green cable to your headphones input on iphone/ipad.

Getting Started. Connect green audio output of SpikerBox/SpikerShield using green cable to your headphones input on iphone/ipad. Getting Started First thing you should do is to connect your iphone or ipad to SpikerBox with a green smartphone cable. Green cable comes with designators on each end of the cable ( Smartphone and SpikerBox

More information

A Framework for Segmentation of Interview Videos

A Framework for Segmentation of Interview Videos A Framework for Segmentation of Interview Videos Omar Javed, Sohaib Khan, Zeeshan Rasheed, Mubarak Shah Computer Vision Lab School of Electrical Engineering and Computer Science University of Central Florida

More information

Evaluation of Serial Periodic, Multi-Variable Data Visualizations

Evaluation of Serial Periodic, Multi-Variable Data Visualizations Evaluation of Serial Periodic, Multi-Variable Data Visualizations Alexander Mosolov 13705 Valley Oak Circle Rockville, MD 20850 (301) 340-0613 AVMosolov@aol.com Benjamin B. Bederson i Computer Science

More information

ECE532 Digital System Design Title: Stereoscopic Depth Detection Using Two Cameras. Final Design Report

ECE532 Digital System Design Title: Stereoscopic Depth Detection Using Two Cameras. Final Design Report ECE532 Digital System Design Title: Stereoscopic Depth Detection Using Two Cameras Group #4 Prof: Chow, Paul Student 1: Robert An Student 2: Kai Chun Chou Student 3: Mark Sikora April 10 th, 2015 Final

More information

IMAGE AESTHETIC PREDICTORS BASED ON WEIGHTED CNNS. Oce Print Logic Technologies, Creteil, France

IMAGE AESTHETIC PREDICTORS BASED ON WEIGHTED CNNS. Oce Print Logic Technologies, Creteil, France IMAGE AESTHETIC PREDICTORS BASED ON WEIGHTED CNNS Bin Jin, Maria V. Ortiz Segovia2 and Sabine Su sstrunk EPFL, Lausanne, Switzerland; 2 Oce Print Logic Technologies, Creteil, France ABSTRACT Convolutional

More information

h t t p : / / w w w. v i d e o e s s e n t i a l s. c o m E - M a i l : j o e k a n a t t. n e t DVE D-Theater Q & A

h t t p : / / w w w. v i d e o e s s e n t i a l s. c o m E - M a i l : j o e k a n a t t. n e t DVE D-Theater Q & A J O E K A N E P R O D U C T I O N S W e b : h t t p : / / w w w. v i d e o e s s e n t i a l s. c o m E - M a i l : j o e k a n e @ a t t. n e t DVE D-Theater Q & A 15 June 2003 Will the D-Theater tapes

More information

CHAPTER 8 CONCLUSION AND FUTURE SCOPE

CHAPTER 8 CONCLUSION AND FUTURE SCOPE 124 CHAPTER 8 CONCLUSION AND FUTURE SCOPE Data hiding is becoming one of the most rapidly advancing techniques the field of research especially with increase in technological advancements in internet and

More information

Interactive Tic Tac Toe

Interactive Tic Tac Toe Interactive Tic Tac Toe Stefan Bennie Botha Thesis presented in fulfilment of the requirements for the degree of Honours of Computer Science at the University of the Western Cape Supervisor: Mehrdad Ghaziasgar

More information

Music Genre Classification

Music Genre Classification Music Genre Classification chunya25 Fall 2017 1 Introduction A genre is defined as a category of artistic composition, characterized by similarities in form, style, or subject matter. [1] Some researchers

More information