Shasha Zhang Art College, JinggangshanUniversity, Ji'an343009,Jiangxi, China

Size: px
Start display at page:

Download "Shasha Zhang Art College, JinggangshanUniversity, Ji'an343009,Jiangxi, China"

Transcription

1 doi: / Intelligent Recognition Model for Music Emotion Shasha Zhang Art College, JinggangshanUniversity, Ji'an343009,Jiangxi, China Abstract This paper utilizes intelligent means to make relatively detailed analysis on the relevant technologies regarding the music emotion recognition, through research and the improvement to the recognition of the main melody of music, to carry out in-depth analysis on MIDI file format, extract the characteristic parameters of the notes of various melodies and audio tracks of the MIDI music files (time value, pitch, sound intensity, etc.), based on the statistics of the pitch interval (the absolute sound intensity difference) of each audio track and note series, make use of an improved BP neural network algorithm, to successfully construct the music feature space model. On this basis, this paper establishes an automatic recognition model of BP neural network algorithm, and finally, verifies the effectiveness of the BP network design in the recognition of music emotion through instantiation. Keywords: Music Emotion, Main Melody Recognition, Intelligent Recognition, Bp Neural Network. 1. INTRODUCTION With full research and adequate application of computer technology in the field of multimedia, the multimedia business has also achieved rapid development and become one of the fastest growing industries with largest scale in the 21st century. The content based multimedia information processing technology is an important research hot topic in this field, especially after the introduction of animation, audio, video and other dynamic media, the multimedia technology has greatly enriched the manifestation of human emotions(agustus, Mahoney and Downey, et al., 2015; Wu, Zhong and Horner, et al., 2014). Computer music is an important component of the multimedia technology, though a lot of practical results have been achieved so far, the music signal digital encoding, digital compression and digital storage technology has developed rapidly, promoting the popularization and application of VCD, digital broadcasting and multimedia, etc., and demonstrating broad market prospect. However, the higher goal of computer music is to utilize the computer to simulate human emotion recognition and the creative intelligence for music, which involves the music theory, psychology, artificial intelligence, information processing, pattern recognition and other disciplines, and thus is of great difficulty(kimand André, 2008; Yangand Lee, 2004). The content of emotion computing research includes the real-time acquisition and modeling of the dynamic emotional information in the three-dimensional space, based on the emotion recognition and understanding of multimodal and dynamic timing characteristics, the information integration theory and methodology, the emotion automatic generation theory and the multimodal emotional expression, as well as the establishment of large scale dynamic emotion database based on physiological and behavioral characteristics, etc.(fritz, Jentschke and Gosselin, et al., 2009; Gosselin, Peretz, and Johnsen, et al.,2014). So far, there is almost no automatic generation model for artificial emotion both in line with human emotion laws and adaptable to machine implementation. Therefore, the research in the field of music emotion is of great theoretical and practical significance(juslin and Laukka, 2003). Overall speaking, the general analysis process of music emotion recognition includes three main aspects: Firstly, to analyze the music feature space, and make the corresponding feature modeling(balkwill, Thompson and Matsunaga, 2004); secondly, to explore the emotion space, and make the appropriate emotion modeling(peretz, Gagnon and Bouchard, 1998); and thirdly, based on the sample space of music emotion, make the recognition modeling for music emotion. On the basis of relatively accurate music recognition, the applications of music emotion can be proposed in the aspect of artificial intelligence, such as the musical fountain, music and stage lighting etc. designed based on the emotion that is expressed by music, and can also be applied in the music retrieval, intelligent creation and other fields(han, Rho and Jun, et al., 2010). 2. MUSIC EMOTION REPRESENTATION MODEL 2.1Music Emotion Analysis Model General speaking, the analysis model for music emotion recognition(gosselin, Peretz and Noulhiane, et al., 2015)is shown in Figure 1.

2 Figure 1. Music EmotionAnalysis Model Diagram It mainly includes three aspects of content, the first is the feature model in the music feature space, the second is the emotion model in the emotion space, and the third is the music emotion recognition model based on the sample space. Among them, the establishment of emotional model is the first step to carry out the music emotion analysis, which mainly completes the construction for the music emotion space. It is the fundamental premise for the next two steps, to lay the foundation for the establishment of the recognition model. Figure 2. Architecture of the Fuzzy Emotion Model Music Emotion Linguistic Variables The recognition model of music emotion can be divided into two levels: One level is the mental model of music emotion, which belongs to the cognitive class of model; and the other is the computational model of music emotion, which belongs to the analytic class of model. The mental models mainly explore the human emotional characteristics from the psychological point of view. Currently the classic models are mainly Hevner model and Thayer model.and the computational model mainly performs more in-depth analysis from the perspective of the computer analysis on music emotion. The current music computational models are all computational models based on linguistic value, including: Linguistic value calculation model based on the fuzzy set theory, and linguistic value calculation model based on the semantic similarity relationship linguistic value (Koelsch, Fritz and Müller, et al., 2009). 2.2 Computational Model formusic Emotion Computational Model Based on Fuzzy Theory Based on fuzzy set theory, this research takes the linguistic value for the academic thought value of human natural language, by certain grammar rules (such as mood operator) the composite linguistic value can be

3 obtained, while the semantics of the linguistic value can be characterized by the fuzzy membership functions. The basic features of this model are as the following: 1. The establishment of emotion model based on Hevner emotion ring; 2. Each emotion adopts the degree of membership to indicate its strength; 3.At certain moment, the intensity of certain emotion may be greater than that of all the other emotions, thus this emotion represents the major feature at this moment of the entire emotional state, which is called the dominant emotion; 4. The construction of a flowing emotion chain on the timeline. The emotioncomputational model based on fuzzy theory is a kind of emotion model in the digital art field that combines Hevner emotion ring and the fuzzy logic language. And the architecture of its emotion linguistic variables is shown in Figure 2, So far, the methodology of linguistic variables based on fuzzy theory is almost the standard method for the linguistic computational model, such as the group decision model based on numerical information and language information, as well as the multi-granularity linguistic information fusion model, theselinguistic computational models are based on fuzzy membership functions to represent the semantics of thelinguistic value, and apply the fuzzy logic operator as the linguistic aggregation operator. In summary, the fuzzy computational model based on the concept of linguistic variables is quite successful Computational Value ModelBased on Semantic Similarity The computational model based on fuzzy theory integrates the fuzzy characteristics of the connotation of the music emotion, and at the same time takes into account that the fuzzy system does not require the precise mathematical model, to facilitate the utilization of human experience and knowledge, with nonlinear, robust and other advantages, and adopts the fuzzy set to describe the music features, to perform analysis and recognition by fuzzy logic and fuzzy reasoning, whichcan be said to be closer to people's cognitive process for music, though there are still limitations in the fuzzy computational model. In order to overcome these limitations, the linguistic computational model based on semantic similarity relations is developed, and such linguistic computational model emphasizes the semantic similarity relationship between the linguistic values, and assumes that this similarity relationship is a basic characteristic of the human brain linguistic cognition, which is in line with the behavioral patterns of music emotion recognition. However, when this model definesthe logical operation between different linguistic expressions, it specifies that different expressions have the same priority, which may be somewhat different from the actual situation. Liu Tao et al. further develops and makes improvement on this basis, apply the linguistic computational model based on semantic similarity to the study on music emotion, and define the music emotion linguistic value model and the music emotion vectors as the following: Definition 1 (Linguistic Model): Diad <LA, R>represents the linguistic model: LA L1. L2. L3,, Ln (1) R rij n n, rij 0,1, i, j 1,2,, n (2) Where LA is a set constituted by finite linguistic values, and R is defined as the fuzzy similarity relationship in LA, n represents the number of elements in the set of linguistic values, the elements in R represent thedegree of overlap in the semantics between two linguistic values Li and Lj. Obviously the fuzzy relation matrix R is a symmetric matrix, which meets the following two properties: rij rji and rii 1. For the studies on music emotion, this model set the basic linguistic value set in LA into signature emotion words of eight subclasses of the emotion space: LAoM = {Sacred, Sorrow, Longing, Lyrical, Light, Joyous,, 1,2,,8 r Music, LAoMi to Enthusiastic, Vigorous}, abbreviated as : LAoM LAoMi i. And use represent the degree of similarity of the i-th element in the music and linguistic value set. Definition 2 (Music Emotion Vector): For the music with independent emotion semantics, its emotional connotation is defined as the eight-dimensional vector E in the Hevner emotion ring, element ei r Music, LAoMi represents the semantic similarity relationship of music and each sub emotion value linguistic value, with values 0-1 to representtheir degree of similarity, and the vector is called music emotion vector E : E r Music, LAoM1,, r Music, LAoMi,, r Music, LAoM 8 (3) Wherein, sub-emotion with the largest value is defined as the dominant emotion of the music Edon: Edon max ei, i 1,2,,8 max () represents the LAoM corresponding to the maximum value, for example, through certain process of reasoning, the emotion of a piece of music M can be expressed as: (0.2, 0.6, 0.9, 0.4, 0.3, 0.1, 0.0, 0.1), its dominant emotion is "longing", and its emotional semantic similarity value is 0.9, and at the same time the music also contains the emotional connotation of "sorrow", though in relatively low degree, 0.6, which can be called secondary emotion, therefore, the emotional connotation of music M is described as "Yearning very much, and somewhat a little sorrow". Of course, some music has several dominant emotions and secondary emotions.

4 The linguistic value model proposed by Liu Tao et al. mainly has makes development in two aspects: Firstly, define the music emotion vector and its similarity measurement rule through the semantic similarityrelationship; and secondly, expand the logic of the emotional vector expression and rules of addition and subtraction, so as to enhance its knowledge representation capacity. It is precisely based on these two points that the result of the integration of a number of complicated emotions of dominant and secondary emotions can be obtained through the reasoning of music emotion in this model, which is completely consistent with the complex feature of music emotion. Direct test method is applied to obtain the similarity matrix as shown in Table 1. Table 1. Similarity Emotion Model Test Obtained Similarity Matrix # Sacred Sorrow Longing Lyrical Light Joyous Enthusiastic Vigorous Sacred Sorrow Longing Lyrical Light Joyous Enthusiastic Vigorous AUTOMATIC RECOGNITION MODEL FOR MUSIC EMOTION 3.1 Automatic Recognition Model Based on Neural Network Pre-processing of Data (1)Extraction ofelements Three elements including pitch, sound intensity and sound length are extracted, which constitute the essential elements of music feature space, in accordance with the characteristic main melody emotion feature vector model, in the expectationto expand into an 8-dimensional vector space with the elements including:(pitch register, intensity, intensity stability, intensity orientation, melody orientation, pitch stability, interval stability, and interval span). However, as in the large number of samples, the amplitude of the same song shows theuniform distribution state, the intensity stability of different emotional categories of music and the intensity orientation do not have much distinction, with relatively small contribution to the music emotion classification, which means that the outline of the emotional connotation of music is characterized by a small number of elements that have relatively important roles. Herein the eight-dimensional feature space is streamlined into six dimensions, removing two vectors of the intensity stability and orientation. (2)Establishment ofsamples Through professional and popular music websites, Shanghai Conservatory of MusicElectronic Music Library and other channels and resources, a total of 580 pieces of MIDI music is obtained. Then according to the typical expression of emotions of the music and the production effects of MIDI music files, 190 songs are carefully selected, including 48 in the sorrow class, 51 in the joyous class, 50 in the sacred class, and 41 in the longingclass. Its approximate distribution is shown in Table 2 as follows. Table 2.Experimental Music Statistics Table Music Type Quantity Demonstration Music Film and television music 32 Knife Man Theme Song, " Princess Pearl", etc. Popular music 51 "Beautiful Mood", "If the Cloud Knows" etc. Classic music 16 "Hungarian Dance", "William Tell Overture" etc Religious music 43 Anthem, choirteaching music etc. Chinese folk music 48 "Happiness", "Fengyang Drum Dance", "Happy New Year" etc. (3)Screening of Samples For the experimental samples established heretofore, due to the inevitableerror in the measurement, if they are adopted directly as the test data, the result of classification may be slightly rough. In fact, select any of 15 samples randomly from each of the four emotions directly as the training data, and take the rest as the test data, the correct rate obtained for the neural network classification is only about 50%, which is caused by the relatively huge error of the data. Therefore, the screening on the test samples is necessary. Take the element of

5 music emotion in the sacred class of pitch register to make the description. In the judgment aspect of the data outliers, select the quantile instead of the mean value or variance, and the reason is that the outliers will affect the mean value and the variance but will not have any interference to the quantile. For the sacred class of 44 sets of data, make the box diagram, as shown in Figure 3, inside the box is the data within the 25% --75% of the fractile, and the length of tentacle is 1.5 times of thequartile deviation, and the longitudinal axis is the pitch register value, and the horizontal axis is the pitch value corresponding to the box midline. As can be seen, at both ends of the outer tentacles there is still some data, which isrelatively further away from the group, as the outliers.make the same box diagram for the other three types of emotional elements, and some outliers will also be found. Thus, study the samples one by one first. Ifamong the sample vectors that have been examined, one component is identified as an outlier, mark this sample, after the investigation on all samples is completed, remove the marked samples, so that the rest of the data has relatively focused trend. At the time when encountering the outlier, instead of removing it immediately, only mark it, so as not to cause the problem of different screening results due to the order of investigation, and also to retain as many samples as possible. Figure 3.Sacred Emotion Data Distribution Box Diagram (4)Selection of the Training Data Although the removal of the outlier results in four groups of data with relatively smaller error compared to the previous data, the selection of training data can directly affectthe results of the calculation, it is expected that the training data can reflect the most primary characteristics of the four categories of emotions as much as possible. Therefore, it is necessary to select a portion of the representative data from the screened data. For the four categories of emotions, calculate their respective center vector Si, i 1, 2,3, 4, by the definition of distance in the Euclidean space, calculate the distance dij for each element from its center vector in the four categories of emotions. The smaller dij is, the closer it means that this sample is from the center vector, that is, the more it can reflect the main characteristics of such emotions. Therefore, for each i, choose the smallest 15 dij as the training data, and a total of 60 training data is obtained Design of Network (1) CharacteristicData of the Sample Input: Take the first 60 training data which is previously obtained, and each sample is a six-dimensional vector. (2) Target Sample: In order to verify the experiment model more intuitively, herein the mapping of the target sample space, that is, the emotion representation space is temporarily taken as four-dimensional vector, and select the fourdimensional vector with the relatively large degree of distinction: Sacred, joyous, sorrow and longing. Take a four-dimensional vector: (1, 0, 0, 0), (0, 1, 0, 0), (0, 0, 1, 0) and (0, 0, 0, 1) to correspond to the four emotions of the sacred, joyous, longing and sorrow. (3) Structure of the Network: According to Kolmogorov theorem, adopt a N 2N 1M 3-layer BP network as the condition classifier. Where N represents the number of components of the input feature vector, M is the total number of categories of the output status. In accordance with the general design, the intermediate layer neuron transfer function S -type tangent function, the output layer neuron transfer function is S -type logarithmic function. The selection of the S -type logarithmic function is because the function is 0-1 function, just to meet the output requirements of the classifier.

6 The S -type logarithmic function graphs are shownrespectively in figure 4(1),4(2) as follows: Figure 4 (1). S-type Tangent Function Figure 4 (2).S-type Logarithmic Function Parameter Setting Set the maximum number of cycles as 100, the training error as , and the training function as Levenberg-Marquardt BP training function. The learning function takes momentum gradient descent learning function, and the performance function takes the mean square error function. Levenberg-Marquardt BP method introduction: Take d for the gradient direction, J for the Jacobi matrix, and the algorithm is as follows: (1)Take the initial point T k k (2)If k k k 1 x and v set precision : Set k 1 ; 1 0 J x r x, then stop. Otherwise solve the linear equations T T k x J x J x v I d J x r x to obtain (3)Set k 1 k k x x d ; And calculate (4)If k 0.25, then set v k1 4v k, if 0.75 (5)Set k k 1, go to(2) Result of Recognition k x d ; k1 k f x f x T T 1 T T k k k k k k k J x r x d d J x J x d k, k k 1 2 v v, otherwise, set v 1 2 k v Firstly, BP network training, the MATLAB BP network operating result is as shown in Figure 5 as follows: When processing tothe 63 step of iteration, the network training error is less than so as to converge. In MATLAB the specific training process of the BP network is shown in Figure 6 as below: k

7 Figure 5. BP NetworkTraining Process Note: The test environment is Windows Vista system, Matlab 7.0 platform. So far, the BP network training process is completed. Select 131 sets of data for the test, of which the first 22 sample emotions are sacred, the subsequent 37 sample emotions are joyous, then the next 39 sample emotions are longing, and the last 33 sample emotions are sorrow. And the determination results are shown in Figure 7 as the following. Wherein, in the BP network, use four sets of different four-element column vectors, to represent four different corresponding music emotionsrespectively: [1,0,0,0] represents sacred; [0,1,0,0] represents joyous; [0,0,1,0] represents longing; [0,0,0,1] represents sorrow. As shown in the figure, since the recognition model cannot achieve 100% recognition accuracy, each emotion may have different misjudgment phenomenon. Figure 6. Recognition Result of the Neural Network 4. EXPERIMENTAL COMPARISON We make the experimental comparison on the algorithm proposed in this paper and the automatic recognition model based on the statistical classification. The automatic recognition model based on statistic classification also makes pre-treatment on the sample data, then adopts the Fisher discriminant method to construct the music emotion recognition model in MATLAB, and distinguishes these 131 sampleslikewise, finally getsthe recognition results as shown in Figure 7 as follows: Figure 7.Recognition Result of the Statistical Classification

8 Make comparison and analysis on both recognition models, and do statistics on the music emotion recognition results obtained by the BP network recognition model and the statistical classification model, as shown in Figure 8. Figure 8.Comparison of the Impact of the Pre-processing of Statistical Classification Data to the Experimental Results As can be seen, the neural network classification results after the data processing (with the accuracy of 80.9%) is significantly better than the results before the data processing (71%). However, although the data has been processed, the actual error is still inevitable; therefore, the calculation accuracy of 80.9% is a pretty good recognition result. On the other hand, after the data is processed, apply the Fisher discriminant method, the classification result is about 2% higher than the result before the data is processed. Therefore, the following conclusions can be obtained: (1) As there is error in the data, all kinds of emotion data points will be interwoven in the vector space, since the projection process is a linear transformation, when the data integration applies the Fisher discriminant method to be projected on the vector, this will havea pretty huge impact on its accuracy. (2) In fact, the Fisher discriminant method is suitable for processing the linearly separable issue, which is more similar to the linear neural network. While the BP network can also handle the complicated nonlinear problems very perfectly, once the data error is not too huge, the classification results will be satisfactory (3) For the fault tolerance, an error of data will have certain influence on the Fisher classification method, however, it may almost certainly have no impact on BP network at all, because the destruction of a few neurons will not affect the characteristics of the overall network, therefore, for data that contains noise, BP network is more competent. (4) BP network has relatively strong adaptive capacity. (5) Neural network adopts the fuzzy logic to simulate the way people think, thus, the application of neural network for the fuzzy classification issues such as emotion classification has more reference value than the statistical methods. (6) In the aspect of the calculation of the time, as the neural network is a numerical iterative method, therefore, it is inevitably slower than the speed of statistical classification, but for the issue with the size of scale that is not particularly big, the time is acceptable. 5.CONCLUSION From the perspective of emotion computing, this paper analyzed the emotion recognition and the information processing characteristics according to artificial neural network, and proposed the research ideas by the application of BP network for music emotion classification and recognition. The paper analyzed the acoustic principles and the composition of MIDI music files, as the basis for the extraction of the basic music features; and summarized the development, features, models, operating principles and learning algorithms of the neural network as the basic method of recognition. And on the basis of the aforementioned two aspects, it extracted the melody characteristic parameters of the MIDI music files; designed the BP network model suitable for the music emotion recognition, and realized the music emotion recognition based on BP network, and performed verification on the algorithm through the samples. REFERENCES Agustus J.L., Mahoney C.J., Downey L.E., et al. (2015) Functional MRI of Music Emotion Processing in Frontotemporal Dementia, Annals of the New York Academy of Sciences, 1337(1), pp Balkwill L.L., Thompson W.F., Matsunaga R.I. (2004) Recognition of Emotion in Japanese, Western, and Hindustani Music by Japanese Listeners, Japanese Psychological Research, 46(4), p.p Fritz T., Jentschke S., Gosselin, N., et al. (2009) Universal Recognition of Three Basic Emotions in Music, Current Biology, 19(7), pp

9 Gosselin N., Peretz I., Johnsen E., et al. (2014) Amygdala Damage Impairs Emotion Recognition from Music,Neuropsychologia, 52(2), pp Gosselin N., Peretz I., Noulhiane M., et al. (2015) Impaired Recognition of Scary Music Following Unilateral Temporal Lobe Excision,Brain, 138(3), pp Han B.J., Rho S., Jun S., et al. (2010) Music Emotion Classification and Context-based Music Recommendation,Multimedia Tools and Applications, 47(3), pp Juslin, P.N., Laukka P. (2003) Communication of Emotions in Vocal Expression and Music Performance: Different Channels, Same Code,Psychological Bulletin, 129(5), pp Kim J., André E. (2008) Emotion Recognition Based on Physiological Changes in Music Listening,Pattern Analysis and Machine Intelligence, IEEE Transactions on, 30(12), pp Koelsch S., Fritz T., Müller K., et al. (2009) Investigating Emotion with Music: An FMRI Study,Human Brain Mapping, 30(3), pp Peretz I., Gagnon L., Bouchard B. (1998) Music and Emotion: Perceptual Determinants, Immediacy, and Isolation after Brain Damage,Cognition, 68(2), pp Wu B., Zhong E., Horner A., et al. (2014) Music Emotion Recognition by Multi-label Multi-layer Multi-instance Multiview Learning, In Proceedings of the ACM International Conference on Multimedia, pp Yang D., Lee W.S. (2004) Disambiguating Music Emotion Using Software Agents,In ISMIR, 4(2), pp

CS229 Project Report Polyphonic Piano Transcription

CS229 Project Report Polyphonic Piano Transcription CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project

More information

Distortion Analysis Of Tamil Language Characters Recognition

Distortion Analysis Of Tamil Language Characters Recognition www.ijcsi.org 390 Distortion Analysis Of Tamil Language Characters Recognition Gowri.N 1, R. Bhaskaran 2, 1. T.B.A.K. College for Women, Kilakarai, 2. School Of Mathematics, Madurai Kamaraj University,

More information

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance Methodologies for Expressiveness Modeling of and for Music Performance by Giovanni De Poli Center of Computational Sonology, Department of Information Engineering, University of Padova, Padova, Italy About

More information

Subjective Similarity of Music: Data Collection for Individuality Analysis

Subjective Similarity of Music: Data Collection for Individuality Analysis Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp

More information

Chord Classification of an Audio Signal using Artificial Neural Network

Chord Classification of an Audio Signal using Artificial Neural Network Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------

More information

2. AN INTROSPECTION OF THE MORPHING PROCESS

2. AN INTROSPECTION OF THE MORPHING PROCESS 1. INTRODUCTION Voice morphing means the transition of one speech signal into another. Like image morphing, speech morphing aims to preserve the shared characteristics of the starting and final signals,

More information

Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting

Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting Dalwon Jang 1, Seungjae Lee 2, Jun Seok Lee 2, Minho Jin 1, Jin S. Seo 2, Sunil Lee 1 and Chang D. Yoo 1 1 Korea Advanced

More information

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

Music Emotion Recognition. Jaesung Lee. Chung-Ang University Music Emotion Recognition Jaesung Lee Chung-Ang University Introduction Searching Music in Music Information Retrieval Some information about target music is available Query by Text: Title, Artist, or

More information

Broken Wires Diagnosis Method Numerical Simulation Based on Smart Cable Structure

Broken Wires Diagnosis Method Numerical Simulation Based on Smart Cable Structure PHOTONIC SENSORS / Vol. 4, No. 4, 2014: 366 372 Broken Wires Diagnosis Method Numerical Simulation Based on Smart Cable Structure Sheng LI 1*, Min ZHOU 2, and Yan YANG 3 1 National Engineering Laboratory

More information

Keywords: Edible fungus, music, production encouragement, synchronization

Keywords: Edible fungus, music, production encouragement, synchronization Advance Journal of Food Science and Technology 6(8): 968-972, 2014 DOI:10.19026/ajfst.6.141 ISSN: 2042-4868; e-issn: 2042-4876 2014 Maxwell Scientific Publication Corp. Submitted: March 14, 2014 Accepted:

More information

ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC

ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC Vaiva Imbrasaitė, Peter Robinson Computer Laboratory, University of Cambridge, UK Vaiva.Imbrasaite@cl.cam.ac.uk

More information

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng

The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng The Research of Controlling Loudness in the Timbre Subjective Perception Experiment of Sheng S. Zhu, P. Ji, W. Kuang and J. Yang Institute of Acoustics, CAS, O.21, Bei-Si-huan-Xi Road, 100190 Beijing,

More information

A Music Retrieval System Using Melody and Lyric

A Music Retrieval System Using Melody and Lyric 202 IEEE International Conference on Multimedia and Expo Workshops A Music Retrieval System Using Melody and Lyric Zhiyuan Guo, Qiang Wang, Gang Liu, Jun Guo, Yueming Lu 2 Pattern Recognition and Intelligent

More information

Exploring Relationships between Audio Features and Emotion in Music

Exploring Relationships between Audio Features and Emotion in Music Exploring Relationships between Audio Features and Emotion in Music Cyril Laurier, *1 Olivier Lartillot, #2 Tuomas Eerola #3, Petri Toiviainen #4 * Music Technology Group, Universitat Pompeu Fabra, Barcelona,

More information

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM A QUER B EAMPLE MUSIC RETRIEVAL ALGORITHM H. HARB AND L. CHEN Maths-Info department, Ecole Centrale de Lyon. 36, av. Guy de Collongue, 69134, Ecully, France, EUROPE E-mail: {hadi.harb, liming.chen}@ec-lyon.fr

More information

A Categorical Approach for Recognizing Emotional Effects of Music

A Categorical Approach for Recognizing Emotional Effects of Music A Categorical Approach for Recognizing Emotional Effects of Music Mohsen Sahraei Ardakani 1 and Ehsan Arbabi School of Electrical and Computer Engineering, College of Engineering, University of Tehran,

More information

Research on sampling of vibration signals based on compressed sensing

Research on sampling of vibration signals based on compressed sensing Research on sampling of vibration signals based on compressed sensing Hongchun Sun 1, Zhiyuan Wang 2, Yong Xu 3 School of Mechanical Engineering and Automation, Northeastern University, Shenyang, China

More information

Music Genre Classification

Music Genre Classification Music Genre Classification chunya25 Fall 2017 1 Introduction A genre is defined as a category of artistic composition, characterized by similarities in form, style, or subject matter. [1] Some researchers

More information

A Case Based Approach to the Generation of Musical Expression

A Case Based Approach to the Generation of Musical Expression A Case Based Approach to the Generation of Musical Expression Taizan Suzuki Takenobu Tokunaga Hozumi Tanaka Department of Computer Science Tokyo Institute of Technology 2-12-1, Oookayama, Meguro, Tokyo

More information

MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC

MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC Lena Quinto, William Forde Thompson, Felicity Louise Keating Psychology, Macquarie University, Australia lena.quinto@mq.edu.au Abstract Many

More information

Hidden Markov Model based dance recognition

Hidden Markov Model based dance recognition Hidden Markov Model based dance recognition Dragutin Hrenek, Nenad Mikša, Robert Perica, Pavle Prentašić and Boris Trubić University of Zagreb, Faculty of Electrical Engineering and Computing Unska 3,

More information

MUSI-6201 Computational Music Analysis

MUSI-6201 Computational Music Analysis MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)

More information

Supervised Learning in Genre Classification

Supervised Learning in Genre Classification Supervised Learning in Genre Classification Introduction & Motivation Mohit Rajani and Luke Ekkizogloy {i.mohit,luke.ekkizogloy}@gmail.com Stanford University, CS229: Machine Learning, 2009 Now that music

More information

Experiments on musical instrument separation using multiplecause

Experiments on musical instrument separation using multiplecause Experiments on musical instrument separation using multiplecause models J Klingseisen and M D Plumbley* Department of Electronic Engineering King's College London * - Corresponding Author - mark.plumbley@kcl.ac.uk

More information

Speech and Speaker Recognition for the Command of an Industrial Robot

Speech and Speaker Recognition for the Command of an Industrial Robot Speech and Speaker Recognition for the Command of an Industrial Robot CLAUDIA MOISA*, HELGA SILAGHI*, ANDREI SILAGHI** *Dept. of Electric Drives and Automation University of Oradea University Street, nr.

More information

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES PACS: 43.60.Lq Hacihabiboglu, Huseyin 1,2 ; Canagarajah C. Nishan 2 1 Sonic Arts Research Centre (SARC) School of Computer Science Queen s University

More information

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular Music Mood Sheng Xu, Albert Peyton, Ryan Bhular What is Music Mood A psychological & musical topic Human emotions conveyed in music can be comprehended from two aspects: Lyrics Music Factors that affect

More information

LSTM Neural Style Transfer in Music Using Computational Musicology

LSTM Neural Style Transfer in Music Using Computational Musicology LSTM Neural Style Transfer in Music Using Computational Musicology Jett Oristaglio Dartmouth College, June 4 2017 1. Introduction In the 2016 paper A Neural Algorithm of Artistic Style, Gatys et al. discovered

More information

Expressive performance in music: Mapping acoustic cues onto facial expressions

Expressive performance in music: Mapping acoustic cues onto facial expressions International Symposium on Performance Science ISBN 978-94-90306-02-1 The Author 2011, Published by the AEC All rights reserved Expressive performance in music: Mapping acoustic cues onto facial expressions

More information

A Framework for Segmentation of Interview Videos

A Framework for Segmentation of Interview Videos A Framework for Segmentation of Interview Videos Omar Javed, Sohaib Khan, Zeeshan Rasheed, Mubarak Shah Computer Vision Lab School of Electrical Engineering and Computer Science University of Central Florida

More information

The Design of Teaching Experiment System Based on Virtual Instrument Technology. Dayong Huo

The Design of Teaching Experiment System Based on Virtual Instrument Technology. Dayong Huo 3rd International Conference on Management, Education, Information and Control (MEICI 2015) The Design of Teaching Experiment System Based on Virtual Instrument Technology Dayong Huo Department of Physics,

More information

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods

Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods Kazuyoshi Yoshii, Masataka Goto and Hiroshi G. Okuno Department of Intelligence Science and Technology National

More information

Analysis of local and global timing and pitch change in ordinary

Analysis of local and global timing and pitch change in ordinary Alma Mater Studiorum University of Bologna, August -6 6 Analysis of local and global timing and pitch change in ordinary melodies Roger Watt Dept. of Psychology, University of Stirling, Scotland r.j.watt@stirling.ac.uk

More information

Detection of Panoramic Takes in Soccer Videos Using Phase Correlation and Boosting

Detection of Panoramic Takes in Soccer Videos Using Phase Correlation and Boosting Detection of Panoramic Takes in Soccer Videos Using Phase Correlation and Boosting Luiz G. L. B. M. de Vasconcelos Research & Development Department Globo TV Network Email: luiz.vasconcelos@tvglobo.com.br

More information

Type-2 Fuzzy Logic Sensor Fusion for Fire Detection Robots

Type-2 Fuzzy Logic Sensor Fusion for Fire Detection Robots Proceedings of the 2 nd International Conference of Control, Dynamic Systems, and Robotics Ottawa, Ontario, Canada, May 7 8, 2015 Paper No. 187 Type-2 Fuzzy Logic Sensor Fusion for Fire Detection Robots

More information

Outline. Why do we classify? Audio Classification

Outline. Why do we classify? Audio Classification Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification Implementation Future Work Why do we classify

More information

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng

Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the

More information

Singer Traits Identification using Deep Neural Network

Singer Traits Identification using Deep Neural Network Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic

More information

DISTRIBUTION STATEMENT A 7001Ö

DISTRIBUTION STATEMENT A 7001Ö Serial Number 09/678.881 Filing Date 4 October 2000 Inventor Robert C. Higgins NOTICE The above identified patent application is available for licensing. Requests for information should be addressed to:

More information

Opening musical creativity to non-musicians

Opening musical creativity to non-musicians Opening musical creativity to non-musicians Fabio Morreale Experiential Music Lab Department of Information Engineering and Computer Science University of Trento, Italy Abstract. This paper gives an overview

More information

Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors *

Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors * Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors * David Ortega-Pacheco and Hiram Calvo Centro de Investigación en Computación, Instituto Politécnico Nacional, Av. Juan

More information

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY Eugene Mikyung Kim Department of Music Technology, Korea National University of Arts eugene@u.northwestern.edu ABSTRACT

More information

MODELING MUSICAL MOOD FROM AUDIO FEATURES AND LISTENING CONTEXT ON AN IN-SITU DATA SET

MODELING MUSICAL MOOD FROM AUDIO FEATURES AND LISTENING CONTEXT ON AN IN-SITU DATA SET MODELING MUSICAL MOOD FROM AUDIO FEATURES AND LISTENING CONTEXT ON AN IN-SITU DATA SET Diane Watson University of Saskatchewan diane.watson@usask.ca Regan L. Mandryk University of Saskatchewan regan.mandryk@usask.ca

More information

Brain.fm Theory & Process

Brain.fm Theory & Process Brain.fm Theory & Process At Brain.fm we develop and deliver functional music, directly optimized for its effects on our behavior. Our goal is to help the listener achieve desired mental states such as

More information

Normalized Cumulative Spectral Distribution in Music

Normalized Cumulative Spectral Distribution in Music Normalized Cumulative Spectral Distribution in Music Young-Hwan Song, Hyung-Jun Kwon, and Myung-Jin Bae Abstract As the remedy used music becomes active and meditation effect through the music is verified,

More information

Research on the Development of Education Level of University Sports Aesthetics Based on AHP

Research on the Development of Education Level of University Sports Aesthetics Based on AHP OPEN ACCESS EURASIA Journal of Mathematics Science and Technology Education ISSN: 1305-8223 (online) 1305-8215 (print) 2017 13(8):5133-5140 DOI: 10.12973/eurasia.2017.00988a Research on the Development

More information

Music Information Retrieval with Temporal Features and Timbre

Music Information Retrieval with Temporal Features and Timbre Music Information Retrieval with Temporal Features and Timbre Angelina A. Tzacheva and Keith J. Bell University of South Carolina Upstate, Department of Informatics 800 University Way, Spartanburg, SC

More information

The Design of Efficient Viterbi Decoder and Realization by FPGA

The Design of Efficient Viterbi Decoder and Realization by FPGA Modern Applied Science; Vol. 6, No. 11; 212 ISSN 1913-1844 E-ISSN 1913-1852 Published by Canadian Center of Science and Education The Design of Efficient Viterbi Decoder and Realization by FPGA Liu Yanyan

More information

Wipe Scene Change Detection in Video Sequences

Wipe Scene Change Detection in Video Sequences Wipe Scene Change Detection in Video Sequences W.A.C. Fernando, C.N. Canagarajah, D. R. Bull Image Communications Group, Centre for Communications Research, University of Bristol, Merchant Ventures Building,

More information

Keywords Separation of sound, percussive instruments, non-percussive instruments, flexible audio source separation toolbox

Keywords Separation of sound, percussive instruments, non-percussive instruments, flexible audio source separation toolbox Volume 4, Issue 4, April 2014 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Investigation

More information

VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS. O. Javed, S. Khan, Z. Rasheed, M.Shah. {ojaved, khan, zrasheed,

VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS. O. Javed, S. Khan, Z. Rasheed, M.Shah. {ojaved, khan, zrasheed, VISUAL CONTENT BASED SEGMENTATION OF TALK & GAME SHOWS O. Javed, S. Khan, Z. Rasheed, M.Shah {ojaved, khan, zrasheed, shah}@cs.ucf.edu Computer Vision Lab School of Electrical Engineering and Computer

More information

Characterization and improvement of unpatterned wafer defect review on SEMs

Characterization and improvement of unpatterned wafer defect review on SEMs Characterization and improvement of unpatterned wafer defect review on SEMs Alan S. Parkes *, Zane Marek ** JEOL USA, Inc. 11 Dearborn Road, Peabody, MA 01960 ABSTRACT Defect Scatter Analysis (DSA) provides

More information

Subjective evaluation of common singing skills using the rank ordering method

Subjective evaluation of common singing skills using the rank ordering method lma Mater Studiorum University of ologna, ugust 22-26 2006 Subjective evaluation of common singing skills using the rank ordering method Tomoyasu Nakano Graduate School of Library, Information and Media

More information

Interactive Classification of Sound Objects for Polyphonic Electro-Acoustic Music Annotation

Interactive Classification of Sound Objects for Polyphonic Electro-Acoustic Music Annotation for Polyphonic Electro-Acoustic Music Annotation Sebastien Gulluni 2, Slim Essid 2, Olivier Buisson, and Gaël Richard 2 Institut National de l Audiovisuel, 4 avenue de l Europe 94366 Bry-sur-marne Cedex,

More information

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring 2009 Week 6 Class Notes Pitch Perception Introduction Pitch may be described as that attribute of auditory sensation in terms

More information

Deep Neural Networks Scanning for patterns (aka convolutional networks) Bhiksha Raj

Deep Neural Networks Scanning for patterns (aka convolutional networks) Bhiksha Raj Deep Neural Networks Scanning for patterns (aka convolutional networks) Bhiksha Raj 1 Story so far MLPs are universal function approximators Boolean functions, classifiers, and regressions MLPs can be

More information

Robert Alexandru Dobre, Cristian Negrescu

Robert Alexandru Dobre, Cristian Negrescu ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Automatic Music Transcription Software Based on Constant Q

More information

Detection and demodulation of non-cooperative burst signal Feng Yue 1, Wu Guangzhi 1, Tao Min 1

Detection and demodulation of non-cooperative burst signal Feng Yue 1, Wu Guangzhi 1, Tao Min 1 International Conference on Applied Science and Engineering Innovation (ASEI 2015) Detection and demodulation of non-cooperative burst signal Feng Yue 1, Wu Guangzhi 1, Tao Min 1 1 China Satellite Maritime

More information

Color Image Compression Using Colorization Based On Coding Technique

Color Image Compression Using Colorization Based On Coding Technique Color Image Compression Using Colorization Based On Coding Technique D.P.Kawade 1, Prof. S.N.Rawat 2 1,2 Department of Electronics and Telecommunication, Bhivarabai Sawant Institute of Technology and Research

More information

LOW-COMPLEXITY BIG VIDEO DATA RECORDING ALGORITHMS FOR URBAN SURVEILLANCE SYSTEMS

LOW-COMPLEXITY BIG VIDEO DATA RECORDING ALGORITHMS FOR URBAN SURVEILLANCE SYSTEMS LOW-COMPLEXITY BIG VIDEO DATA RECORDING ALGORITHMS FOR URBAN SURVEILLANCE SYSTEMS Ling Hu and Qiang Ni School of Computing and Communications, Lancaster University, LA1 4WA, UK ABSTRACT Big Video data

More information

Sudhanshu Gautam *1, Sarita Soni 2. M-Tech Computer Science, BBAU Central University, Lucknow, Uttar Pradesh, India

Sudhanshu Gautam *1, Sarita Soni 2. M-Tech Computer Science, BBAU Central University, Lucknow, Uttar Pradesh, India International Journal of Scientific Research in Computer Science, Engineering and Information Technology 2018 IJSRCSEIT Volume 3 Issue 3 ISSN : 2456-3307 Artificial Intelligence Techniques for Music Composition

More information

Analysis on the Value of Inner Music Hearing for Cultivation of Piano Learning

Analysis on the Value of Inner Music Hearing for Cultivation of Piano Learning Cross-Cultural Communication Vol. 12, No. 6, 2016, pp. 65-69 DOI:10.3968/8652 ISSN 1712-8358[Print] ISSN 1923-6700[Online] www.cscanada.net www.cscanada.org Analysis on the Value of Inner Music Hearing

More information

Speech Recognition and Signal Processing for Broadcast News Transcription

Speech Recognition and Signal Processing for Broadcast News Transcription 2.2.1 Speech Recognition and Signal Processing for Broadcast News Transcription Continued research and development of a broadcast news speech transcription system has been promoted. Universities and researchers

More information

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm Georgia State University ScholarWorks @ Georgia State University Music Faculty Publications School of Music 2013 Chords not required: Incorporating horizontal and vertical aspects independently in a computer

More information

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 AN HMM BASED INVESTIGATION OF DIFFERENCES BETWEEN MUSICAL INSTRUMENTS OF THE SAME TYPE PACS: 43.75.-z Eichner, Matthias; Wolff, Matthias;

More information

Automatic Piano Music Transcription

Automatic Piano Music Transcription Automatic Piano Music Transcription Jianyu Fan Qiuhan Wang Xin Li Jianyu.Fan.Gr@dartmouth.edu Qiuhan.Wang.Gr@dartmouth.edu Xi.Li.Gr@dartmouth.edu 1. Introduction Writing down the score while listening

More information

Journal Citation Reports on the Web. Don Sechler Customer Education Science and Scholarly Research

Journal Citation Reports on the Web. Don Sechler Customer Education Science and Scholarly Research Journal Citation Reports on the Web Don Sechler Customer Education Science and Scholarly Research don.sechler@thomsonreuters.com Introduction JCR distills citation trend data for over 10,000 journals from

More information

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis

Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis Fengyan Wu fengyanyy@163.com Shutao Sun stsun@cuc.edu.cn Weiyao Xue Wyxue_std@163.com Abstract Automatic extraction of

More information

Voice & Music Pattern Extraction: A Review

Voice & Music Pattern Extraction: A Review Voice & Music Pattern Extraction: A Review 1 Pooja Gautam 1 and B S Kaushik 2 Electronics & Telecommunication Department RCET, Bhilai, Bhilai (C.G.) India pooja0309pari@gmail.com 2 Electrical & Instrumentation

More information

Predicting Performance of PESQ in Case of Single Frame Losses

Predicting Performance of PESQ in Case of Single Frame Losses Predicting Performance of PESQ in Case of Single Frame Losses Christian Hoene, Enhtuya Dulamsuren-Lalla Technical University of Berlin, Germany Fax: +49 30 31423819 Email: hoene@ieee.org Abstract ITU s

More information

The Inspiration of Folk Fine Arts based on Common Theoretical Model to Modern Art Design

The Inspiration of Folk Fine Arts based on Common Theoretical Model to Modern Art Design Abstract The Inspiration of Folk Fine Arts based on Common Theoretical Model to Modern Art Design Wenquan Wang Yanan University Art Institute of LuXun, Yan an 716000, China Cultural connotation and humanity

More information

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu

More information

Machine Vision System for Color Sorting Wood Edge-Glued Panel Parts

Machine Vision System for Color Sorting Wood Edge-Glued Panel Parts Machine Vision System for Color Sorting Wood Edge-Glued Panel Parts Q. Lu, S. Srikanteswara, W. King, T. Drayer, R. Conners, E. Kline* The Bradley Department of Electrical and Computer Eng. *Department

More information

A Novel Video Compression Method Based on Underdetermined Blind Source Separation

A Novel Video Compression Method Based on Underdetermined Blind Source Separation A Novel Video Compression Method Based on Underdetermined Blind Source Separation Jing Liu, Fei Qiao, Qi Wei and Huazhong Yang Abstract If a piece of picture could contain a sequence of video frames, it

More information

6.UAP Project. FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System. Daryl Neubieser. May 12, 2016

6.UAP Project. FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System. Daryl Neubieser. May 12, 2016 6.UAP Project FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System Daryl Neubieser May 12, 2016 Abstract: This paper describes my implementation of a variable-speed accompaniment system that

More information

Automatic Rhythmic Notation from Single Voice Audio Sources

Automatic Rhythmic Notation from Single Voice Audio Sources Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung

More information

DEVELOPMENT OF MIDI ENCODER "Auto-F" FOR CREATING MIDI CONTROLLABLE GENERAL AUDIO CONTENTS

DEVELOPMENT OF MIDI ENCODER Auto-F FOR CREATING MIDI CONTROLLABLE GENERAL AUDIO CONTENTS DEVELOPMENT OF MIDI ENCODER "Auto-F" FOR CREATING MIDI CONTROLLABLE GENERAL AUDIO CONTENTS Toshio Modegi Research & Development Center, Dai Nippon Printing Co., Ltd. 250-1, Wakashiba, Kashiwa-shi, Chiba,

More information

Music Segmentation Using Markov Chain Methods

Music Segmentation Using Markov Chain Methods Music Segmentation Using Markov Chain Methods Paul Finkelstein March 8, 2011 Abstract This paper will present just how far the use of Markov Chains has spread in the 21 st century. We will explain some

More information

Topics in Computer Music Instrument Identification. Ioanna Karydi

Topics in Computer Music Instrument Identification. Ioanna Karydi Topics in Computer Music Instrument Identification Ioanna Karydi Presentation overview What is instrument identification? Sound attributes & Timbre Human performance The ideal algorithm Selected approaches

More information

Design of Fault Coverage Test Pattern Generator Using LFSR

Design of Fault Coverage Test Pattern Generator Using LFSR Design of Fault Coverage Test Pattern Generator Using LFSR B.Saritha M.Tech Student, Department of ECE, Dhruva Institue of Engineering & Technology. Abstract: A new fault coverage test pattern generator

More information

Comparison of Dictionary-Based Approaches to Automatic Repeating Melody Extraction

Comparison of Dictionary-Based Approaches to Automatic Repeating Melody Extraction Comparison of Dictionary-Based Approaches to Automatic Repeating Melody Extraction Hsuan-Huei Shih, Shrikanth S. Narayanan and C.-C. Jay Kuo Integrated Media Systems Center and Department of Electrical

More information

TERRESTRIAL broadcasting of digital television (DTV)

TERRESTRIAL broadcasting of digital television (DTV) IEEE TRANSACTIONS ON BROADCASTING, VOL 51, NO 1, MARCH 2005 133 Fast Initialization of Equalizers for VSB-Based DTV Transceivers in Multipath Channel Jong-Moon Kim and Yong-Hwan Lee Abstract This paper

More information

Adaptive Key Frame Selection for Efficient Video Coding

Adaptive Key Frame Selection for Efficient Video Coding Adaptive Key Frame Selection for Efficient Video Coding Jaebum Jun, Sunyoung Lee, Zanming He, Myungjung Lee, and Euee S. Jang Digital Media Lab., Hanyang University 17 Haengdang-dong, Seongdong-gu, Seoul,

More information

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC Scientific Journal of Impact Factor (SJIF): 5.71 International Journal of Advance Engineering and Research Development Volume 5, Issue 04, April -2018 e-issn (O): 2348-4470 p-issn (P): 2348-6406 MUSICAL

More information

Detecting and Analyzing System for the Vibration Comfort of Car Seats Based on LabVIEW

Detecting and Analyzing System for the Vibration Comfort of Car Seats Based on LabVIEW Detecting and Analyzing System for the Vibration Comfort of Car Seats Based on LabVIEW Ying Qiu Key Laboratory of Conveyance and Equipment, Ministry of Education School of Mechanical and Electronical Engineering,

More information

Brain-Computer Interface (BCI)

Brain-Computer Interface (BCI) Brain-Computer Interface (BCI) Christoph Guger, Günter Edlinger, g.tec Guger Technologies OEG Herbersteinstr. 60, 8020 Graz, Austria, guger@gtec.at This tutorial shows HOW-TO find and extract proper signal

More information

Formalizing Irony with Doxastic Logic

Formalizing Irony with Doxastic Logic Formalizing Irony with Doxastic Logic WANG ZHONGQUAN National University of Singapore April 22, 2015 1 Introduction Verbal irony is a fundamental rhetoric device in human communication. It is often characterized

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

Various Artificial Intelligence Techniques For Automated Melody Generation

Various Artificial Intelligence Techniques For Automated Melody Generation Various Artificial Intelligence Techniques For Automated Melody Generation Nikahat Kazi Computer Engineering Department, Thadomal Shahani Engineering College, Mumbai, India Shalini Bhatia Assistant Professor,

More information

HUMMING METHOD FOR CONTENT-BASED MUSIC INFORMATION RETRIEVAL

HUMMING METHOD FOR CONTENT-BASED MUSIC INFORMATION RETRIEVAL 12th International Society for Music Information Retrieval Conference (ISMIR 211) HUMMING METHOD FOR CONTENT-BASED MUSIC INFORMATION RETRIEVAL Cristina de la Bandera, Ana M. Barbancho, Lorenzo J. Tardón,

More information

Computational Modelling of Harmony

Computational Modelling of Harmony Computational Modelling of Harmony Simon Dixon Centre for Digital Music, Queen Mary University of London, Mile End Rd, London E1 4NS, UK simon.dixon@elec.qmul.ac.uk http://www.elec.qmul.ac.uk/people/simond

More information

Smart Traffic Control System Using Image Processing

Smart Traffic Control System Using Image Processing Smart Traffic Control System Using Image Processing Prashant Jadhav 1, Pratiksha Kelkar 2, Kunal Patil 3, Snehal Thorat 4 1234Bachelor of IT, Department of IT, Theem College Of Engineering, Maharashtra,

More information

Retrieval of textual song lyrics from sung inputs

Retrieval of textual song lyrics from sung inputs INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Retrieval of textual song lyrics from sung inputs Anna M. Kruspe Fraunhofer IDMT, Ilmenau, Germany kpe@idmt.fraunhofer.de Abstract Retrieving the

More information

ONE SENSOR MICROPHONE ARRAY APPLICATION IN SOURCE LOCALIZATION. Hsin-Chu, Taiwan

ONE SENSOR MICROPHONE ARRAY APPLICATION IN SOURCE LOCALIZATION. Hsin-Chu, Taiwan ICSV14 Cairns Australia 9-12 July, 2007 ONE SENSOR MICROPHONE ARRAY APPLICATION IN SOURCE LOCALIZATION Percy F. Wang 1 and Mingsian R. Bai 2 1 Southern Research Institute/University of Alabama at Birmingham

More information

SDR Implementation of Convolutional Encoder and Viterbi Decoder

SDR Implementation of Convolutional Encoder and Viterbi Decoder SDR Implementation of Convolutional Encoder and Viterbi Decoder Dr. Rajesh Khanna 1, Abhishek Aggarwal 2 Professor, Dept. of ECED, Thapar Institute of Engineering & Technology, Patiala, Punjab, India 1

More information

DELTA MODULATION AND DPCM CODING OF COLOR SIGNALS

DELTA MODULATION AND DPCM CODING OF COLOR SIGNALS DELTA MODULATION AND DPCM CODING OF COLOR SIGNALS Item Type text; Proceedings Authors Habibi, A. Publisher International Foundation for Telemetering Journal International Telemetering Conference Proceedings

More information

INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR NPTEL ONLINE CERTIFICATION COURSE. On Industrial Automation and Control

INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR NPTEL ONLINE CERTIFICATION COURSE. On Industrial Automation and Control INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR NPTEL ONLINE CERTIFICATION COURSE On Industrial Automation and Control By Prof. S. Mukhopadhyay Department of Electrical Engineering IIT Kharagpur Topic Lecture

More information

CS 591 S1 Computational Audio

CS 591 S1 Computational Audio 4/29/7 CS 59 S Computational Audio Wayne Snyder Computer Science Department Boston University Today: Comparing Musical Signals: Cross- and Autocorrelations of Spectral Data for Structure Analysis Segmentation

More information

Embodied music cognition and mediation technology

Embodied music cognition and mediation technology Embodied music cognition and mediation technology Briefly, what it is all about: Embodied music cognition = Experiencing music in relation to our bodies, specifically in relation to body movements, both

More information

Reducing False Positives in Video Shot Detection

Reducing False Positives in Video Shot Detection Reducing False Positives in Video Shot Detection Nithya Manickam Computer Science & Engineering Department Indian Institute of Technology, Bombay Powai, India - 400076 mnitya@cse.iitb.ac.in Sharat Chandran

More information