An Automated Music Selector Derived from Weather Condition and its Impact on Human Psychology

Similar documents
The relationship between properties of music and elicited emotions

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

Expressive information

THE EFFECT OF EXPERTISE IN EVALUATING EMOTIONS IN MUSIC

A Mood-Based Music Classification and Exploration System. Owen Craigie Meyers

Affective response to a set of new musical stimuli W. Trey Hill & Jack A. Palmer Psychological Reports, 106,

1. BACKGROUND AND AIMS

This slideshow is taken from a conference presentation (somewhat modified). It summarizes the Temperley & Tan 2013 study, and also talks about some

Using machine learning to decode the emotions expressed in music

A Categorical Approach for Recognizing Emotional Effects of Music

The Role of Time in Music Emotion Recognition

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset

Music Mood Classification - an SVM based approach. Sebastian Napiorkowski

Automatic Music Clustering using Audio Attributes

MODELING MUSICAL MOOD FROM AUDIO FEATURES AND LISTENING CONTEXT ON AN IN-SITU DATA SET

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

Exploring Relationships between Audio Features and Emotion in Music

Electronic Musicological Review

Automatic Detection of Emotion in Music: Interaction with Emotionally Sensitive Machines

Can Song Lyrics Predict Genre? Danny Diekroeger Stanford University

Research & Development. White Paper WHP 228. Musical Moods: A Mass Participation Experiment for the Affective Classification of Music

Expressive performance in music: Mapping acoustic cues onto facial expressions

Compose yourself: The Emotional Influence of Music

MUSICAL MOODS: A MASS PARTICIPATION EXPERIMENT FOR AFFECTIVE CLASSIFICATION OF MUSIC

Music Similarity and Cover Song Identification: The Case of Jazz

Searching for the Universal Subconscious Study on music and emotion

MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC

Quality of Music Classification Systems: How to build the Reference?

EMOTIONS IN CONCERT: PERFORMERS EXPERIENCED EMOTIONS ON STAGE

HOW TO STUDY: YEAR 11 MUSIC 1

Emotions perceived and emotions experienced in response to computer-generated music

Audio Feature Extraction for Corpus Analysis

The Human Features of Music.

PRESCHOOL (THREE AND FOUR YEAR-OLDS) (Page 1 of 2)

DIGITAL AUDIO EMOTIONS - AN OVERVIEW OF COMPUTER ANALYSIS AND SYNTHESIS OF EMOTIONAL EXPRESSION IN MUSIC

Measuring a Measure: Absolute Time as a Factor in Meter Classification for Pop/Rock Music

TOWARDS AFFECTIVE ALGORITHMIC COMPOSITION

Improving Music Mood Annotation Using Polygonal Circular Regression. Isabelle Dufour B.Sc., University of Victoria, 2013

MOTIVATION AGENDA MUSIC, EMOTION, AND TIMBRE CHARACTERIZING THE EMOTION OF INDIVIDUAL PIANO AND OTHER MUSICAL INSTRUMENT SOUNDS

Enhancing Music Maps

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

Subjective Emotional Responses to Musical Structure, Expression and Timbre Features: A Synthetic Approach

Improving Piano Sight-Reading Skills of College Student. Chian yi Ang. Penn State University

Keywords: Edible fungus, music, production encouragement, synchronization

AudioRadar. A metaphorical visualization for the navigation of large music collections

Foundation - MINIMUM EXPECTED STANDARDS By the end of the Foundation Year most pupils should be able to:

Emotional Remapping of Music to Facial Animation

MUSICAL EAR TRAINING THROUGH ACTIVE MUSIC MAKING IN ADOLESCENT Cl USERS. The background ~

Sound visualization through a swarm of fireflies

Singer Traits Identification using Deep Neural Network

Detecting Musical Key with Supervised Learning

A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES

Therapeutic Sound for Tinnitus Management: Subjective Helpfulness Ratings. VA M e d i c a l C e n t e r D e c a t u r, G A

Influence of timbre, presence/absence of tonal hierarchy and musical training on the perception of musical tension and relaxation schemas

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

A Recipe for Emotion in Music (Music & Meaning Part II)

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

Dimensional Music Emotion Recognition: Combining Standard and Melodic Audio Features

Music Information Retrieval

18 Benefits of Playing a Musical Instrument

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

INFLUENCE OF MUSICAL CONTEXT ON THE PERCEPTION OF EMOTIONAL EXPRESSION OF MUSIC

ATSSB Bb clarinet (revised February 2016) Artistic Studies Book I from the French School David Hite/Southern Music

An Indian Journal FULL PAPER ABSTRACT KEYWORDS. Trade Science Inc.

STRAND I Sing alone and with others

TongArk: a Human-Machine Ensemble

Notes on David Temperley s What s Key for Key? The Krumhansl-Schmuckler Key-Finding Algorithm Reconsidered By Carley Tanoue

An Integrated Music Chromaticism Model

Analysis on the Value of Inner Music Hearing for Cultivation of Piano Learning

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm

Chapter Five: The Elements of Music

they in fact are, and however contrived, will be thought of as sincere and as producing music from the heart.

Interactive Visualization for Music Rediscovery and Serendipity

Robert Alexandru Dobre, Cristian Negrescu

Does Music Directly Affect a Person s Heart Rate?

Active learning will develop attitudes, knowledge, and performance skills which help students perceive and respond to the power of music as an art.

Lyric-Based Music Mood Recognition

ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE

Perceptual dimensions of short audio clips and corresponding timbre features

The purpose of this essay is to impart a basic vocabulary that you and your fellow

Brain.fm Theory & Process

Using Genre Classification to Make Content-based Music Recommendations

CTP431- Music and Audio Computing Music Information Retrieval. Graduate School of Culture Technology KAIST Juhan Nam

Environment Expression: Expressing Emotions through Cameras, Lights and Music

Montana Instructional Alignment HPS Critical Competencies Music Grade 3

Therapeutic Function of Music Plan Worksheet

Arts, Computers and Artificial Intelligence

ART I: UNIT THREE DESIGN PERSONALITY

Surprise & emotion. Theoretical paper Key conference theme: Interest, surprise and delight

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Arts Education Essential Standards Crosswalk: MUSIC A Document to Assist With the Transition From the 2005 Standard Course of Study

The Million Song Dataset

PEP-Lower Elementary Report Card 12-13

2014 Music Performance GA 3: Aural and written examination

Katie Rhodes, Ph.D., LCSW Learn to Feel Better

Amplitude and Loudness 1

Interactive Music: Compositional Techniques for Communicating Different Emotional Qualities

HST 725 Music Perception & Cognition Assignment #1 =================================================================

mood into an adequate input for our procedural music generation system, a scientific classification system is needed. One of the most prominent classi

Transcription:

DOI 10.7603/s4060101400148 An Automated Music Selector Derived from Weather Condition and its Impact on Human Psychology Debajyoti Karmaker, Md. Al Imran, Niaj Mohammad, Mohaiminul Islam and Md. Nafees Mahbub Received 20 May 2015 Accepted 10 Jun 2015 emotion. For example, a listeners with happy mood may want to listen Don`t Worry, Be Happy or Hotat Roud, they will not care about the artist, they just need that specific speech to satisfy them. Abstract Sometimes it is disquieting to generate a playlist to listen music for a specific moment. Though listening of music basically depends on our mood and it s also been said that there exists a relation between our mood and weather, so our approach is to build an automated system to create a music playlist based on users mood and defined weather. Method is to measure the weight of each music files respect to defined mood and weather by using data mining algorithms. Keywords Decision Rule, Sentiment Analysis, Music Mining, Machine learning, MSV, WSV, NLP. I. INTRODUCTION Verbis defectis musica incipit that s Latin for, Music takes us where word cannot. Music is an important part of our life. There is a strong relationship between music and human mind. In most researchers opinion, people value music primarily because of the emotions it evokes. As Juslin says People use music to change emotions, to release emotions, to match their current emotion, to enjoy or comforts themselves and to relieve stress. He also represents a novel theoretical framework featuring six additional mechanisms through which music listening may induce emotions: brain stem reflexes, evaluative conditioning, emotional contagion, visual imagery, episodic memory, and musical expectancy [4]. As listening of music varies with emotions, a better music classification is recommended to create a welldefined automated music playlist. Music usually classified according to genre and for limitation to cover these huge number of music collection subgenre are introduced. If we go little deep while we are listening music, an artist or genre is not the primary key for music selection to satisfy our emotional state. There is two main roles go with our emotion, one is musical instrument and another one is tone. Sometimes lyrics reflect on our Weather is another factor for a music listening which we cannot realize most of the time. As we listen music to satisfy our emotion and mood states changes with weather. Different categories of weather are responsible for changing in emotions, which indicates a character either positive or negative for the organism, but these changes, varies from one to another [6]. For example, in rainy weather a listener may want to listen a sad music to evoke his/her emotion; on other hand another one wants to a happy music. In other word we can say, weather helps us to feel our emotion more deeply. In our work, we propose a different list of music based on mood and weather. It is recommended that listening of music depends on our mood. The interesting part is that, from this model we can define the relation between human mind and weather. II. BACKGROUND STUDY With the development of modern science people are more and more reliable on newly invented technologies. From professional needs at work place or personal activities like listening to music they rely on modern tools and techniques. Today s music listener faces various obstacles in finding music for a specific context. With the extension of digital music library day by day it is getting more complicated to classify or categorize music according to user s emotional specifications. As a result the need for music classification tools is becoming more and more apparent. Listeners now a day s need tools that could classify their large music library into mood based category so that they could enjoy music in a new and exciting way. Music has various attributes that influence DOI: 10.5176/22513043_4.3.327 13

a tool to create playlists for every occasion, with the ability to tag and share music with everyone you know [2]. basic human emotions as like as human emotions changes with the changing of weather state. The work presented here is to find a relation amongst music, weather and human emotion and why they influence each other. Also updating our classification [31] through user feedback and machine learning. With the development of digital music technology, it is essential to develop the music recommendation system which recommends music for users. Some work has been done on the personalized music recommendation to recommend based on the users preference. There exist two major approaches for the personalized music recommendation. One is the contentbased filtering approach which analyzes the content of music that users liked in the past and recommends the music with relevant content. The other is the collaborative filtering approach which recommends music that peer group of similar preference liked. Both recommendation approaches are based on the users preferences observed from the listening behavior [7]. Many of the current music recommendation systems follow the rule of social tagging rather than feature extraction of song. Now days there are many online social tagging sites which provide tagging system for classifying songs last.fm3 is one of them. Although these social tagging sites are useful but, they cannot keep it up to the mark when it comes to actual user preferences. Which is important in classifying mood based music. As mentioned previously recommending personalized music based on the users preference one of the two approaches is collaborative filtering approach. Ringo is a pioneering music recommendation system based on the collaborative filtering approach, where the preference of a user is acquired by the user s rating of music. Similar users are identified by comparing the preferences. Ringo predicts the preference of new music for a user by computing a weighted average of all ratings given by peer group of similar preference [1]. 1 http://www.stereomood.fm MusicCat2 is a music recommendation agent based on user modeling, where the user model is defined by the user. It contains information about user s habit, preference or user defined features, etc. MusicCat can automatically choose music from user s music collection according to user s model. MRS is a system which provides music recommendation based on music grouping and user interests. MoodLogic3 allows the user to generate "mood" based playlist based on the mood of the user. It uses a central database to allow users to collaboratively profile music by mood. The software was also capable of organizing a music collection based on a "fingerprint" of the song. Mood Logic would generate this fingerprint of the song, based on some social tagging websites like last.fm4, Qloud5, MyStrands6, upload it to the server and wait for a response from the user. Once the feedback tag is received from the user MoodLogic updates the tag for that track and use it for future recommendations for the users. This meant a user could have a collection of incorrectly tagged mp3's and the software would be able to correctly identify, tag, and even organize the songs into categories [3]. Human mind is changing consequently and weather plays an important role when it comes to music selection. Above mentioned models and application focuses on the mood of the user. Our goal is to develop a model that can analyze users preferred mood and weather to come up with a more comprehensive outcome for our endusers. A. Psychology of Music and Human Mind Here we will discuss about those musical emotions which are mostly related with music and also about those musical acoustic features will be introduced, which has a great role to influence on human mind. 1) Musical Emotion/Mood The relationship between emotion and music is important when mapping various musical parameters to an emotion space. Emotion is a complex set of interactions among subjective and objective factors, lead by neural/hormonal systems, which can give rise to affective experiences such as feelings of arousal, pleasure/displeasure [14]. There are vast array of emotional states that human are capable of experiencing. It relates to our everyday life as a result it is comprised of various terms and definitions. Research indicates that people value music primarily because of the emotions it evokes. Affect, mood, emotion and arousal are often used interchangeably though each is unique and differentiable 2 http://www.musiccat.com 3 http://www.moodlogic.com 4 http://www.last.fm 5 http://www.qloud.com 6 http://www.mystrands.com Stereomood1 is a music streaming service that plays music tailored to your mood and daily activities. It's 14

Literature Emotions Imberty 1979 Balkwill 1999 Thompson 1992 Maher 1982 Gundlach 1935 Watson 1942 Krumhansl 1997 Lindstrom 1997 Hevner 1936 Excitement Serenity Happiness Sadness Anger Fear Disgust Boredom Agitation Surprise Tender Dignified Gloom Triumphant Uneasy Sentimental LIST OF EMOTIONS AND THE THEIR CITATIONS IN LITERATURE ON MUSIC AND EMOTION [13] Juslin 1997 With the flow of time various approaches were implied to represent human mood in some sort of categorical form. Such approaches include James Russell's twodimensional bipolar space (valencearousal) [16], Robert Thayer's energystress model [18, 19] (see Figure 1), where contentment is defined as low energy/low stress, depression as low energy/ high stress, exuberance as high energy/low stress, and anxious/frantic as high energy, high stress [17]. TABLE I. Scherer 1977 from each other. Mood affective state, which is the broadest of emotional states, may have some degree of positive or negative valence, which is the measure of the state's emotional charge [15]. An emotional state is largely influenced by the underlying mood and affective state of the individual. A highly aroused emotional state will be very apparent in the individual. TABLE II. LIST OF EMOTIONS AND THE THEIR CITATIONS IN LITERATURE ON MUSIC AND EMOTION [13] Literature Emotions Wexner 1954 Kouwer 1949 Wright 1962 Valdez 1994 Kaya 2004 Rusinek 2004 Terwogt 1995 Kandinsky 1912 Cullari 2000 Excitement Serenity Happiness Sadness Anger Fear Disgust Boredom Agitation Quiet Lazy Secure Serious Radiant Romantic Comfort Kopacz 2003 The basic twodimensional models have also been expanded to circular models, such as Russell's circumplex model [24] and Kate Hevner's adjective circle [20]. Russell's has expressed the basic mood of human kind in a circular format in his circumplex model (see Figure 2), which contains a quadrant graph format in which some general mood types like happy, sad, angry, serene are expressed in each of the four quadrants with some of its similar listed moods. Poffenberger 1924 Figure 1: Thayer's twodimensional energystress model [18] 2) Music & Emotion : In the mid 20th century, scholars and researchers began to think how various audio features can play a role in transforming human mood. Researchers like Hevner, Russell, Rigg, Thayer and Watson began to make progress in relating specific musical features, such as mode, harmony, tempo, rhythm, and dynamics (loudness), to emotions and moods. Such models have classified Figure 2 : Russell's multidimensional circumplex model [24] Researchers from time to time have researched through the complex mode of human emotions and psychology in order to find some sort of pattern that may lay between human moods on music selection. Some of the famous citations are stated in the Table 1 and Table 2. 15

while negative values translate to minor mode, complex harmony, slow tempo, irregular rhythm, and low loudness. emotions along some axes, such as valence (pleasure), arousal (activity). Hevner's studies [23, 20, 21, 22] focus on the affective value of six musical features and how they relate to emotion. The results of these studies are summarized in Table 3. The six musical elements explored in these studies include mode, tempo, pitch (register), rhythm, harmony, and melody. These features are mapped to a circular model of affect encompassing eight different emotional categories. Major modes are often associated with happiness, gracefulness and solemnity while minor modes are related to the emotions of sadness, dreaminess, disgust, and anger. Simple harmonies, or consonant chords, such as major chords, are often pleasant, happy, and relaxed. Complex harmonies contain dissonant notes that create instability in a piece of music and activate emotions of excitement, tension, anger, and sadness. In Rigg s experiment; joy is described as having iambic rhythm (staccato notes), fast tempo, high register, major mode, simple harmony, and loud dynamics (forte). TABLE III. HEVNER S WEIGHTING OF MUSICAL CHARACTERISTICS IN 8 AFFECTIVE STATES [22] Musical Element Mode Tempo Pitch Rhythm Harmony Melody Musical Element Mode Tempo Pitch Rhythm Harmony Melody Dignified/ Solemn major 4 slow 14 low 10 firm 18 simple 3 ascend 4 Sad/ Heavy minor 20 12 slow low 19 firm 3 complex 7 Dreamy/ Sentimental minor 12 slow 16 high 6 flowing 9 simple 4 Serene/ Gentle major 3 slow 20 high 8 flowing 2 simple 10 ascend 3 Graceful/ Sparkling major 21 fast 6 high 16 flowing 8 simple 12 descend 3 Happy/ Bright major 24 20 fast high 6 flowing 10 simple 16 Exciting/ Elated fast 21 low 9 firm 2 complex 14 descend 7 Vigorous/ Majestic fast 6 low 13 firm 10 complex 8 descend 8 Watson s studies differ from those of Hevner and Rigg because he uses fifteen adjective groups in conjunction with the musical attributes pitch (lowhigh), volume (softloud), tempo (slowfast), sound (prettyugly), dynamics (constantvarying), and rhythm (regularirregular). Watson s research reveals many important relationships between these musical attributes and the perceived emotion of the musical excerpt. Mayers states that, loudness aligns itself roughly along the yaxis of arousal. High arousal and excitement are generally the result of loud music and peaceful and delicate emotions which are triggered by soft music [5]. The weightings for each feature and emotion are shown in Table 4. Positive values translate to major mode, simple harmony, fast tempo, regular rhythm, and high loudness, TABLE IV. MAPPING OF MUSICAL CIRCUMPLEX MODEL OF EMOTION [5] Mood Mode Pleasure Excitement 12 24 0 0 16 8 12 3 Arousal Distress Displeasure Depression Sleepiness Relaxation FEATURES TO RUSSELL S Harmony Tempo Rhythm Loudness 11 16 14 14 2 4 4 10 7 20 21 21 14 7 16 20 5 10 2 2 3 10 9 2 0 10 20 10 0 10 20 10 During the last decade, many researchers (Feng, Zhuang & Pan, 2003; Juslin & Sloboda, 2001; Lu, Liu & Zhang, 2006) have investigated the influence of music factors such as loudness and tonality on the perceived emotional expression. They analyzed those data using diverse techniques, some of which are involved in measuring psychological and physiological correlation between the state of a particular musical factor and emotion evocation. Researchers to researchers the musical factor and emotional model varies, but they not say so far from each other. Table 5 describes the emotional models of Russell, Schubert, and Hevner and details their position in both the circumplex space and the twodimensional valencearousal space [5]. TABLE V. COMPARISON OF THREE EMOTIONAL MODELS, IN TERMS OF VALENCE AND AROUSAL [5] Degree Russell Schubert 0 Pleasure Lyrical 45 90 135 Excitement Arousal Distress Bright Dramatic Tense 180 Displeasure Tragic 225 Depression 270 315 Sleepiness Relaxation Dark Majestic Dreamy Calm Hevner Serene/ Graceful Happy Exciting Exciting Sad/ Dreamy Dignified/ Sad/Vigorous Dreamy Serene Valence Arousal o o o o 3) Psychology of Weather and Human Mind Human emotions is a complex but influential structure, which is influenced by the slightest change in the environment. There is a weak but significant relationship between weather and human mood. Researchers have hypnotized that mood state mediates the relationship between weather and human behavior. Weather is widely believed to influence people s mood. For example, the majority of people think they feel happier 16

enhance the user s music listening experience to a new level and to remove the hassle of searching through large music database that suits the user need. on days with a lot of sunshine as compared to dark and rainy days. Although this association seems to be common sense [25]. Researchers like Howard and Hoffman (1984) have found that there is a significant effect on mood correlated with the weather, especially with regards to humidity (a component of weather not always measured) [25]. III. PROPOSED ARCHITECTURE A. Humidity, temperature, and hours of sunshine had the greatest effect on mood. High levels of humidity lowered scores on concentration while increasing reports of sleepiness. Rising temperatures lowered anxiety and scepticism mood scores [25]. The number of hours of sunshine was found to predict optimism scores significantly. As the number of hours of sunshine increased, optimism scores also increased [25]. Another researcher named Keller and his colleagues (2005) examined 605 participants responses in three separate studies to examine the connection between mood states and weather [26]. They found that: Pleasant weather (higher temperature or barometric pressure) was related to higher mood, better memory, and broadened cognitive style during the spring as time spent outside increased. The same relationships between mood and weather were not observed during other times of year, and indeed hotter weather was associated with lower mood in the summer [26]. Research has proven that warm temperatures and exposure to sunshine have the greatest positive impact on moods. A report published in the British Journal of Psychology found that warmer temperatures lowered anxiety and skepticism while more hours of sunshine increased positive thinking. The same study showed that high levels of humidity made it hard to concentrate, increasing fatigue and sleepiness, depresses [27]. Music Theory The field of music psychology dates to the 18th century, beginning with J.P.Rameau in 1772 [7].The psychologically based fields of music perception and explore how scientific representations of audio signals in the environment differ from representations within human mind. This includes the representation of the pitch, harmony, loudness, mode, tempo, rhythm [7]. Henver s studies focus on these six musical features related to human mood [8, 9]. Our proposed system maintains a stable database populated with the most popular tracks. Each of the tracks are taken and driven into the systems feature extractor to classify the mood for the track. The feature extractor process depends on two term; 1) Musical Acoustic Feature and 2) Lyric mining. 1) Musical Acoustic Feature Based on the discussion states above we select six musical acoustic features for our system; are Mode, Tempo, Pitch, Rhythm, Harmony and Dynamics, which play vital role in deciding the human emotions. Below is a brief description about them and their key role on human mind. Mode indicates the modality (major or minor) of a track, the type of scale from which it s melodic is derived; is a set of musical notes forming a scale and from which melodies and harmonies are constructed [10]. Major modes are often associated with happiness, gracefulness and solemnity while minor modes are related to the emotions of sadness, dreaminess and anger [11]. For example at high temperature and when barometric wind pressure is high there is a pleasant weather which gives a positive effect on mood[26, 27].On the other hand low temperature high levels of humidity and low hours on sunshine give a negative effect[26]. Tempo is defined as the speed at which a passage of music is or should be played [10], and is typically measured in beats per minute (bpm). A fast tempo falls into the range of 140 to 200bpm and slow tempo could be anywhere between 40 and 80 bpm. Depending on other musical factors, a fast tempo can trigger such emotions as excitement, joy, surprise or fear. Similarly a slow tempo is typically of calmness, dignity, sadness, tenderness, boredom or disgust [13]. After the process it gives us a float value which is referenced to our emotion model to get a relevant mood. The research presented in this thesis attempts to implement the relationship between human emotions and why it is influenced by various attributes of music and weather. The goal of this work was to classify music based on human emotion and current weather to recommend music according to user mood. Also to clarify the classification [31] and update it through machine learning. The moodweather based classification system is meant to Pitch (highlow) convey emotional responses include the amplitude envelops and interactions between factors [11]. Pitch level has a natural influence in musical expression. High pitch music is often perceived as happy, 17

serene, dreamy and expressive of surprise, anger and fear. Low pitch is associated with sadness, solemnity and boredom but sometimes with pleasantness, depending on the overall musical context. Pitch variations may also account for expressiveness, high variations associate with happiness and small variations with anger and fear [13]. pressure, wind, sunshine, cloudiness and precipitation; which are mostly responsible to effect on human mind. Each one is described below: Rhythm is defined as the systematic arrangement of musical sounds, principally according to duration and periodic stress [10]. Rhythm can be categorized as regular/irregular (Watson), smooth/rough (Gundlach), firm/flowing (Hevner) and simple/complex (Vercoe). The variations of the regularity or complexity of a rhythmic pattern in a piece of music trigger emotional responses. Regular and smooth rhythms are representative of happiness, dignity, majesty, and peace, while irregular and rough rhythms pair with amusement, uneasiness and anger. Harmony (simplecomplex) is the combination of simultaneously sounded musical notes to produce chords [12]. Simple harmonies, or constant chords, such as major chords, are often pleasant, happy and relaxed. Complex harmonies contain dissonant notes that create instability in a piece of music and activate emotions of excitement, tension, anger and sadness [11]. Dynamics represent varying volume levels of perceived intensity of sound. The dynamics of a piece of music may be either soft or loud [11]. A loud passage of music is associated with intensity, tension, anger, and joy and soft passages are associated with tenderness, sadness, solemnity, and fear. Large dynamic ranges signify fear, rapid changes in dynamics signify playfulness, and minimal variations relate to sadness and peacefulness [13]. 2) Lyric Mining The areas of natural language processing and textual analysis are relevant to the field of music recommendations and classification in that they provide tools with which to extract meaning and context from cultural metadata, such as music reviews or collaborative content websites. A valuable NLP tool is commonsense reasoning, which is practically suited to the analysis of song lyrics as it enables the mining of key concepts and contexts from the lyric. Common sense is defined simply as good sense and sound judgment in practical matters [10]. B. Weather Theory Temperature is the hotness or coldness of a substance and measured with a thermometer. Most temperature scales today are expressed in degrees Celsius (ºC), although one will sometime see Fahrenheit (ºF) in use [28]. Humidity is the amount of water vapour in the air. There are three main measurements of humidity: absolute, relative and specific [29]. Humidity gives rise to most weather phenomena: clouds, rain, snow, dew and fog [28]. Pressure is a force, or weight, exerted on a surface per unit area, and is measured in any unit of force divided by any unit of area, which called Pascal (Pa) [28]. Wind is the flow of gases on a large scale. The air is nearly always in motion, and this is felt as wind. Two factors are necessary to specify wind, its speed and direction. The wind speed can be expressed in miles or kilometers per hour, meters per second, and knots or as a force on the Beaufort scale [28]. Sunshine is the direct solar radiation is not blocked by clouds, a combination of bright light and radiant heat. When it is blocked by the clouds or reflects off other objects, it is experienced as diffused light [28]. Cloudiness is used to define the measurement of cloud; where cloud is a visible mass of liquid droplets or frozen crystals made of water or various chemicals suspended in the atmosphere above the surface of a planetary body [30]. Precipitation is the amount of rain, sleet; snow or hail which falls in a specified time is expressed as the depth of water it would produce on a large, level impermeable surface. Usually it is expressed in millimeters; although inches may sometimes be used [28]. i) The model which we have proposed has three variants: ii) Positive Effect (excited, happy, satisfied, serene, delighted) iii) Negative Effect (sad, distressed, tense, and angry) iv) Tiredness (sleepy, tired, bored) Based on the discussion states above (from chapter 2, section 2.4) we can get a view that weather has some significant contribution on influencing human emotions. One of our systems job is to get weather related data from the current location of the user locations. The key weather attributes are used in our system is temperature, humidity, In order to get the weather data our system call through the World Weather Online (WWO) API. The API first gathers the location of the user and the collects the local weather report of the particular location. Then the raw data is processed through the above three variants. 18

C. MusicEmotion Model D. MIR and MSV & WSV weight Generate As there are various musical emotions related to music listening. In these sense of classifying these emotions we are concentrating on 12 musical emotion or mood state (Excited, Delighted, Happy, Satisfied, Serene Sleepy, Tired, Bored, Sad, Distressed, Tense and Angry) for our system, which represent the core aspects of human emotions and all other emotional mood. By using these twelve mood state and six musical acoustic features we generates a mood state flow model depicted in Figure 3, which is fall under the four quadrate of Russell s circumplex model of emotion [24] and are also listed as adjacent mood in Henver s adjective circle [21]. 1) First it takes a single music file from Large Music Collection and generates a unique key. i) Extract Meta data (location in data store, title, artist name, album name, genre etc). ii) Extracts raw data from audio and retrieves six acoustic feature data (mode, tempo, pitch, rhythm, harmony, and dynamics). iii) For a lyric containing music file Lyric weight retrieves by using text mining. TABLE VI. RULESET FOR MSV WEIGHT WITH AFV DATA Acoustic Feature Value Pitch Tempo Rhythm Dynamics Lyric TABLE VII. Harmony Figure 3 : Mood State flow with Acoustic Feature 0 30 60 90 120 150 180 210 240 270 300 330 Mode Happy Delighted Excited Angry Tense Distressed Sad Bore Tired Sleepy Serene Satisfied Degree Mood State Value RULESET FOR MSV WEIGHT WITH WSV Overall system architecture is depicted in Figure 4. Weather State Value Humidity Pressure Wind Sunshine Cloudiness Precipitation Happy Delighted Excited Angry Tense Distressed Sad Bore Tired Sleepy Serene Satisfied Temperature Mood State Value 1) Next it generates MSV (Mood State Value) weight for each music file related with their acoustic features by using RuleSet for MSV weight with AFV data given in Table 6 which is followed by mood state flow diagram with acoustic feature (see Figure 3) and from those MSV weight corresponding WSV weight will be determined by Figure 4 : System Architecture 19

using RuleSet for MSV weight with WSV in given Table 7. E. Music Playlist Generator The playlist generator performs the task of generating the music playlist depending on the preferred mood and current weather by the user. 2) After retrieving and generating all meta data, AFV data, MSV and WSV weight the system stores them to the database in a single row. i) Collects listener current mood state and weather state from listener current location. ii) Makes a call for a list of music s from database which is highly corresponded with specified mood state and weather state. iii) Generate a music playlist ordered by higher weighted MSV and WSV. In this process higher weighted music will be suggested by the playlist generator which corresponds to the identified mood state (given by listener) and weather state (collected from listener current geometric location). If a mood state is not specified by listener then the system is able to create a music playlist without listener interaction by selecting those music file which corresponds to the current weather state. F. Update MSV & WSV weight from Listener Feedback One of the main and essential features of the system is updating MSV and WSV based on user feedback. In this part when music will be played, system will update the MSV & WSV weight for that particular music for which mood and weather state it was previously suggested. If the music only played by a suggested weather state then WSV weight value will only be updated. Weight value will be updated for each mood and weather state by the given direct/indirect feedback by the listener with value range between (0.1/0.1). Figure 5 : Automated music playlist generate system G. Learning Mechanism Machine learning is one of the core features of the system. In this learning process system will generate three separate dataset from database; containing each row for music with their AFV data, MSV weight and WSV weight, which is enriched by listener s feedback. These three dataset are named as AFV&MSV, AFV&WSV and MSV&WSV. From AFV&MSV dataset a decision rule will be generated to define a more appropriate weight for MSV corresponding to their AFV by using data mining algorithm. Similarly from AFV& WSV dataset a decision rule will be generate to define a more appropriate weight for WSV corresponding to their AFV by using data mining algorithm. Also a relation will be generated between mood and weather from MSV&WSV dataset. From these decisions Rule Set for MSV weight for AFV data and Rule Set for WSV with MSV will be updated. Day by day with more Figure 6 : Update process for MSV and/or WSV weight 20

[4] Patrik N. Juslin and Daniel Vastfjall. Emotional responses to music: The need to consider underlying mechanisms. Behavioral and Brain Sciences, 2008. [5] Owen Craigie Meyers. A MoodBased Music Classification and Exploration System. B.Mus., McGill University, 2004. [6] Z. Spasova. The effect of weather and its changes on emotional state individual characteristics that make us vulnerable. National Center of Public Health and Analyses, Sofia, Bulgaria, 27 March 2012. [7] FangFeiKuo, MengFen Chiang, ManKwan Shan and SuhYin Lee. Emotionbased Music Recommendation by Association Discovery from Film Music. [8] Huron D., History of Music Psychology. http://musiccog.ohiosate.edu/music829f/timeline.html, 02 February 2007. The main goal of this research paper was to find the relevant factors that exist amongst music, weather and human mind and to classify music based on user mood and current weather. A model that was built to classify music based on musical acoustic features and weather state. James Russel s dimensional model on emotions and Kate Henver s studies in music and emotions was especially useful for correlating the music features of a song with specific emotions. [9] K. Henver. The affect character of the major and minor in music. The American Journal of Psychology, 47(1):103118, 1935. To gain our ultimate goal, our system performs some tasks like extracting core music features like pitch, harmony, dynamics, rhythm, mode form a single music track also gathering local weather data and feeding it to our proposed musicemotion model to classify a mood for that particular song. [13] RemonaBehravan. Automatic Mapping of Emotion in Music to Abstract Visual Arts. A dissertation submitted for the degree of Engineering Doctorate of the University of London, April 2007. The combination of audio and weather data content is a circular factor in our mood classification system. Many recent attempts at music mood classification have relied solely on audio content. But adding weather based classification has improved the efficiency of the task thus gathering the knowledge about the relation between human emotions and weather. [15] J.A.Sloboda and P.N.Juslin. Music and Emotion: Theory and Research, chapter Psycho;ogical Perspectives on Music and Emotion, pages 71104. Oxford University Press, 2001. enriched data the accuracy of these three decision rules will be more accurate and system will be able to generate an automated music playlist more near to listener s satisfaction level. IV. CONCLUSION The research presented in this thesis is to find relation amongst music, weather and human mind. Human emotion is a complex subject to analyze as it is not static, which changes with the slightest change in environment. Music also is not an easy nut to crack as it contains a magnitude of independent and dependent parameters. [10] K.Henver. Experimental studies of the elements of expression in music. The American Journal of Psychology, 48(2):246268, 1936. [11] E.McKean. New Oxford American Dictionary. Oxford University Press, New York, 2nd edition, 2005. [12] A.Gabrielsson and E.Lindstrom. Music and Emotion: Theory and Research, chapter The Influence of Musical Structure on Emotions Expression, pages 223248. Oxford University Press, 2001. One of the weaknesses of classifying system lies in its audio lies in its audio feature extraction stage. A misclassified musical feature such as tempo or mode will have a large on the mood classification of song. Fault is also part of learning process. By incorporating this type of mood based digital music classifier into one s everyday music pattern enhances and simulates one s overall music experience. [14] P.R.Kleinginna and A.M.Kleinginna. A categorized lists of emotions definitions, with a suggestion from a consensual definition. Motivation and Emotion, 5(4):34579, 1981. [16] J.A.russel. Affective space is bipolar. Journal of Personality and Social Psychology, 37(3):345356, 1979. [17] A. Mehrabian. pleasurearousaldominance: A general framework for describing and measuring individual. Current Psychology, 14(4):261292, 1996. [18] R.E.thayer. The Biopsychology of Mood and Arousal. Oxford University Press, Oxford, 1989. [19] R.E.thayer. The origin of everyday moods: managing energy, tension and stress. Oxford University Press, Oxford, 1996. [20] K.Henver. The affective value of pitch and tempo in music. The American Journal of Psychology, 49(4):621630, 1937. [21] K.Henver. Experimental studies of the elements of expression in music. The American Journal of Psychology, 49(4):246268, 1936. [22] K.Henver. Expression in music: A discussion of experimental studies and theories. Psychology Review, 42:186204, 1935. REFERENCES [1] Shardanan&Mayers. 1995. [2] Giovanni Ferron, Daniele Novaga, Maurizio PraticiEleonoraViviani, Silvia Pianelli. Stereomood: Website for listening to music according to their emotions or activities, February 2007. [3] [23] K.Henver. The affective value of pitch and tempo in music. The American Journal of Psychology, 47(1):103118, 1935. Ringo: A Music Recommender System, [24] J.A.russel. A circumplex model of effect. Journal of Personality and Social Psychology, 39(6):11611178, 1980. [25] Howarth E. & Hoffman M.S. A multidimensional approach to the relationship between mood and weather. British Journal of Psychology, 75(1), 1523, 1984. Tom Sulzer, Christian Pirkner, Elion Chin and Tom Andreas Weigend; MoodLogic: An Online Recommendation System, 1998. [26] Keller Matthew C., Fredrickson Barbara L., Ybarra Oscar, Côté, Stéphane, Johnson Kareem, Mikels Joe, Conway Anne, Wager 21

Tor. A Warm Heart and a Clear Head: The Contingent Effects of Weather on Mood and Cognition. Psychological Science, 16(9),724731, 2005. Bachelors and Master s degree in Computer Science from American International University Bangladesh (AIUB). [27] Jennie Wood. The Weather and Your Mood. British Journal of Psychology, Association for Psychological Science, American Academy of Allergy, Asthma, & Immunology. [28] Physical Attributes of Weather, available on http://www.weatherclimate.org.uk/05.php, last accessed in August 2013. [29] WikipediaHumidity, available on http://en.wikipedia.org/wiki/humidity, last accessed in August 2013. [30] WikipediaCloud, available on http://en.wikipedia.org/wiki/cloud, last accessed in August 2013. [31] Debajyoti Karmaker, Hafizur Rahman, Mohammad Saiedur Rahaman, and Md. Kamrul Bari; A Fine Grained Technique for Viral Marketing Based on Social Network: A Machine Learning Approach. International Journal of Science and Technology, 1(2):8995, 2011. Niaj Mohammad completed his B.Sc. in Computer Science and Engineering in 2014 from American International University Bangladesh (AIUB). After completing the bachelors degree, his interests for research inspired him for higher education in his preferred area of knowledge. He is currently enrolled in Data Warehousing and Data Mining in Khulna University of Engineering and Technology. AUTHOR S PROFILE Debajyoti Karmaker was enthralled by the power of computers and intrigued by the ideal of becoming a computer scientist from his childhood. With the evolution of time, he is now working as an Assistant Professor in the Department of Computer Science at American International University Bangladesh (AIUB) from where he earned both of his B.Sc. and M.Sc. in Computer Science. Throughout his academic and professional career, he was always a curious and research active person and at this point he has the urge to undergo further research and delve deeper into the subject matters at an institute of international eminence under the guidance of distinguished researchers which would strengthen his knowledge and enable him to realize his potential to the maximum. Mohaiminul Islam is a Software Engineer at the Bengal Solutions Ltd. He is currently focused on research and development in the field of Data Mining and Software Engineering. He has completed his Bachelor degree in Computer Science and Engineering in 2013 from American International University Bangladesh (AIUB). Md. Nafees Mahbub is currently involved in software developnemt after completing his Bachelors in 2013 from American International University Bangladesh (AIUB). Alongside with software development, he has a greate interest on research based activities. This article is distributed under the terms of the Creative Commons Attribution License which permits any use, distribution, and reproduction in any medium, provided the original author(s) and the source are credited. Md. Al Imran possessing a diverse knowledge in various technologies related to web platform as well as software development currently giving his full concentration in teaching at the department of Computer Science in American International University Bangladesh (AIUB). He has a keen interest in the research related activities and contributing in the various research labs of AIUB including the Internet and Web Technologies, Data Mining and Data Management Systems. He has completed his 22