A Perceptual and Affective Evaluation of an Affectively-Driven Engine for Video Game Soundtracking

Size: px
Start display at page:

Download "A Perceptual and Affective Evaluation of an Affectively-Driven Engine for Video Game Soundtracking"

Transcription

1 A Perceptual and Affective Evaluation of an Affectively-Driven Engine for Video Game Soundtracking 39 DUNCAN WILLIAMS, JAMIE MEARS, ALEXIS KIRKE AND EDUARDO MIRANDA, Plymouth University IAN DALY, ASAD MALIK, JAMES WEAVER, FAUSTINA HWANG, AND SLAWOMIR NASUTO, University of Reading We report on a player evaluation of a pilot system for dynamic video game soundtrack generation. The system being evaluated generates music using an AI-based algorithmic composition technique to create score in real-time, in response to a continuously varying emotional trajectory dictated by gameplay cues. After a section of gameplay, players rated the system on a Likert scale according to emotional congruence with the narrative, and also according to their perceived immersion with the gameplay. The generated system showed a statistically meaningful and consistent improvement in ratings for emotional congruence, yet with a decrease in perceived immersion, which might be attributed to the marked difference in instrumentation between the generated music, voiced by a solo piano timbre, and the original, fully orchestrated soundtrack. Finally, players rated selected stimuli from the generated soundtrack dataset on a two-dimensional model reflecting perceived valence and arousal. These ratings were compared to the intended emotional descriptor in the meta-data accompanying specific gameplay events. Participant responses suggested strong agreement with the affective correlates, but also a significant amount of interparticipant variability. Individual calibration of the musical feature set, or further adjustment of the musical feature set are therefore suggested as useful avenues for further work. Categories and Subject Descriptors: H.5.5 [Information Interfaces and Presentation]: Sound and Music Computing Methodologies and techniques; H.1.2 [Models and Principles]: User/Machine Systems Human Information Processing; [Pattern Recognition]: Models Statistical General Terms: Design, Algorithms, Testing Additional Key Words and Phrases: Algorithmic composition, affect, music perception, immersion, emotional congruence ACM Reference Format: Williams, D., Kirke, A., Miranda. E., Daly. I., Weaver. J., Roesch. E., Hwang. F., and Nasuto. S., A perceptual and affective evaluation of an affectively-driven engine for video game soundtracking. ACM Comput. In. Entertain. 14, 3, Online (December 2016), 19 pages. URL: This work is supported by the Engineering and Physical Sciences Research Council (EPSRC), under grants EP/J003077/1 and EP/J002135/1. Author s addresses: D. Williams et al, Interdisciplinary Centre for Computer Music Research (ICCMR), The House, Plymouth University, Drake Circus, Plymouth, Devon, PL4 8AA, United Kingdom; I. Daly et al, Brain Embodiment Laboratory, School of Systems Engineering, University of Reading, Whiteknights, PO Box 217, Reading, Berkshire, RG6 6AY, United Kingdom. Permission to make digital or hardcopies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies show this notice on the first page or initial screen of a display along with the full citation. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credits permitted. To copy otherwise, to republish, to post on servers, to redistribute to lists, or to use any component of this work in other works requires prior specific permission and/or a fee. Permissions may be requested from Publications Dept., ACM, Inc., 2 Penn Plaza, Suite 701, New York, NY USA, fax +1 (212) , or permissions@acm.org ACM /2010/03-ART39 $15.00 URL:

2 39:2 D. Williams et al. 1. INTRODUCTION High quality soundtracking has the potential to enhance player immersion in video games [Grimshaw et al. 2008; Lipscomb and Zehnder 2004]. Combining emotionally congruent sound-tracking with game narrative has the potential to create significantly stronger affective responses than either stimulus alone the power of multimodal stimuli on affective response has been shown both anecdotally and scientifically [Camurri et al. 2005]. Game audio requires at least two additional challenges over other sound-for-picture work; firstly, the need to be dynamic (responding to gameplay states) and secondly to be emotionally congruent whilst adapting to non-linear narrative changes [Collins 2007]. Thus, creating congruent sound-tracking for video games is a non-trivial task, as their interactive nature necessitates dynamic and potentially non-linear sound-tracking. This requirement is essentially due to the unpredictable element of player control over the narrative, without which the game would cease to be interactive. This problem has been approached with various solutions. A commonly used solution is to loop a precomposed passage of music until a narrative break, such as the end of a level, death of a player, victory in a battle, and so on. This type of system is illustrated in Figure 1. However, this approach can become repetitive, and potentially irritating to the player if the transition points are not carefully managed, as musical repetition has been shown to have its own impact on the emotional state of the listener [Livingstone et al. 2012]. transition (1) loop transition (2) Beginning node 1 node 2 Coda time Fig. 1. A game soundtrack adapting to play dynamically the middle section is looped until a narrative breakpoint is reached. All passages, and transitions between passages (cadences etc) are precomposed. This approach has been used extensively in a wide range of games (see, for example, the Final Fantasy series). An alternative solution would be to use divergent musical sequences to create a less repetitive soundtrack, as shown in Figure 2. However, the trade-off becomes one of practicality and complexity, as each divergent branch requires both storage space on the physical medium (which is often at a premium in the world of video gaming), and a human cost in the time taken to compose enough material. arousal (a(n)) (a+) (a+) (a-) (a-) (a+) (a-) time (s) Beginning node 1 node 2 Coda Fig. 2. Divergent branching system used to sequence a musical score without looping, determined by arousal cues as meta-tags corresponding to gameplay narrative (arousal increase or decrease at beginning,

3 A Perceptual and Affective Evaluation of an Affectively-Driven Engine for Video Game Soundtracking 39:3 middle, and end). Even for a very simple narrative with just three stages, 7 discrete pieces of music are required, with a resulting impact on composition time and storage space. 1.1 Defining Player Immersion Player immersion is well-understood by the gaming community and as such is a desirable, measurable attribute [Qin et al. 2009]. Some researchers consider that the enduring popularity of videogames is due to the total immersion in a player mediated world [Weibel and Wissmath 2011]. Defining immersion in this context is, however, not trivial. A Player derived definition has been suggested as a sense of the loss of time perception [Sanders and Cairns 2010], but this was refuted in later work [Nordin et al. 2013] as a solely anecdotal definition which requires further experimental analysis in order to consider correlation between time and immersion perception. Some studies had previously attempted to address this in an ecological context [Tobin et al. 2010], i.e. Whilst actually playing, with the suggestion that players fundamentally underestimate the length of time spent if they evaluate the duration prospectively rather than retrospectively, and that the number of hours involved in regular play was a reliable predictor of perceived gameplay duration estimates from the participating players. Time perception is therefore difficult to linearly correlate to perceived immersion. Nordin further suggested that attention might be a better direct correlate of immersion, with time perception being implicated when immersion is experienced as a by-product of attention rather than being a direct correlate itself [Nordin et al. 2013]. The cognitive processes involved in the player considering the experience immersive are likely to be narrative dependent, in other words the music needs to be emotionally congruent with gameplay events [Jørgensen 2008]. This has implications for soundtrack timing, and for emotional matching of soundtrack elements with gameplay narrative [Williams, Kirke, E. R. Miranda, et al. 2015]. If the music is appropriate to the context of the gameplay, is likely that there will be A direct relationship between greater emotional impact and an increase in perceived immersion [Grimshaw et al. 2008; Lipscomb and Zehnder 2004]. Emotional congruence between sound-tracking and gameplay is also a measurable attribute, provided the players evaluating this attribute have a shared understanding of exactly what it is they are being asked to evaluate [Bensa et al. 2005]. Lucasarts implemented a dynamic system, imuse (see [Strank 2013] for a full treatment) to accompany their role-playing games series in the late 1980 s (which included the Indiana Jones series and perhaps most famously, the Monkey Island series of games) [Warren 2003]. This system implemented two now commonplace solutions, horizontal re-sequencing and vertical re-orchestration, both of which were readily implementable due to the use of MIDI orchestration. However, the move towards real audio made many of these transformations more complex beyond the compositional aspect alone. Music has been shown in the past to be able to influence the perceived emotion of images, for example in the exception of happy or sad facial expressions recognition [Aubé et al. 2014; Schubert 2004]. Beyond simply reflecting the existing narrative, music can be used to project particular emotional qualities onto the imagery, helping to sustain engagement, ought to incentivise particular gameplay objectives, potentially leading to engrossment. The ability of the composer to influence player engagement (and encourage the players to continue playing) beyond that of the game designer, by reinforcing narrative through emotionally congruent soundtracking has become an exciting and

4 39:4 D. Williams et al. fertile area for work. Other theoretical approaches to engagement by congruent sound tracking have also been suggested. Listener expectation is a powerful cue for emotions [Huron 2006], and has been related to engagement with videogames by means of schemata confirmation and violation [Douglas and Hargadon 2000]. These mechanisms demonstrate many similarities with the world of sound for film, which may offer some explanation as to how they would be enjoyable to a first time gamer. Examples include fast tempos with action sequences, or when the player is otherwise under pressure. If these preconceptions are violated, some studies find that there is some suggestion that experienced gamers may find this sensation enjoyable on a neurological level [Janata 1995] [Sridharan et al. 2007], as correlates of predictability which might be used to enhance attention and help move a player higher in their level of immersion. Engrossment follows engagement, and might be considered to be the point where the player becomes emotionally invested in the game, beyond simply enjoying it. The player may find interacting with the game no longer requires any cognitive effort (in the same way that inexperienced automobile drivers have to pay attention to the process of driving, but this cognitive load changes over time partly due to time spent behind the wheel) [Arsenault 2005]. Soundtracking at this stage becomes synonymous with the emotional content of the narrative, such that one might influence the other symbiotically [Mar et al. 2011]. Total immersion or flow, the holistic sensation do people feel when the act with total involvement [Csikszentmihalyi 1975] is a state where the player would not only act automatically in controlling the gameplay, but would not necessarily consider themselves playing the game any more but rather actually being in the game. In this state there is A merging of actions and self-awareness, the place full attention will be immersed in the game. Considering the affective potential of music in video games is a useful way of understanding, and potentially enhancing, the player experience of emotion from the gameplay narrative. 1.2 Defining Emotional Congruence Three types of emotional responses are commonly found in music psychology literature: emotion, affect, and mood, though the differences between these are less often explained [Russell and Barrett 1999]. The general trend is for emotions to be short episodes, with moods longer-lived. The literature often makes a strong distinction between perceived and induced emotions (see for example (Västfjäll, 2001; Vuoskoski and Eerola, 2011; Gabrielsson, 2001); though the precise terminology used to differentiate the two also varies, the over-arching difference is whether the emotion is communicated to, or experienced by, the listener. The reader can find exhaustive reviews on the link between music and emotion in [Scherer 2004], which is explored further in the recent special issue of Musciae Scientiae [Lamont and Eerola 2011]. There are a number of emotion models that can be used when approaching emotional responses to musical stimuli, including categorical models, which describe affective responses with discrete labels, and dimensional models, which approach affect as coordinates, often in a two-dimensional space Russell s circumplex model [Russell 1980] provides a way of parameterising affective responses to musical stimuli in two dimensions: valence (a scale of positivity) and arousal (a scale of energy or activation strength), although three dimensional spaces are also common [Eerola and Vuoskoski 2010]. Emotional descriptors from Hevner s adjective cycle

5 A Perceptual and Affective Evaluation of an Affectively-Driven Engine for Video Game Soundtracking 39:5 can be mapped quite closely onto the two-dimensional model [Hevner 1936], in order to create a semantic space. This creates a dimensional-categorical approach [Schubert 1999, p.22], whereby intensely negative states, such as anger or fear would occur at the opposite end of a 2-dimensional space from low-intensity positive states such as calmness or relaxation. Thus, categorical and dimensional approaches are not necessarily irreconcilable, as emotional labels from categorical models can be mapped onto dimensional spaces without any particular difficulty. Both types of models have been used to carry out affective evaluations of music in a large number of studies (Juslin & Sloboda, 2010). Recently, music-specific approaches have been developed, notably in (Zentner, Grandjean, & Scherer, 2008), where the Geneva Emotion Music Scale (GEMS) describes nine dimensions covering a complete semantic space of musically evoked emotions. The relative importance to the gamer of immersion and emotional congruence is not necessarily evenly weighted. Immersion is essential, or the player will likely cease playing the game; keeping the player hooked is an important goal of game design. Emotional congruence, on the other hand, may enhance player immersion, but is likely to hold a lower place in the player s perceptual hierarchy. One notable exception might be in situations where the player deliberately controls the music (as can be seen in games like Guitar Hero, for example). The process then becomes a complex feedback loop wherein the player not only influences the selection of music according to their mood but the selection of music also has a subsequent impact on the player s mood. 1.3 Defining Affectively-Driven Algorithmic Composition Affectively-driven algorithmic composition (AAC) is an emerging field combining computer music research and perceptual/psychological approaches to music cognition [Mattek 2011; Williams et al. 2013; Williams et al. 2014]. AAC systems attempt to communicate specific emotions to the listener. A disadvantage to looping-based approaches for gameplay soundtracking is the high amount of repetition involved. This can become distracting or worse, irritating, at transition points, which can have a knock-on negative effect on player immersion. The resolution of looping systems can be improved by adding divergent score branches at narrative breakpoints within the soundtrack, which results in more complex, less repetitive musical sequences. However, the need to create the contributory score fragments in such a manner that they can be interchanged whilst maintaining the intended aesthetic-congruency with the narrative poses a significant challenge to the video-game composer. In simple terms, the over-arching challenge is that video game music can get repetitive and thereby ruin player immersion, but composing large quantities of music with particular moods and emotions is not practical for most games, both in terms of storage on the media (whether that be disc, cartridge, or simply bandwidth in, for example, online streaming games), and also in terms of human cost (i.e., that of the composers time when constructing large numbers of interchangeable musical sequences). Thus, the adaptability of a branching system to emotional responses for these purposes is somewhat compromised. This paper addresses such challenges by considering an AAC system for creation of music on the fly. AAC creation of music in this manner has the potential to overcome some of these restrictions, yet many systems for algorithmic composition

6 39:6 D. Williams et al. fail to directly target emotional responses [Williams et al. 2013], and thus the necessary narrative congruency demanded by game soundtracking might still be compromised by such systems. In this prototype we evaluate a system incorporating a range of musical features with known affective correlates; tempo, mode, pitch range, timbre, and amplitude envelope. For computational efficiency the value range of each of these musical features was discretized to 3 possible levels resulting in a generative ruleset of an AAC system defined over discretized 3x3 partition of a 2-D affective space based on the circumplex model [Russell 1980]. The specification of the system under evaluation in this paper, which attempts to tackle these challenges, has been previously presented as a proof-of-concept [Williams et al. 2015], but has not yet been the subject of a perceptual or affective evaluation. 2. METHOD The AAC pilot described here uses a transformative algorithm based on a second order Markov-model [Ames 1989] with a musical feature matrix that allows for discrete control over five musical parameters in order to imply various affective descriptors in a categorical/dimensional model loosely arranged over two dimensions, after the circumplex model of affect [Russell 1980]. 2.1 Musical Structure Representation, Analysis, and Generation Markov generation of musical structures has been frequently described in the literature (see, for example, [Ames 1989; Visell 2004; Papadopoulos and Wiggins 1999]. The second-order markov model used here is defined over the finite space of 5 musical features mentioned above. The model consists of a transition probability matrix, with each row corresponding to a conditional probability vector defined over the array of possible next states (musical feature 5-tuples),given the last two states. The model, i.e. the entries of the state transition probability matrix, is learned from the musical training material. The Markov model is generative, i.e. once it is learned, it can be used to create new state (musical feature 5-tuples) sequences according to the likelihood of a particular state occurring after the current and proceeding states. The generated musical state sequences are subsequently further transformed according to the distance between the current features and the features which correlate to a given affective target (the transformations indicating the affective correlates are shown in Table 1). The transformed data are then synthesized using a piano timbre. The system can create a surprisingly large variety of material from a very limited amount of seed data. The input material used to train the system in the case of this pilot study was twelve monophonic bars of a Mozart piano concerto in the key of C major. Material can be can be generated so quickly it could in future be used to create score and performance data in pseudo real-time (in normal operation, listeners would not be aware of any noticeable latency between generation and triggering of synthesized or sampled audio), but in this case, a range of pre-rendered sequences were produced as amplitude normalized PCM wave files. For each of the nine affective correlates (see section 2.2, Figure 3), 11 sequences were generated, each one-minute in duration. During gameplay, these files are cross-faded to create continuous soundtracking either within a single affective co-ordinate or across a variety depending on the gameplay cues. An overview of the generation and transformation process is shown in Figure 3.

7 A Perceptual and Affective Evaluation of an Affectively-Driven Engine for Video Game Soundtracking 39:7 Start 2nd order Markov probability matrix (from training material) Generate notes generated note sequence Target affective state {a, v} (according to gameplay) determine selected feature transformations affective correlate value(s) of target affective correlate matrix (mode, tempo, timbre, melodic and dynamic range) Calculate difference between generated sequence correlate values and target correlate values distance from current affective co-ordinate to target INTERPOLATE musical features according to distance between generated affective correlate values and target affective correlate values transformed feature set SYNTHESISE audio output end Fig. 3. Generative process. A Markov Chain is used to generate new material, which is then transformed according to the distance between the current set of musical features and the target musical features which correlate to a given point in the emotion space. Note that this flow is applied sequentially to each generated Markov state (5-tuple of musical features) at a time without the need for a the sequence of transformation there is a generated musical features state with its implied affective counterpart and the target affective state and the transformation thus adjust the features to meet that target affective state.

8 39:8 D. Williams et al. Broader musical structure, including thematic variation and repetition, is not addressed in this pilot system beyond the matching of a single emotional trajectory according to the player character. Thus, this evaluation should not be generalized beyond MMORPG games, and in future would be best expanded to an evaluation across a wide range of games for each participant. Moreover, the possible positive influence of repetition on player immersion should not be discounted [Pichlmair and Kayali 2007; Lipscomb and Zehnder 2004]. Structural composition techniques remain challenging for all automatic composition systems [Edwards 2011], and as such present a fertile area for continued investigation beyond the scope of the work presented here. 2.2 Affective Model The algorithmic composition system references affective correlates according to the narrative of the gameplay to derive an affective target for the generated music. The system uses a combined dimensional and categorical approach for this affective target. The two-dimensional circumplex model of affect (two dimensions: valence (a scale of positivity on the horizontal axis and arousal as a scale of energy or activation strength on the vertical axis),is divided into 9 sectors that are indexed with metatags corresponding to Cartesian co-ordinate values, each a discrete affective descriptor, as shown in Figure 4. Thus, a range of basic affective descriptors are represented across the sectors of this model, with Euclidean distances for lower and higher arousal levels vertically across the affective space (though as with any affective descriptor, some degree of perceptual overlap is present amongst descriptors, see for example, pleased and happy, or pleased and content such descriptors cannot be universally discrete). In this manner, a co-ordinate of { v3, a3 } would refer to excited. arousal angry pleased excited sad content happy tired frustrated calm valence Fig. 4. Two-dimensional model divided into 9 sectors with discrete affective adjectives mapped to the circumplex model (angry, sad, tired, pleased, content, frustrated, excited, happy and calm).

9 A Perceptual and Affective Evaluation of an Affectively-Driven Engine for Video Game Soundtracking 39:9 Table I. Showing affective correlates, and the corresponding musical parameter mappings used by the generative model Descriptor Tempo range Mode Timbre range Mean melodic range Dynamic range (loudness) Tired Slow Minor Dark High Low Sad Medium Minor Medium Medium Medium Angry Fast Minor Bright Low High Frustrated Slow Chromatic Dark Low Low Content Medium Chromatic Medium Medium Medium Pleased Fast Chromatic Bright High High Calm Slow Major Dark Low Low Happy Medium Major Medium Medium Medium Excited Fast Major Bright High High A quest (section of gameplay) from World of Warcraft (a massively multiplayer online role-playing game, MMORPG), was marked up with various affective targets as meta-tags (for example, fighting scenes were tagged with {v1, a3}, or angry). Two screen-shots illustrating the application of this meta-tagging are shown in Figures 5 and 6, where a battle scene is accompanied by angry music until the player is victorious, at which point, content is cued. Stimuli corresponding to the affective meta-tag were selected randomly from the pre-generated stimulus pool (for each affective co-ordinate the pool contained 11 audiofiles each one minute in duration) during gameplay via a dedicated audio playback engine built using Max/MSP. The engine selects one of the 11 stimuli randomly, cross-fading with the next stimulus choice until the entire pool has been used (or until a new meta-tag trigger is received), using a unique random number function. Timbral changes in the generated soundtrack were created by means of piano dynamics (loud performances resulting in harder and brighter timbres, with more pronounced upper harmonics and a correspondingly higher spectral centroid). Participants were then asked to complete the quest three times; once with the original musical soundtrack, once with a soundtrack provided by the algorithmic composition system, and once with no musical accompaniment at all (sound effects were still used, for example, action sounds). Each playback was recorded so that the stimulus selections could be repeated for subsequent affective evaluation.

10 39:10 D. Williams et al. Fig. 5. Screen capture of a sequence of gameplay from a generative-soundtrack playthrough, marked up with { v1, a3 } or angry, which accompanies a battle scene (the process of entering battle cues the change in affective target). Fig. 6. Screen capture of a sequence of gameplay from a generative-soundtrack playthrough, marked up with { v2, a2 }, or content, triggered after the player had successfully completed the battle sequence above. Participants were asked to rate emotional congruence and immersion after each playthrough, using a 11-point Likert scale presented via an interactive webbrowser form, as shown in Figure 7. Short definitions for both terms were included in the pre-experiment instructions for participants.

11 A Perceptual and Affective Evaluation of an Affectively-Driven Engine for Video Game Soundtracking 39:11 Fig. 7. Listener interface for evaluating emotional congruence of music and immersion for each playthrough. Having evaluated emotional congruence and immersion for each of the musical playthroughs, participants were also asked to rate the perceived emotion of each stimulus that they had been exposed to in the generative soundtrack playthrough, using a two-dimensional space labeled with the self-assessment manikin [Bradley and Lang 1994] showing valence on the horizontal scale and arousal on the vertical scale, allowing both valence and arousal to be estimated in a single rating. Valence was defined to the participants as a measure of positivity, and arousal as a measure of activation strength in a pre-experiment familiarization and training stage. This participant interface was also implemented in Max/MSP. In total, 11 participants took part in the experiment, 6 males and 5 females. 9 participants were ages 18-21, while the remaining 2 participants were aged In a pre-evaluation questionnaire, the majority of the participants reported that they enjoyed listening to music, while only 45% of them had experience of composing or performing music. 72% of participants reported that they enjoyed playing or watching video games, although 45% of participants answered that they only spend 0.5 hours a week watching or playing video games. Participants undertook the experiment using circumaural headphones in a quiet room with a dry acoustic. Gameplay was presented on a 15 laptop computer. The game was not connected to the internet (i.e., there were no other playercharacters, all characters apart from the avatar were computer controlled). The exact duration of playthroughs varied on a player-by-player basis, from 4 minutes to a cut off of 10 minutes. A variety of affective states may be entered in each playthrough, depending on the actions of the player, for example exploration, fighting, fighting and being victorious, fighting and failure, evading danger, and interacting positively with other (non-player) character avatars. 3. RESULTS The Likert scale responses for emotional congruence and immersion are shown in Table 2. The mean emotional congruence is improved by 1, with a standard deviation of 0.7 and a p-value of less than This strongly suggests that listeners found the generated soundtrack more emotionally congruent with the gameplay than that of the original soundtrack, and that this marked improvement was both consistent across all participants, and statistically significant. This is a promising result for the

12 39:12 D. Williams et al. AAC system. However, the mean immersion decreased by in the ratings of the generated soundtrack playthrough in comparison to the original soundtrack playthrough, with a p-value still below the threshold of significance (0.05), at Despite the high accompanying standard deviation, at 2.1 for the generated soundtrack immersion, this suggests that player immersion was consistently reduced in the generated soundtrack playthrough. Overall, these results suggest that an increase in emotional congruence of ~10% can be achieved by the current AAC prototype system, at the expense of a ~30% reduction in immersion. Table 2. Likert-scale responses showing participant reactions to playthroughs with the original soundtrack and the generated soundtrack. Original soundtrack emotional congruence Original soundtrack gameplay immersion Generated soundtrack emotional congruence Generated soundtrack gameplay immersion Mean Standard p-value Deviation Fig. 8. From left to right, mean participant ratings for original soundtrack emotional congruence, original soundtrack gameplay immersion, generated soundtrack emotional congruence, and generated soundtrack gameplay immersion. Error bars indicate 95% confidence interval for each mean. The mean ratings for emotional congruence and immersion in playthroughs with the original soundtrack and the generated soundtrack as shown in Figure 8,

13 A Perceptual and Affective Evaluation of an Affectively-Driven Engine for Video Game Soundtracking 39:13 suggest that although the means are close, the difference between original and generated ratings in emotional congruence and immersion is significant (a clear line can be drawn horizontally from the top of the error bar for original soundtrack emotional congruence below that of the bottom of the error bar for ratings of the generated soundtrack emotional congruence, similarly this can be drawn from the bottom of the CI of original soundtrack immersion to well above the top of the CI of the ratings for generated soundtrack immersion). This suggests that the improvement between emotional congruence in the original soundtrack and the generated soundtrack is consistent. Immersion is generally consistently lower in the generated soundtrack results, though this exhibits the largest variance and consequently the highest standard deviation. However, the sample size is small and these results should therefore be interpreted with some caution. The smallest variation occurs in the ratings for generated soundtrack emotional congruency, which suggest that even though the number of participants was small, removing additional participants would not change the outcome drastically (and thus, including additional participants, from the same small demographic, would likely not significantly influence the outcome of this pilot study). If we compute power after Rosner [Rosner 2010] allowing 5% Type I error rate, using the standard deviations and means reported in Table 2, we find that for the original soundtrack emotional congruence we can be confident of 70% power with 10 participants at the reported standard deviation of 1.188, and for the generated soundtrack emotional congruence we can be confident of 70% power with 10 participants at the reported standard deviation of (both values would require 13 participants to achieve 80% power at the reported standard deviation values). Using the same measures for original soundtrack immersion we find 80% power with 11 participants at the reported standard deviation of 0.916, and 80% power with 11 participants at the reported standard deviation of for generated soundtrack immersion. This suggests reasonable statistical significance despite the small sample size, especially considering the limited amount of variability between participants (who fall closely within the target demographic of this type of video game and shared an agreement in the amount of gameplay they undertook each week). The infancy of the field at the time of conducting these experiments means that there are few precedents for appropriate evaluation paradigms, and to test for normality would take hundreds of trials, which might also have a knock-on effect on the participants involved (repeatedly undertaking the same section of gameplay might have an impact on immersion for example). However, the system under evaluation here has previously been the subject of a number of other trials, including discrete musical feature evaluation [Daly et al. 2014; Williams, Kirke, E. Miranda, et al. 2015] and broader two-dimensional affective evaluation by self-report and biophysiological measurement [Daly et al. 2015; Williams, Kirke, J. Eaton, et al. 2015]. Thus, whilst we still consider the small sample size a caveat, we consider both the low p values and the comparatively high power encouraging for further work with a more fully realized system and larger numbers of participants in the future. The spread of ratings for emotional congruence between the original soundtrack and the generated soundtrack playthroughs is illustrated in Figures 9 and 10. Participants were not asked to rate the emotional congruence of the soundtrack in the silent playthrough. However, 10 out of 11 participants rated the silent playthrough as less immersive than the two playthroughs that included a musical soundtrack.

14 39:14 D. Williams et al. Fig. 9. The spread of ratings for emotional congruence responses to original soundtrack playthroughs. Fig. 10. The spread of ratings for emotional congruence responses to generated soundtrack playthroughs. 3.1 Listener Agreement with Intended Emotional Meta-tag Mean listener agreement between perceived emotion and the intended emotional meta-tag was ~85%, with a standard deviation of 4.9 and a p-value of Full results are shown in Table 3.

15 A Perceptual and Affective Evaluation of an Affectively-Driven Engine for Video Game Soundtracking 39:15 Table 3. Listener agreement with emotional correlate meta-tag for stimuli from each sector (shown to maximum of 3 decimal places). Note the full stimulus set was not evaluated by each listener, only stimuli which were present during the playthrough were repeated for evaluation in the two-dimensional space. Stimulus Label Mean agreement with target Std. Deviation Std. Error of Mean v1, a1 v2, a1 v3, a1 v1, a2 v2, a2 v3, a2 v1, a3 v2, a2 v3, a3 Mean Std. dev Mean listener agreement, as shown in Table 3, was markedly high in stimuli across all sectors, from 76.5% for { v1, a2 } to 90.1% for { v1, a1 }. Standard deviation and error of mean typically decreased with higher mean agreement (with the exception of { v3, a1 }. Initially, this suggests that the musical feature mapping used in the generative system, which derives affective correlates from literature and uses them as features to inform the generation of affectively-driven music, was operating as intended, and that the AAC system was thereby able to generate music correctly according to a specific affective target. However, whilst the overall standard deviation is relatively low, the p-value does not reach a significance threshold of <0.05, which suggests that the range of individual emotions from each participant is still unpredictable. This suggests a strong argument that adapting this type of system to individual responses would be a useful avenue for further work, for example by calibrating the musical feature-set from the generative algorithm to each individual player before commencing, or by using bio-feedback data (e.g., from a brain cap, heart-rate monitor or other biosensor) to calibrate the musical feature-set on a case-by-case basis to attempt to reduce inter-participant variability. The emerging field of brain-computer musical interfacing (BCMI) is steadily making inroads to this type of adaptive control for music (systems which combine other biosensors with brain-control are sometimes referred to as hybrid-bcmi) [Miranda et al. 2011]. 4. DISCUSSION The generated soundtrack was performed via a single synthesized piano timbre, which is in a sharp contrast to that of the original soundtrack, which consisted of a fully orchestrated symphonic piece. Nevertheless, participants seemed to consistently find that the generated music matched the emotional trajectory of the video gameplay more congruently than the original musical soundtrack. The majority of participants also reported that the playthroughs which were accompanied by a musical soundtrack were more immersive than the silent playthrough (which featured sound effects but no musical accompaniment). However, there was also a marked decrease in reported immersion with the gameplay in the generated soundtrack playthroughs, perhaps because of the lack of familiarity and repetition in the generated soundtrack. This could be usefully addressed in further work by evaluating repetition of themes and with generated music across multiple different games with the same player in future. Another possible explanation for this increase in emotional congruence but decrease in immersion might well be the orchestration of the AAC system, (solo piano). The original soundtrack might be lacking in p- val

16 39:16 D. Williams et al. emotional variation, but it offers a fullness and depth of instrumentation which is not easily matched by a single instrumental voice. This is a challenge which could be evaluated in future work by creating a piano reduction of the original gameplay score and subjecting it to subsequent evaluation, or by developing the generative system further such that it can create more fully realized pieces of music using multiple instrumental timbres (this alone presents a significant challenge to algorithmic composition). Call of Duty: Modern Warfare features 17 unique pieces of music, around 52:31 total duration, using 531mb of storage. Players might spend upwards of 100 hours playing the game, and are therefore likely to have heard each piece of music many times over. If a system like the AAC pilot evaluated here could be expanded upon to sustain or improve player immersion as well as improve emotional congruence, the benefits to the video game world would not simply be limited to a reduction in workload for composers, or to less repetitive sound-tracking for players the amount of data storage required for soundtrack storage might also be significantly reduced (a stereo CD-quality PCM wave file takes approximately 10mb of storage per minute of recorded audio). Whilst data space is not as scarce a commodity in the modern world as it might have been in the days of game storage on cartridge or floppy disk, gaming is now increasingly moving onto mobile platforms (phones, tablets) with limited storage space, and online streaming is also a popular delivery platform for gaming. Therefore, a reduction in data by using AAC for soundtracking could represent a significant and valuable contribution to video game delivery in the future. Some participants anecdotally reported that the generated music seemed more responsive to the gameplay narrative, and therefore added an extra dimension to the gameplay (in other words, they felt in control of the music generation system). This suggests that this type of system might have possible applications as a composition engine for pedagogic, or even therapeutic applications beyond the world of entertainment. Beyond our existing caveats regarding the timbral variation this pilot can generate, and the relatively small number of participants who have been involved in this specific evaluation, we also acknowledge that a future test paradigm might usefully explore deliberately incongruous music generation (perhaps playing happy music with scenes of low valence). This type of testing would have implications on the number of trial iterations required, and the trial time that participants were required for (at present, the maximum test time was approximately 30 minutes, after which point listener fatigue might be considered a factor). Our hypothesis would be that emotional congruence, and likely immersion, would go down in such a case, but this has not yet been tested and remains an avenue we intend to explore in further work. There also remains significant useful further work in training the generator with a larger range of input material, and in testing a real-time system so that pregenerated sequences did not need to be cross-faded as in the current paradigm (more complex timbres might create difficulties with simple cross-fades, which the current piano timbre does not seem to be particularly disrupted by).

17 A Perceptual and Affective Evaluation of an Affectively-Driven Engine for Video Game Soundtracking 39:17 5. CONCLUSIONS Using an AAC system to create music according to emotional meta-tagging as part of a video game narrative has clear practical benefits in terms of composer time and hardware/software requirements (file storage and data transmission rates), which could free up processing and storage space for other game elements (e.g., visual processing). This type of system might also therefore be beneficial to mobile gaming platforms (smart phones etc.), where space is at more of a premium than on desktop or home gaming environments. Within the constraints of the test paradigm, this pilot study suggests that emotional congruence could be improved when participants played with a soundtrack generated by the affectively-driven algorithmic composition system. However, player immersion was consistently and significantly reduced at the same time. It might be possible to seek explanation of this in the instrumentation and orchestration of the generated music, but further work would be required to establish the reason for this reported reduction in player immersion, before tackling the problem of developing an AAC system with fuller orchestration, which is in and of itself non-trivial. These algorithmic composition techniques are still in their infancy and the likelihood of replacing a human composer in the successful creation of complex, affectivelycharged musical arrangements is minimal. In fact, as the system presented here (and others like it) require training with musical input, this evaluation suggests that in the future composers working explicitly with video game sound-tracking might use this type of system to generate large pools of material from specific themes, thereby freeing up time to spend on the creative part of the composition process. Participant agreement with the affective meta-tagging used to select musical features as part of the generative system was good, though significant interparticipant variability suggested that either the musical feature set needs further calibration (which would require specific affective experiments), or that a generalized set of affective correlates as musical feature sets is not yet possible. Another solution might be to calibrate this type of generative music system to the individual, using a mapping of musical features documented here in order to attempt to target specific emotional responses in the generated soundtrack. In the future, this could be driven by bio-sensors such as the electroencephalogram (as in the emerging field of braincomputer music interfacing), or by more traditional biosensors such as heart rate sensors or galvanic skin response. ACKNOWLEDGMENTS The authors gratefully acknowledge the support of EPRSC grants EP/J003077/1 and EP/J002135/1.

18 39:18 D. Williams et al. REFERENCES Charles Ames The Markov process as a compositional model: a survey and tutorial. Leonardo (1989), Dominic Arsenault Dark waters: Spotlight on immersion. In Game-On North America 2005 Conference Proceedings William Aubé, Arafat Angulo-Perkins, Isabelle Peretz, Luis Concha, and Jorge L. Armony Fear across the senses: brain responses to music, vocalizations and facial expressions. Soc. Cogn. Affect. Neurosci. (2014), nsu067. Julien Bensa, Danièle Dubois, Richard Kronland-Martinet, and Sølvi Ystad Perceptive and cognitive evaluation of a piano synthesis model. In Computer music modeling and retrieval. Springer, Margaret M. Bradley and Peter J. Lang Measuring emotion: the self-assessment manikin and the semantic differential. J. Behav. Ther. Exp. Psychiatry 25, 1 (1994), A. Camurri, G. Volpe, G. De Poli, and M. Leman Communicating expressiveness and affect in multimodal interactive systems. Multimed. IEEE 12, 1 (2005), Karen Collins Collins, Karen An Introduction to the Participatory and Non-Linear Aspects of Video Games Audio. Eds. Stan Hawkins and John Richardson. Essays on Sound and Vision. Helsinki: Helsinki University Press. pp (2007). Mihaly Csikszentmihalyi Play and intrinsic rewards. J. Humanist. Psychol. (1975). Ian Daly et al Brain-computer music interfacing for continuous control of musical tempo. (2014). Ian Daly et al Towards human-computer music interaction: Evaluation of an affectively-driven music generator via galvanic skin response measures. In Computer Science and Electronic Engineering Conference (CEEC), th. IEEE, Yellowlees Douglas and Andrew Hargadon The pleasure principle: immersion, engagement, flow. In Proceedings of the eleventh ACM on Hypertext and hypermedia. ACM, Michael Edwards Algorithmic composition: computational thinking in music. Commun. ACM 54, 7 (2011), T. Eerola and J.K. Vuoskoski A comparison of the discrete and dimensional models of emotion in music. Psychol. Music 39, 1 (August 2010), DOI: Alf Gabrielsson Emotion perceived and emotion felt: Same or different? Music. Sci. Spec Issue, (2001), Mark Grimshaw, Craig A. Lindley, and Lennart Nacke Sound and immersion in the first-person shooter: mixed measurement of the player s sonic experience. In Proceedings of Audio Mostly Conference K. Hevner Experimental studies of the elements of expression in music. Am. J. Psychol. 48, 2 (1936), David Brian Huron Sweet anticipation: Music and the psychology of expectation, MIT press. Petr Janata ERP measures assay the degree of expectancy violation of harmonic contexts in music. J. Cogn. Neurosci. 7, 2 (1995), Kristine Jørgensen Left in the dark: playing computer games with the sound turned off, Ashgate. Patrik N. Juslin and John A. Sloboda Handbook of music and emotion : theory, research, applications, Oxford: Oxford University Press. A. Lamont and T. Eerola Music and emotion: Themes and development. Music. Sci. 15, 2 (July 2011), DOI: Scott D. Lipscomb and Sean M. Zehnder Immersion in the virtual environment: The effect of a musical score on the video gaming experience. J. Physiol. Anthropol. Appl. Human Sci. 23, 6 (2004), Steven R. Livingstone, Caroline Palmer, and Emery Schubert Emotional response to musical repetition. Emotion 12, 3 (2012), DOI: Raymond A. Mar, Keith Oatley, Maja Djikic, and Justin Mullin Emotion and narrative fiction: Interactive influences before, during, and after reading. Cogn. Emot. 25, 5 (2011), Alison Mattek Emotional Communication in Computer Generated Music: Experimenting with Affective Algorithms. In Proceedings of the 26th Annual Conference of the Society for Electro-Acoustic Music in the United States. Miami, Florida: University of Miami Frost School of Music. Eduardo R. Miranda, Wendy L. Magee, John J. Wilson, Joel Eaton, and Ramaswamy Palaniappan Brain-computer music interfacing (BCMI) from basic research to the real world of special needs. Music Med. 3, 3 (2011), A. Imran Nordin, Jaron Ali, Aishat Animashaun, Josh Asch, Josh Adams, and Paul Cairns Attention, time perception and immersion in games. In CHI 13 Extended Abstracts on Human Factors in Computing Systems. ACM, G. Papadopoulos and G. Wiggins AI methods for algorithmic composition: A survey, a critical view and future prospects. In AISB Symposium on Musical Creativity

TOWARDS AFFECTIVE ALGORITHMIC COMPOSITION

TOWARDS AFFECTIVE ALGORITHMIC COMPOSITION TOWARDS AFFECTIVE ALGORITHMIC COMPOSITION Duncan Williams *, Alexis Kirke *, Eduardo Reck Miranda *, Etienne B. Roesch, Slawomir J. Nasuto * Interdisciplinary Centre for Computer Music Research, Plymouth

More information

Investigating Perceived Emotional Correlates of Rhythmic Density in Algorithmic Music Composition

Investigating Perceived Emotional Correlates of Rhythmic Density in Algorithmic Music Composition Investigating Perceived Emotional Correlates of Rhythmic Density in Algorithmic Music Composition 1 DUNCAN WILLIAMS, ALEXIS KIRKE AND EDUARDO MIRANDA, Plymouth University IAN DALY, JAMES HALLOWELL, JAMES

More information

THE EFFECT OF EXPERTISE IN EVALUATING EMOTIONS IN MUSIC

THE EFFECT OF EXPERTISE IN EVALUATING EMOTIONS IN MUSIC THE EFFECT OF EXPERTISE IN EVALUATING EMOTIONS IN MUSIC Fabio Morreale, Raul Masu, Antonella De Angeli, Patrizio Fava Department of Information Engineering and Computer Science, University Of Trento, Italy

More information

Expressive information

Expressive information Expressive information 1. Emotions 2. Laban Effort space (gestures) 3. Kinestetic space (music performance) 4. Performance worm 5. Action based metaphor 1 Motivations " In human communication, two channels

More information

Subjective Emotional Responses to Musical Structure, Expression and Timbre Features: A Synthetic Approach

Subjective Emotional Responses to Musical Structure, Expression and Timbre Features: A Synthetic Approach Subjective Emotional Responses to Musical Structure, Expression and Timbre Features: A Synthetic Approach Sylvain Le Groux 1, Paul F.M.J. Verschure 1,2 1 SPECS, Universitat Pompeu Fabra 2 ICREA, Barcelona

More information

1. BACKGROUND AND AIMS

1. BACKGROUND AND AIMS THE EFFECT OF TEMPO ON PERCEIVED EMOTION Stefanie Acevedo, Christopher Lettie, Greta Parnes, Andrew Schartmann Yale University, Cognition of Musical Rhythm, Virtual Lab 1. BACKGROUND AND AIMS 1.1 Introduction

More information

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene

However, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene Beat Extraction from Expressive Musical Performances Simon Dixon, Werner Goebl and Emilios Cambouropoulos Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria.

More information

The Space Between Us: Evaluating a multi-user affective braincomputer

The Space Between Us: Evaluating a multi-user affective braincomputer The Space Between Us: Evaluating a multi-user affective braincomputer music interface Joel Eaton, Duncan Williams, Eduardo Miranda Interdisciplinary Centre for Computer Music Research, Plymouth University,

More information

Affective response to a set of new musical stimuli W. Trey Hill & Jack A. Palmer Psychological Reports, 106,

Affective response to a set of new musical stimuli W. Trey Hill & Jack A. Palmer Psychological Reports, 106, Hill & Palmer (2010) 1 Affective response to a set of new musical stimuli W. Trey Hill & Jack A. Palmer Psychological Reports, 106, 581-588 2010 This is an author s copy of the manuscript published in

More information

MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC

MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC Lena Quinto, William Forde Thompson, Felicity Louise Keating Psychology, Macquarie University, Australia lena.quinto@mq.edu.au Abstract Many

More information

Compose yourself: The Emotional Influence of Music

Compose yourself: The Emotional Influence of Music 1 Dr Hauke Egermann Director of York Music Psychology Group (YMPG) Music Science and Technology Research Cluster University of York hauke.egermann@york.ac.uk www.mstrcyork.org/ympg Compose yourself: The

More information

Research & Development. White Paper WHP 228. Musical Moods: A Mass Participation Experiment for the Affective Classification of Music

Research & Development. White Paper WHP 228. Musical Moods: A Mass Participation Experiment for the Affective Classification of Music Research & Development White Paper WHP 228 May 2012 Musical Moods: A Mass Participation Experiment for the Affective Classification of Music Sam Davies (BBC) Penelope Allen (BBC) Mark Mann (BBC) Trevor

More information

The Human Features of Music.

The Human Features of Music. The Human Features of Music. Bachelor Thesis Artificial Intelligence, Social Studies, Radboud University Nijmegen Chris Kemper, s4359410 Supervisor: Makiko Sadakata Artificial Intelligence, Social Studies,

More information

The relationship between properties of music and elicited emotions

The relationship between properties of music and elicited emotions The relationship between properties of music and elicited emotions Agnieszka Mensfelt Institute of Computing Science Poznan University of Technology, Poland December 5, 2017 1 / 19 Outline 1 Music and

More information

SHORT TERM PITCH MEMORY IN WESTERN vs. OTHER EQUAL TEMPERAMENT TUNING SYSTEMS

SHORT TERM PITCH MEMORY IN WESTERN vs. OTHER EQUAL TEMPERAMENT TUNING SYSTEMS SHORT TERM PITCH MEMORY IN WESTERN vs. OTHER EQUAL TEMPERAMENT TUNING SYSTEMS Areti Andreopoulou Music and Audio Research Laboratory New York University, New York, USA aa1510@nyu.edu Morwaread Farbood

More information

Expressive performance in music: Mapping acoustic cues onto facial expressions

Expressive performance in music: Mapping acoustic cues onto facial expressions International Symposium on Performance Science ISBN 978-94-90306-02-1 The Author 2011, Published by the AEC All rights reserved Expressive performance in music: Mapping acoustic cues onto facial expressions

More information

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE Copyright SFA - InterNoise 2000 1 inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering 27-30 August 2000, Nice, FRANCE I-INCE Classification: 7.9 THE FUTURE OF SOUND

More information

Analysis of local and global timing and pitch change in ordinary

Analysis of local and global timing and pitch change in ordinary Alma Mater Studiorum University of Bologna, August -6 6 Analysis of local and global timing and pitch change in ordinary melodies Roger Watt Dept. of Psychology, University of Stirling, Scotland r.j.watt@stirling.ac.uk

More information

PLANE TESSELATION WITH MUSICAL-SCALE TILES AND BIDIMENSIONAL AUTOMATIC COMPOSITION

PLANE TESSELATION WITH MUSICAL-SCALE TILES AND BIDIMENSIONAL AUTOMATIC COMPOSITION PLANE TESSELATION WITH MUSICAL-SCALE TILES AND BIDIMENSIONAL AUTOMATIC COMPOSITION ABSTRACT We present a method for arranging the notes of certain musical scales (pentatonic, heptatonic, Blues Minor and

More information

MUSICAL MOODS: A MASS PARTICIPATION EXPERIMENT FOR AFFECTIVE CLASSIFICATION OF MUSIC

MUSICAL MOODS: A MASS PARTICIPATION EXPERIMENT FOR AFFECTIVE CLASSIFICATION OF MUSIC 12th International Society for Music Information Retrieval Conference (ISMIR 2011) MUSICAL MOODS: A MASS PARTICIPATION EXPERIMENT FOR AFFECTIVE CLASSIFICATION OF MUSIC Sam Davies, Penelope Allen, Mark

More information

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes

DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring 2009 Week 6 Class Notes Pitch Perception Introduction Pitch may be described as that attribute of auditory sensation in terms

More information

Quarterly Progress and Status Report. Perception of just noticeable time displacement of a tone presented in a metrical sequence at different tempos

Quarterly Progress and Status Report. Perception of just noticeable time displacement of a tone presented in a metrical sequence at different tempos Dept. for Speech, Music and Hearing Quarterly Progress and Status Report Perception of just noticeable time displacement of a tone presented in a metrical sequence at different tempos Friberg, A. and Sundberg,

More information

THE INTERACTION BETWEEN MELODIC PITCH CONTENT AND RHYTHMIC PERCEPTION. Gideon Broshy, Leah Latterner and Kevin Sherwin

THE INTERACTION BETWEEN MELODIC PITCH CONTENT AND RHYTHMIC PERCEPTION. Gideon Broshy, Leah Latterner and Kevin Sherwin THE INTERACTION BETWEEN MELODIC PITCH CONTENT AND RHYTHMIC PERCEPTION. BACKGROUND AND AIMS [Leah Latterner]. Introduction Gideon Broshy, Leah Latterner and Kevin Sherwin Yale University, Cognition of Musical

More information

BRAIN-ACTIVITY-DRIVEN REAL-TIME MUSIC EMOTIVE CONTROL

BRAIN-ACTIVITY-DRIVEN REAL-TIME MUSIC EMOTIVE CONTROL BRAIN-ACTIVITY-DRIVEN REAL-TIME MUSIC EMOTIVE CONTROL Sergio Giraldo, Rafael Ramirez Music Technology Group Universitat Pompeu Fabra, Barcelona, Spain sergio.giraldo@upf.edu Abstract Active music listening

More information

Electronic Musicological Review

Electronic Musicological Review Electronic Musicological Review Volume IX - October 2005 home. about. editors. issues. submissions. pdf version The facial and vocal expression in singers: a cognitive feedback study for improving emotional

More information

Opening musical creativity to non-musicians

Opening musical creativity to non-musicians Opening musical creativity to non-musicians Fabio Morreale Experiential Music Lab Department of Information Engineering and Computer Science University of Trento, Italy Abstract. This paper gives an overview

More information

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS

POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS Andrew N. Robertson, Mark D. Plumbley Centre for Digital Music

More information

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance Methodologies for Expressiveness Modeling of and for Music Performance by Giovanni De Poli Center of Computational Sonology, Department of Information Engineering, University of Padova, Padova, Italy About

More information

MOTIVATION AGENDA MUSIC, EMOTION, AND TIMBRE CHARACTERIZING THE EMOTION OF INDIVIDUAL PIANO AND OTHER MUSICAL INSTRUMENT SOUNDS

MOTIVATION AGENDA MUSIC, EMOTION, AND TIMBRE CHARACTERIZING THE EMOTION OF INDIVIDUAL PIANO AND OTHER MUSICAL INSTRUMENT SOUNDS MOTIVATION Thank you YouTube! Why do composers spend tremendous effort for the right combination of musical instruments? CHARACTERIZING THE EMOTION OF INDIVIDUAL PIANO AND OTHER MUSICAL INSTRUMENT SOUNDS

More information

Can parents influence children s music preferences and positively shape their development? Dr Hauke Egermann

Can parents influence children s music preferences and positively shape their development? Dr Hauke Egermann Introduction Can parents influence children s music preferences and positively shape their development? Dr Hauke Egermann Listening to music is a ubiquitous experience. Most of us listen to music every

More information

Enhancing Music Maps

Enhancing Music Maps Enhancing Music Maps Jakob Frank Vienna University of Technology, Vienna, Austria http://www.ifs.tuwien.ac.at/mir frank@ifs.tuwien.ac.at Abstract. Private as well as commercial music collections keep growing

More information

Musical Creativity. Jukka Toivanen Introduction to Computational Creativity Dept. of Computer Science University of Helsinki

Musical Creativity. Jukka Toivanen Introduction to Computational Creativity Dept. of Computer Science University of Helsinki Musical Creativity Jukka Toivanen Introduction to Computational Creativity Dept. of Computer Science University of Helsinki Basic Terminology Melody = linear succession of musical tones that the listener

More information

Subjective Similarity of Music: Data Collection for Individuality Analysis

Subjective Similarity of Music: Data Collection for Individuality Analysis Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp

More information

A combination of approaches to solve Task How Many Ratings? of the KDD CUP 2007

A combination of approaches to solve Task How Many Ratings? of the KDD CUP 2007 A combination of approaches to solve Tas How Many Ratings? of the KDD CUP 2007 Jorge Sueiras C/ Arequipa +34 9 382 45 54 orge.sueiras@neo-metrics.com Daniel Vélez C/ Arequipa +34 9 382 45 54 José Luis

More information

A prototype system for rule-based expressive modifications of audio recordings

A prototype system for rule-based expressive modifications of audio recordings International Symposium on Performance Science ISBN 0-00-000000-0 / 000-0-00-000000-0 The Author 2007, Published by the AEC All rights reserved A prototype system for rule-based expressive modifications

More information

ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC

ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC Vaiva Imbrasaitė, Peter Robinson Computer Laboratory, University of Cambridge, UK Vaiva.Imbrasaite@cl.cam.ac.uk

More information

Modeling memory for melodies

Modeling memory for melodies Modeling memory for melodies Daniel Müllensiefen 1 and Christian Hennig 2 1 Musikwissenschaftliches Institut, Universität Hamburg, 20354 Hamburg, Germany 2 Department of Statistical Science, University

More information

Exploring Relationships between Audio Features and Emotion in Music

Exploring Relationships between Audio Features and Emotion in Music Exploring Relationships between Audio Features and Emotion in Music Cyril Laurier, *1 Olivier Lartillot, #2 Tuomas Eerola #3, Petri Toiviainen #4 * Music Technology Group, Universitat Pompeu Fabra, Barcelona,

More information

A Categorical Approach for Recognizing Emotional Effects of Music

A Categorical Approach for Recognizing Emotional Effects of Music A Categorical Approach for Recognizing Emotional Effects of Music Mohsen Sahraei Ardakani 1 and Ehsan Arbabi School of Electrical and Computer Engineering, College of Engineering, University of Tehran,

More information

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC

TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu

More information

Running head: THE EFFECT OF MUSIC ON READING COMPREHENSION. The Effect of Music on Reading Comprehension

Running head: THE EFFECT OF MUSIC ON READING COMPREHENSION. The Effect of Music on Reading Comprehension Music and Learning 1 Running head: THE EFFECT OF MUSIC ON READING COMPREHENSION The Effect of Music on Reading Comprehension Aislinn Cooper, Meredith Cotton, and Stephanie Goss Hanover College PSY 220:

More information

The Role of Time in Music Emotion Recognition

The Role of Time in Music Emotion Recognition The Role of Time in Music Emotion Recognition Marcelo Caetano 1 and Frans Wiering 2 1 Institute of Computer Science, Foundation for Research and Technology - Hellas FORTH-ICS, Heraklion, Crete, Greece

More information

Keywords: Edible fungus, music, production encouragement, synchronization

Keywords: Edible fungus, music, production encouragement, synchronization Advance Journal of Food Science and Technology 6(8): 968-972, 2014 DOI:10.19026/ajfst.6.141 ISSN: 2042-4868; e-issn: 2042-4876 2014 Maxwell Scientific Publication Corp. Submitted: March 14, 2014 Accepted:

More information

A Comparison between Continuous Categorical Emotion Responses and Stimulus Loudness Parameters

A Comparison between Continuous Categorical Emotion Responses and Stimulus Loudness Parameters A Comparison between Continuous Categorical Emotion Responses and Stimulus Loudness Parameters Sam Ferguson, Emery Schubert, Doheon Lee, Densil Cabrera and Gary E. McPherson Creativity and Cognition Studios,

More information

Quantify. The Subjective. PQM: A New Quantitative Tool for Evaluating Display Design Options

Quantify. The Subjective. PQM: A New Quantitative Tool for Evaluating Display Design Options PQM: A New Quantitative Tool for Evaluating Display Design Options Software, Electronics, and Mechanical Systems Laboratory 3M Optical Systems Division Jennifer F. Schumacher, John Van Derlofske, Brian

More information

DIGITAL AUDIO EMOTIONS - AN OVERVIEW OF COMPUTER ANALYSIS AND SYNTHESIS OF EMOTIONAL EXPRESSION IN MUSIC

DIGITAL AUDIO EMOTIONS - AN OVERVIEW OF COMPUTER ANALYSIS AND SYNTHESIS OF EMOTIONAL EXPRESSION IN MUSIC DIGITAL AUDIO EMOTIONS - AN OVERVIEW OF COMPUTER ANALYSIS AND SYNTHESIS OF EMOTIONAL EXPRESSION IN MUSIC Anders Friberg Speech, Music and Hearing, CSC, KTH Stockholm, Sweden afriberg@kth.se ABSTRACT The

More information

Peak experience in music: A case study between listeners and performers

Peak experience in music: A case study between listeners and performers Alma Mater Studiorum University of Bologna, August 22-26 2006 Peak experience in music: A case study between listeners and performers Sujin Hong College, Seoul National University. Seoul, South Korea hongsujin@hotmail.com

More information

The Effects of Web Site Aesthetics and Shopping Task on Consumer Online Purchasing Behavior

The Effects of Web Site Aesthetics and Shopping Task on Consumer Online Purchasing Behavior The Effects of Web Site Aesthetics and Shopping Task on Consumer Online Purchasing Behavior Cai, Shun The Logistics Institute - Asia Pacific E3A, Level 3, 7 Engineering Drive 1, Singapore 117574 tlics@nus.edu.sg

More information

Crossroads: Interactive Music Systems Transforming Performance, Production and Listening

Crossroads: Interactive Music Systems Transforming Performance, Production and Listening Crossroads: Interactive Music Systems Transforming Performance, Production and Listening BARTHET, M; Thalmann, F; Fazekas, G; Sandler, M; Wiggins, G; ACM Conference on Human Factors in Computing Systems

More information

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY Eugene Mikyung Kim Department of Music Technology, Korea National University of Arts eugene@u.northwestern.edu ABSTRACT

More information

Estimating the Time to Reach a Target Frequency in Singing

Estimating the Time to Reach a Target Frequency in Singing THE NEUROSCIENCES AND MUSIC III: DISORDERS AND PLASTICITY Estimating the Time to Reach a Target Frequency in Singing Sean Hutchins a and David Campbell b a Department of Psychology, McGill University,

More information

Pitch correction on the human voice

Pitch correction on the human voice University of Arkansas, Fayetteville ScholarWorks@UARK Computer Science and Computer Engineering Undergraduate Honors Theses Computer Science and Computer Engineering 5-2008 Pitch correction on the human

More information

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular Music Mood Sheng Xu, Albert Peyton, Ryan Bhular What is Music Mood A psychological & musical topic Human emotions conveyed in music can be comprehended from two aspects: Lyrics Music Factors that affect

More information

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

Music Emotion Recognition. Jaesung Lee. Chung-Ang University Music Emotion Recognition Jaesung Lee Chung-Ang University Introduction Searching Music in Music Information Retrieval Some information about target music is available Query by Text: Title, Artist, or

More information

mood into an adequate input for our procedural music generation system, a scientific classification system is needed. One of the most prominent classi

mood into an adequate input for our procedural music generation system, a scientific classification system is needed. One of the most prominent classi Received, 201 ; Accepted, 201 Markov Chain Based Procedural Music Generator with User Chosen Mood Compatibility Adhika Sigit Ramanto Institut Teknologi Bandung Jl. Ganesha No. 10, Bandung 13512060@std.stei.itb.ac.id

More information

Quarterly Progress and Status Report. Musicians and nonmusicians sensitivity to differences in music performance

Quarterly Progress and Status Report. Musicians and nonmusicians sensitivity to differences in music performance Dept. for Speech, Music and Hearing Quarterly Progress and Status Report Musicians and nonmusicians sensitivity to differences in music performance Sundberg, J. and Friberg, A. and Frydén, L. journal:

More information

TongArk: a Human-Machine Ensemble

TongArk: a Human-Machine Ensemble TongArk: a Human-Machine Ensemble Prof. Alexey Krasnoskulov, PhD. Department of Sound Engineering and Information Technologies, Piano Department Rostov State Rakhmaninov Conservatoire, Russia e-mail: avk@soundworlds.net

More information

Acoustic and musical foundations of the speech/song illusion

Acoustic and musical foundations of the speech/song illusion Acoustic and musical foundations of the speech/song illusion Adam Tierney, *1 Aniruddh Patel #2, Mara Breen^3 * Department of Psychological Sciences, Birkbeck, University of London, United Kingdom # Department

More information

Toward a Computationally-Enhanced Acoustic Grand Piano

Toward a Computationally-Enhanced Acoustic Grand Piano Toward a Computationally-Enhanced Acoustic Grand Piano Andrew McPherson Electrical & Computer Engineering Drexel University 3141 Chestnut St. Philadelphia, PA 19104 USA apm@drexel.edu Youngmoo Kim Electrical

More information

Analysis, Synthesis, and Perception of Musical Sounds

Analysis, Synthesis, and Perception of Musical Sounds Analysis, Synthesis, and Perception of Musical Sounds The Sound of Music James W. Beauchamp Editor University of Illinois at Urbana, USA 4y Springer Contents Preface Acknowledgments vii xv 1. Analysis

More information

Environment Expression: Expressing Emotions through Cameras, Lights and Music

Environment Expression: Expressing Emotions through Cameras, Lights and Music Environment Expression: Expressing Emotions through Cameras, Lights and Music Celso de Melo, Ana Paiva IST-Technical University of Lisbon and INESC-ID Avenida Prof. Cavaco Silva Taguspark 2780-990 Porto

More information

Proceedings of Meetings on Acoustics

Proceedings of Meetings on Acoustics Proceedings of Meetings on Acoustics Volume 19, 2013 http://acousticalsociety.org/ ICA 2013 Montreal Montreal, Canada 2-7 June 2013 Musical Acoustics Session 3pMU: Perception and Orchestration Practice

More information

SYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS

SYNTHESIS FROM MUSICAL INSTRUMENT CHARACTER MAPS Published by Institute of Electrical Engineers (IEE). 1998 IEE, Paul Masri, Nishan Canagarajah Colloquium on "Audio and Music Technology"; November 1998, London. Digest No. 98/470 SYNTHESIS FROM MUSICAL

More information

A perceptual assessment of sound in distant genres of today s experimental music

A perceptual assessment of sound in distant genres of today s experimental music A perceptual assessment of sound in distant genres of today s experimental music Riccardo Wanke CESEM - Centre for the Study of the Sociology and Aesthetics of Music, FCSH, NOVA University, Lisbon, Portugal.

More information

Improving music composition through peer feedback: experiment and preliminary results

Improving music composition through peer feedback: experiment and preliminary results Improving music composition through peer feedback: experiment and preliminary results Daniel Martín and Benjamin Frantz and François Pachet Sony CSL Paris {daniel.martin,pachet}@csl.sony.fr Abstract To

More information

CS229 Project Report Polyphonic Piano Transcription

CS229 Project Report Polyphonic Piano Transcription CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project

More information

Affective Priming. Music 451A Final Project

Affective Priming. Music 451A Final Project Affective Priming Music 451A Final Project The Question Music often makes us feel a certain way. Does this feeling have semantic meaning like the words happy or sad do? Does music convey semantic emotional

More information

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset Ricardo Malheiro, Renato Panda, Paulo Gomes, Rui Paiva CISUC Centre for Informatics and Systems of the University of Coimbra {rsmal,

More information

Timbre blending of wind instruments: acoustics and perception

Timbre blending of wind instruments: acoustics and perception Timbre blending of wind instruments: acoustics and perception Sven-Amin Lembke CIRMMT / Music Technology Schulich School of Music, McGill University sven-amin.lembke@mail.mcgill.ca ABSTRACT The acoustical

More information

Automatic Rhythmic Notation from Single Voice Audio Sources

Automatic Rhythmic Notation from Single Voice Audio Sources Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung

More information

2. AN INTROSPECTION OF THE MORPHING PROCESS

2. AN INTROSPECTION OF THE MORPHING PROCESS 1. INTRODUCTION Voice morphing means the transition of one speech signal into another. Like image morphing, speech morphing aims to preserve the shared characteristics of the starting and final signals,

More information

Real-time Granular Sampling Using the IRCAM Signal Processing Workstation. Cort Lippe IRCAM, 31 rue St-Merri, Paris, 75004, France

Real-time Granular Sampling Using the IRCAM Signal Processing Workstation. Cort Lippe IRCAM, 31 rue St-Merri, Paris, 75004, France Cort Lippe 1 Real-time Granular Sampling Using the IRCAM Signal Processing Workstation Cort Lippe IRCAM, 31 rue St-Merri, Paris, 75004, France Running Title: Real-time Granular Sampling [This copy of this

More information

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm

Chords not required: Incorporating horizontal and vertical aspects independently in a computer improvisation algorithm Georgia State University ScholarWorks @ Georgia State University Music Faculty Publications School of Music 2013 Chords not required: Incorporating horizontal and vertical aspects independently in a computer

More information

INFLUENCE OF MUSICAL CONTEXT ON THE PERCEPTION OF EMOTIONAL EXPRESSION OF MUSIC

INFLUENCE OF MUSICAL CONTEXT ON THE PERCEPTION OF EMOTIONAL EXPRESSION OF MUSIC INFLUENCE OF MUSICAL CONTEXT ON THE PERCEPTION OF EMOTIONAL EXPRESSION OF MUSIC Michal Zagrodzki Interdepartmental Chair of Music Psychology, Fryderyk Chopin University of Music, Warsaw, Poland mzagrodzki@chopin.edu.pl

More information

Emotions perceived and emotions experienced in response to computer-generated music

Emotions perceived and emotions experienced in response to computer-generated music Emotions perceived and emotions experienced in response to computer-generated music Maciej Komosinski Agnieszka Mensfelt Institute of Computing Science Poznan University of Technology Piotrowo 2, 60-965

More information

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? NICHOLAS BORG AND GEORGE HOKKANEN Abstract. The possibility of a hit song prediction algorithm is both academically interesting and industry motivated.

More information

Computational Parsing of Melody (CPM): Interface Enhancing the Creative Process during the Production of Music

Computational Parsing of Melody (CPM): Interface Enhancing the Creative Process during the Production of Music Computational Parsing of Melody (CPM): Interface Enhancing the Creative Process during the Production of Music Andrew Blake and Cathy Grundy University of Westminster Cavendish School of Computer Science

More information

THE SOUND OF SADNESS: THE EFFECT OF PERFORMERS EMOTIONS ON AUDIENCE RATINGS

THE SOUND OF SADNESS: THE EFFECT OF PERFORMERS EMOTIONS ON AUDIENCE RATINGS THE SOUND OF SADNESS: THE EFFECT OF PERFORMERS EMOTIONS ON AUDIENCE RATINGS Anemone G. W. Van Zijl, Geoff Luck Department of Music, University of Jyväskylä, Finland Anemone.vanzijl@jyu.fi Abstract Very

More information

Ben Neill and Bill Jones - Posthorn

Ben Neill and Bill Jones - Posthorn Ben Neill and Bill Jones - Posthorn Ben Neill Assistant Professor of Music Ramapo College of New Jersey 505 Ramapo Valley Road Mahwah, NJ 07430 USA bneill@ramapo.edu Bill Jones First Pulse Projects 53

More information

Topics in Computer Music Instrument Identification. Ioanna Karydi

Topics in Computer Music Instrument Identification. Ioanna Karydi Topics in Computer Music Instrument Identification Ioanna Karydi Presentation overview What is instrument identification? Sound attributes & Timbre Human performance The ideal algorithm Selected approaches

More information

A repetition-based framework for lyric alignment in popular songs

A repetition-based framework for lyric alignment in popular songs A repetition-based framework for lyric alignment in popular songs ABSTRACT LUONG Minh Thang and KAN Min Yen Department of Computer Science, School of Computing, National University of Singapore We examine

More information

A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES

A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES A FUNCTIONAL CLASSIFICATION OF ONE INSTRUMENT S TIMBRES Panayiotis Kokoras School of Music Studies Aristotle University of Thessaloniki email@panayiotiskokoras.com Abstract. This article proposes a theoretical

More information

Topic 10. Multi-pitch Analysis

Topic 10. Multi-pitch Analysis Topic 10 Multi-pitch Analysis What is pitch? Common elements of music are pitch, rhythm, dynamics, and the sonic qualities of timbre and texture. An auditory perceptual attribute in terms of which sounds

More information

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu

More information

A PSYCHOACOUSTICAL INVESTIGATION INTO THE EFFECT OF WALL MATERIAL ON THE SOUND PRODUCED BY LIP-REED INSTRUMENTS

A PSYCHOACOUSTICAL INVESTIGATION INTO THE EFFECT OF WALL MATERIAL ON THE SOUND PRODUCED BY LIP-REED INSTRUMENTS A PSYCHOACOUSTICAL INVESTIGATION INTO THE EFFECT OF WALL MATERIAL ON THE SOUND PRODUCED BY LIP-REED INSTRUMENTS JW Whitehouse D.D.E.M., The Open University, Milton Keynes, MK7 6AA, United Kingdom DB Sharp

More information

Palmer (nee Reiser), M. (2010) Listening to the bodys excitations. Performance Research, 15 (3). pp ISSN

Palmer (nee Reiser), M. (2010) Listening to the bodys excitations. Performance Research, 15 (3). pp ISSN Palmer (nee Reiser), M. (2010) Listening to the bodys excitations. Performance Research, 15 (3). pp. 55-59. ISSN 1352-8165 We recommend you cite the published version. The publisher s URL is http://dx.doi.org/10.1080/13528165.2010.527204

More information

Construction of a harmonic phrase

Construction of a harmonic phrase Alma Mater Studiorum of Bologna, August 22-26 2006 Construction of a harmonic phrase Ziv, N. Behavioral Sciences Max Stern Academic College Emek Yizre'el, Israel naomiziv@013.net Storino, M. Dept. of Music

More information

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval DAY 1 Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval Jay LeBoeuf Imagine Research jay{at}imagine-research.com Rebecca

More information

Speech Recognition and Signal Processing for Broadcast News Transcription

Speech Recognition and Signal Processing for Broadcast News Transcription 2.2.1 Speech Recognition and Signal Processing for Broadcast News Transcription Continued research and development of a broadcast news speech transcription system has been promoted. Universities and researchers

More information

Making Progress With Sounds - The Design & Evaluation Of An Audio Progress Bar

Making Progress With Sounds - The Design & Evaluation Of An Audio Progress Bar Making Progress With Sounds - The Design & Evaluation Of An Audio Progress Bar Murray Crease & Stephen Brewster Department of Computing Science, University of Glasgow, Glasgow, UK. Tel.: (+44) 141 339

More information

Using the new psychoacoustic tonality analyses Tonality (Hearing Model) 1

Using the new psychoacoustic tonality analyses Tonality (Hearing Model) 1 02/18 Using the new psychoacoustic tonality analyses 1 As of ArtemiS SUITE 9.2, a very important new fully psychoacoustic approach to the measurement of tonalities is now available., based on the Hearing

More information

Understanding Compression Technologies for HD and Megapixel Surveillance

Understanding Compression Technologies for HD and Megapixel Surveillance When the security industry began the transition from using VHS tapes to hard disks for video surveillance storage, the question of how to compress and store video became a top consideration for video surveillance

More information

Detecting Musical Key with Supervised Learning

Detecting Musical Key with Supervised Learning Detecting Musical Key with Supervised Learning Robert Mahieu Department of Electrical Engineering Stanford University rmahieu@stanford.edu Abstract This paper proposes and tests performance of two different

More information

MusCat: A Music Browser Featuring Abstract Pictures and Zooming User Interface

MusCat: A Music Browser Featuring Abstract Pictures and Zooming User Interface MusCat: A Music Browser Featuring Abstract Pictures and Zooming User Interface 1st Author 1st author's affiliation 1st line of address 2nd line of address Telephone number, incl. country code 1st author's

More information

A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS

A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS A STATISTICAL VIEW ON THE EXPRESSIVE TIMING OF PIANO ROLLED CHORDS Mutian Fu 1 Guangyu Xia 2 Roger Dannenberg 2 Larry Wasserman 2 1 School of Music, Carnegie Mellon University, USA 2 School of Computer

More information

LOUDNESS EFFECT OF THE DIFFERENT TONES ON THE TIMBRE SUBJECTIVE PERCEPTION EXPERIMENT OF ERHU

LOUDNESS EFFECT OF THE DIFFERENT TONES ON THE TIMBRE SUBJECTIVE PERCEPTION EXPERIMENT OF ERHU The 21 st International Congress on Sound and Vibration 13-17 July, 2014, Beijing/China LOUDNESS EFFECT OF THE DIFFERENT TONES ON THE TIMBRE SUBJECTIVE PERCEPTION EXPERIMENT OF ERHU Siyu Zhu, Peifeng Ji,

More information

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models

Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Aric Bartle (abartle@stanford.edu) December 14, 2012 1 Background The field of composer recognition has

More information

Brain.fm Theory & Process

Brain.fm Theory & Process Brain.fm Theory & Process At Brain.fm we develop and deliver functional music, directly optimized for its effects on our behavior. Our goal is to help the listener achieve desired mental states such as

More information

Modelling Perception of Structure and Affect in Music: Spectral Centroid and Wishart s Red Bird

Modelling Perception of Structure and Affect in Music: Spectral Centroid and Wishart s Red Bird Modelling Perception of Structure and Affect in Music: Spectral Centroid and Wishart s Red Bird Roger T. Dean MARCS Auditory Laboratories, University of Western Sydney, Australia Freya Bailes MARCS Auditory

More information

Social Interaction based Musical Environment

Social Interaction based Musical Environment SIME Social Interaction based Musical Environment Yuichiro Kinoshita Changsong Shen Jocelyn Smith Human Communication Human Communication Sensory Perception and Technologies Laboratory Technologies Laboratory

More information