Opening musical creativity to non-musicians

Similar documents
THE EFFECT OF EXPERTISE IN EVALUATING EMOTIONS IN MUSIC

Designing new experiences of music making

Expressive information

YARMI: an Augmented Reality Musical Instrument

The relationship between properties of music and elicited emotions

Toward the Adoption of Design Concepts in Scoring for Digital Musical Instruments: a Case Study on Affordances and Constraints

Expressive performance in music: Mapping acoustic cues onto facial expressions

Affective response to a set of new musical stimuli W. Trey Hill & Jack A. Palmer Psychological Reports, 106,

Musical Entrainment Subsumes Bodily Gestures Its Definition Needs a Spatiotemporal Dimension

Lian Loke and Toni Robertson (eds) ISBN:

Supporting Creative Confidence in a Musical Composition Workshop: Sound of Colour

Computer Coordination With Popular Music: A New Research Agenda 1

Aesthetics and Design for Group Music Improvisation

Sudhanshu Gautam *1, Sarita Soni 2. M-Tech Computer Science, BBAU Central University, Lucknow, Uttar Pradesh, India

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

Embodied music cognition and mediation technology

Using machine learning to support pedagogy in the arts

Exploring Choreographers Conceptions of Motion Capture for Full Body Interaction

Ben Neill and Bill Jones - Posthorn

Mappe per Affetti Erranti: a Multimodal System for Social Active Listening and Expressive Performance

MAKING INTERACTIVE GUIDES MORE ATTRACTIVE

TOWARDS AFFECTIVE ALGORITHMIC COMPOSITION

15th International Conference on New Interfaces for Musical Expression (NIME)

Crossroads: Interactive Music Systems Transforming Performance, Production and Listening

Emotions perceived and emotions experienced in response to computer-generated music

Algorithmic Music Composition

Social Interaction based Musical Environment

Vuzik: Music Visualization and Creation on an Interactive Surface

a Collaborative Composing Learning Environment Thesis Advisor: Barry Vercoe Professor of Media Arts and Sciences MIT Media Laboratory

Electronic Musicological Review

"The mind is a fire to be kindled, not a vessel to be filled." Plutarch

Construction of a harmonic phrase

A prototype system for rule-based expressive modifications of audio recordings

Musical Creativity. Jukka Toivanen Introduction to Computational Creativity Dept. of Computer Science University of Helsinki

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

Music Mood. Sheng Xu, Albert Peyton, Ryan Bhular

The Reactable: Tangible and Tabletop Music Performance

This slideshow is taken from a conference presentation (somewhat modified). It summarizes the Temperley & Tan 2013 study, and also talks about some

DUNGOG HIGH SCHOOL CREATIVE ARTS

Composing Affective Music with a Generate and Sense Approach

Evolutionary Computation Applied to Melody Generation

Evolutionary jazz improvisation and harmony system: A new jazz improvisation and harmony system

Aalborg Universitet. Flag beat Trento, Stefano; Serafin, Stefania. Published in: New Interfaces for Musical Expression (NIME 2013)

The Effects of Web Site Aesthetics and Shopping Task on Consumer Online Purchasing Behavior

Shimon: An Interactive Improvisational Robotic Marimba Player

GESTURECHORDS: TRANSPARENCY IN GESTURALLY CONTROLLED DIGITAL MUSICAL INSTRUMENTS THROUGH ICONICITY AND CONCEPTUAL METAPHOR

UWE has obtained warranties from all depositors as to their title in the material deposited and as to their right to deposit such material.

THE SOUND OF SADNESS: THE EFFECT OF PERFORMERS EMOTIONS ON AUDIENCE RATINGS

12 Lynch & Eilers, 1992 Ilari & Sundara, , ; 176. Kastner & Crowder, Juslin & Sloboda,

DIGITAL AUDIO EMOTIONS - AN OVERVIEW OF COMPUTER ANALYSIS AND SYNTHESIS OF EMOTIONAL EXPRESSION IN MUSIC

This full text version, available on TeesRep, is the post-print (final version prior to publication) of:

MELODIC AND RHYTHMIC CONTRASTS IN EMOTIONAL SPEECH AND MUSIC

ESP: Expression Synthesis Project

Eliciting Domain Knowledge Using Conceptual Metaphors to Inform Interaction Design: A Case Study from Music Interaction

Advances in Algorithmic Composition

Controlling Musical Tempo from Dance Movement in Real-Time: A Possible Approach

Frankenstein: a Framework for musical improvisation. Davide Morelli

ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC

Robert Alexandru Dobre, Cristian Negrescu

3/2/11. CompMusic: Computational models for the discovery of the world s music. Music information modeling. Music Computing challenges

CONDUCT: An Expressive Conducting Gesture Dataset for Sound Control

BayesianBand: Jam Session System based on Mutual Prediction by User and System

Application of a Musical-based Interaction System to the Waseda Flutist Robot WF-4RIV: Development Results and Performance Experiments

Improving music composition through peer feedback: experiment and preliminary results

Emotional Remapping of Music to Facial Animation

Curriculum Standard One: The student will listen to and analyze music critically, using the vocabulary and language of music.

Sound visualization through a swarm of fireflies

Harmony and tonality The vertical dimension. HST 725 Lecture 11 Music Perception & Cognition

Quantifying Tone Deafness in the General Population

Improvised Duet Interaction: Learning Improvisation Techniques for Automatic Accompaniment

BRAIN-ACTIVITY-DRIVEN REAL-TIME MUSIC EMOTIVE CONTROL

Evolutionary Computation Systems for Musical Composition

EmbodiComp: Embodied Interaction for Mixing and Composition

PRESCOTT UNIFIED SCHOOL DISTRICT District Instructional Guide January 2016

PARTICIPATORY DESIGN RESEARCH METHODOLOGIES: A CASE STUDY IN DANCER SONIFICATION. Steven Landry, Myounghoon Jeon

Action and expression in music performance

Usability of Computer Music Interfaces for Simulation of Alternate Musical Systems

Automatic Notes Generation for Musical Instrument Tabla

1. BACKGROUND AND AIMS

Automatic music transcription

Interacting with a Virtual Conductor

SEEING IS BELIEVING: THE CHALLENGE OF PRODUCT SEMANTICS IN THE CURRICULUM

QUALITY OF COMPUTER MUSIC USING MIDI LANGUAGE FOR DIGITAL MUSIC ARRANGEMENT

Exploring the Effect of Interface Constraints on Live Collaborative Music Improvisation

Playful Sounds From The Classroom: What Can Designers of Digital Music Games Learn From Formal Educators?

An action based metaphor for description of expression in music performance

Comparison, Categorization, and Metaphor Comprehension

The Role of Time in Music Emotion Recognition

Designing a Musical Playground in the Kindergarten

2018 Indiana Music Education Standards

MUSICAL MOODS: A MASS PARTICIPATION EXPERIMENT FOR AFFECTIVE CLASSIFICATION OF MUSIC

MODELING MUSICAL MOOD FROM AUDIO FEATURES AND LISTENING CONTEXT ON AN IN-SITU DATA SET

Various Artificial Intelligence Techniques For Automated Melody Generation

Visualizing Euclidean Rhythms Using Tangle Theory

Praxis Music: Content Knowledge (5113) Study Plan Description of content

Subjective Similarity of Music: Data Collection for Individuality Analysis

INFLUENCE OF MUSICAL CONTEXT ON THE PERCEPTION OF EMOTIONAL EXPRESSION OF MUSIC

THEORY AND COMPOSITION (MTC)

An Investigation into the Tuition of Music Theory using Empirical Modelling

Access from the University of Nottingham repository:

Transcription:

Opening musical creativity to non-musicians Fabio Morreale Experiential Music Lab Department of Information Engineering and Computer Science University of Trento, Italy Abstract. This paper gives an overview of my PhD research that aims at contributing toward the definition of a class of interfaces for music creation that target non-musicians. In particular, I am focusing on the differences on design and evaluation procedures with respect to traditional interfaces for music creation that are usually intended to be used by musicians. Supported by a number of preliminary findings we developed the first interactive system: The Music Room is an interactive installation which enables people to compose tonal music in pairs by communicating emotion expressed by moving throughout a room. Keywords: Musical interfaces, user-experience, performing art, active listening. 1 Research questions I am a third year PhD candidate at the HCI group of the University of Trento, guided by Antonella De Angeli. The focus of my study is to design interactive systems to allow everybody to experience musical creativity. So far, the inherent complexity of music composition limits the access of traditional musical interfaces to musicians due to the extensive presence of musical notation. Novel technologies (e.g. tabletops, mobile apps, motion capture sensors) have been adopted to replace traditional instruments with more intuitive devices [1, 2] and lead a new set of design issues. As musical notation fails on giving everybody access to music creation, what kind of interaction paradigm can be used in order to ease this art to a wider and lay audience? In order to answer this question, we explored new interaction metaphors that have to meet a series of requirements: they have to be available to everybody, intuitive, with a proper level of affordance and naturally connected with music. Emotion seems to be the element that best meets these requirements. Music is one of the arts that can most effectively elicit emotions [3, 4] and it has always been connoted as emotional [5]. In interactive systems, emotions need to be mediated by specific artefacts in order to be communicated to the system. Bodily movements, which, in the different declinations of dancing and conducting, are traditionally associated to music, can be the most appropriate medium through which emotions can be conveyed [5].

2 Our contribution The first interface we developed is The Music Room, an installation that provides a space where people can compose music expressing their emotions through movements. The visitors experience the installation in couple by informing the system on the emotions they intend to elicit. The couple directs the generation of music by providing information about the emotionality and the intensity of the music they wish to create. To communicate emotions, the analogy with love is used: the proximity between them affects the pleasantness of the music, while their speed affects the dynamic and intensity. We decided to limit the interaction dimensions to closeness and speed in order to keep the experience as simple and intuitive as possible. Proxemics information is acquired by a vision tracking system. It is then converted into emotional cues and finally passed to the musical engine. These intuitive compositional rules provide everybody with unlimited musical outcomes. As regard the generation of music, we developed Robin, an algorithmic composer that composes original tonal music in piano 1. Fig. 1. The Music Room. 1 Examples of pieces generated at The Music Room can be listened at goo.gl/ulhgz

3 Related works This project spans several research areas. The adoption of the metaphor of gestures and emotions is partially influenced by previous collaborative interactive systems for music generation. The rules of the compositional system are founded research on music perception, while Robin is inspired by existing approaches for algorithmic composition. 3.1 Interactive Musical System Research on the design of interactive systems for generative music has been growing in the last decade. A number of tangible musical interfaces such as the Reactable [1], Jam-O-Drum [17], and GarageBand for the ipad, target users that have at least a minimum musical training as sonic and musical inputs are adopted. A category of interfaces addresses users to collaborate. In particular, several works exploit the concept of active listening, an approach where listeners can interactively control the music content by modifying it in real-time while listening to it [18, 19]. TouchMeDare aims at encouraging strangers to collaborate for reaching a common creative goal: pre-composed music samples are triggered when both simultaneously touch a canvas [22]. In the Urban Musical Game, users manipulate pre-composed music by playing with sensors-equipped foam balls [21]. With Sync n Move music is also experienced by collaborative means [23]. Two users freely move their mobile phones and the level of music orchestration depends on the synchronization of their movements. In Mappe per Affetti Erranti, a group of people can explore pre-composed music by navigating a physical and emotional space [20]. Once again, collaborative situations are encouraged as music can only be listened to in its full complexity if the participants cooperate. 3.2 Eliciting emotions in music Related works suggest that the perception of emotions in music depends on compositional parameters (e.g. tempo, melody direction, mode) and performance behaviors (articulations, timing, dynamics) whose combinations elicit different emotional responses in the listener [5 7]. Measurement and classification of emotions in music, most of the works in the music domain are operates on Russell s Circumplex model [8]. This model describes emotions as a continuum along two dimensions: valence and arousal. In 1937, Hevren identified the most important compositional factors in terms of emotions elicitation by labelling them on the music s expressive character [9]. Juslin and Sloboda later reviewed this categorization by representing the emotions along the valence/arousal dimensions [10]. There is a consensus that at the compositional level, mode and rhythm are responsible for valence, while tempo and dynamics are responsible for arousal. Despite the remarkable amount of works on this area, no significant study has been tried to understand to which extent expertise has a role on judging, appreciating and perceiving musical pieces. How do non-musicians perceive and

describe music? What are the musical parameters and semantic elements that are more relevant for them? 3.3 Algorithmic Composition Generative music composition has been widely explored in the last decade. The most common approaches are: rule-based, learning-based and evolutionary composition [10]. In rule-based approach, algorithmic rules inspired from music theory are manually coded into the system [11, 12]. In learning-based approach, the system is trained with existing musical excerpts and a number of rules are automatically included [13, 14]. Even though this method manages to decrease the reliance on designer skills on music theory, the output heavily depends on the training set. Lastly, evolutionary algorithms allow the creation of original and complex melodies by means of computational approaches inspired by biological evolution [15]. The generated music is original and unpredictable but it might sound unnatural and lack ecological validity if compared to rule-based systems that are generally superior in contexts of tonal music [16]. 4 Results achieved A number of personal contributions for each of the three research areas were recently published. At the Interactivity session of CHI 2013, we demoed The Music Room [24], whose objectives, development, findings and evaluation are better discussed on the forthcoming publication at the Special Issue of Pervasive and Ubiquitous Computing on Designing Collaborative Interactive Spaces. The role of expertise on the evaluation of induced emotions in music was analysed in a experiment we conducted in 2012 whose results were recently published on Proceedings of ICME3 [25]. The details on the ideation and implementation of Robin, the algorithmic composer, are going to be published at Proceedings of SMC2013 [26]. 5 Future works The last year of my PhD will be mainly devoted toward a formal definition of interactive systems for music creation that target non-musicians. The first objective is to investigate similarities and differences with traditional digital musical interfaces. By date, just a few studies highlighted the differences between interfaces for artistic experience and for musical expression but these works didn t have a follow-up in the last decade [27]. However, we believe that a number of relevant differences exist. By combining personal intuitions with related literature findings, we propose a list of potential differences between the two sets. Possibly, the output of this study will consist of a categorization of musical interfaces. The idea is to exhaustively describe musical interfaces by means of a model composed of a space of number of dimensions such as:

Target user Ultimate goal Learning curve Interaction metaphor Level of direction Musical possibilities Role of the audience The successive step would consist in testing the proposed dimensions with a series of existing interfaces. Once validated, I will elaborate on defining a series of evaluation principles for each dimension. This will allow interface designers to position their projects on the model and to evaluate them consequently. I d also wish to tackle a number of challenges regarding The Music Room. Even though the current implementation received a lot of interest, there is room for several improvements. A number of innovations to music engine are currently under development: the quality of music will be enhanced by introducing new genres and instruments as well as by teaching Robin new compositional rules. I also aim at further elaborating on the communication of intended emotions to the system. Temporal aspects will be taken into consideration in order to determine a general evolution of the experience, by considering recurrence of patterns of moving close and getting far. Also, we are likely to introduce head pose tracking in order to have information whether the two people are looking at each other. This additional information will be used to differentiate the situations in which the user are facing or turned and direct the music generation consequently. References 1. Jordà, S., Geiger, G., Alonso, M., & Kaltenbrunner, M. (2007). The reactable: exploring the synergy between live music performance and tabletop tangible interfaces. Proceedings of the 1st international conference on Tangible and embedded interaction (pp. 139-146). ACM. 2. Camurri, a., Volpe, G., De Poli, G., & Leman, M. (2005). Communicating expressiveness and affect in multimodal interactive systems. IEEE Multimedia, 12(1), 43-53. 3. Balkwill, L.-L., Thompson, W. F., & Matsunaga, R. I. E. (2004). Recognition of emotion in Japanese, Western, and Hindustani music by Japanese listeners. Japanese Psychological Research, 46(4), 337-349. 4. Fritz, T., Jentschke, S., Gosselin, N., Sammler, D., Peretz, I., Turner, R., Friederici, A.D., Koelsch, S.,(2009). Universal Recognition of Three Basic Emotions in Music, Current Biology, Volume 19, Issue 7, 14 April 2009, Pages 573-576, ISSN 0960-9822 5. Juslin P.N., & Sloboda J.A. (2009). Handbook of music and emotion: theory, research, applications. Oxford University Press. 6. Temperley, D. (2001). The Cognition of Basic Musical Structures. Mit Press. 7. Gabrielsson, A., Lindstrm, E., (2001). The influence of musical structure on emotional expression. Music and emotion: Theory and research. Series in affective science., (pp. 223-248) 8. Russell, J. (1980). A circumplex model of affect. Journal of personality and social psychology.

9. Hevner, K. (1937). The Affective Value of Pitch and Tempo in Music. The American Journal of Psychology, 49(4), 621-630 CR - Copyright & 169; 1937 University of Illinois Press. 10. Todd, P. M., & Werner, G. M. (1999). Frankensteinian methods for evolutionary music composition. Musical networks: Parallel distributed perception and performance, 313-339. 11. Rader, G. (1974). A method for composing simple traditional music by computer. IT press, Cambridge, MA 12. Steedman, M. (1984). A Generative grammar for jazz chord sequences. Music Perception Vol. 2, No. 1, Fall, 1984 13. Simon, I., Morris, D., & Basu, S. (2008). MySong: automatic accompaniment generation for vocal melodies. Proceeding of the twenty-sixth annual SIGCHI conference on Human factors in computing systems (pp. 725-734). ACM. 14. Hiller, L., Isaacson, L. (1992). Musical composition with a high-speed digital computer, Machine models of music. 15. Wiggins, G., Papadopoulos, G., Phon-amnuaisuk, S., & Tuson, A. (1998). Evolutionary Methods for Musical Composition. United Kingdom 06 - Biological and medical sciences. 16. Nierhaus, G. (2009). Algorithmic Composition: Paradigms of Automated Music Generation. Springer 17. Blaine, T., & Perkis, T. (2000). The Jam-O-Drum interactive music system: a study in interaction design. In Proceedings of the 3rd conference on Designing interactive systems: processes, practices, methods, and techniques (DIS 00). ACM. 18. Rowe, R. (1993). Interactive music systems: Machine listening and composition. MIT Press, Cambridge MA, 1993. 19. Camurri, A., Canepa, C., & Volpe, G. (2007). Active listening to a virtual orchestra through an expressive gestural interface: The Orchestra Explorer. Proceedings of the 7th international conference on New interfaces for musical expression (pp. 56-61). 20. Camurri, A., Varni, G., & Volpe, G. (2010). Towards analysis of expressive gesture in groups of users: computational models of expressive social interaction. Gesture in Embodied Communication and Human-Computer Interaction, (1947), 122-133. 21. Van Boerdonk K.,Tieben Rob, Klooster S., Van den Hoven E., (2009). Contact through canvas: an entertaining encounter. Personal and Ubiquitous Computing. Springer. 22. Rasamimanana, N., Bevilacqua, F., Bloit, J., Schnell, N., Flty, E., Cera, A., Frechin, J.-L., et al. (2012). The urban musical game: using sport balls as musical interfaces. CHI 2012 Proceedings, 1027-1030. 23. Varni, G., Mancini, M., Volpe, G., & Camurri, A. (2010). Sync n Move: social interaction based on music and gesture. User Centric Media, 31-38. 24. Morreale, F., Masu, R., De Angeli, A., Rota, P. (2013). The Music Room. CHI 13 Extended Abstracts on Human Factors in Computing Systems, 3099-3102 (2013). 25. Morreale, F., Masu, R., De Angeli, A., Fava, P. (2013). The Effect of Expertise in Evaluating Emotions in Music. Proceedings of the 3rd International Conference on Music & Emotion. 26. Morreale, F., Masu, R., De Angeli, A. (2013). Robin: An Algorithmic Composer For Interactive Scenarios. Proceedings of Sound And Music Computing. 27. Blaine,T., Fels, S. Contexts of collaborative musical experiences. In Proceedings of the 2003 conference on New interfaces for musical expression (NIME 03).