Visual enhancement using multiple audio streams in live music performance
|
|
- Gwendoline Caldwell
- 5 years ago
- Views:
Transcription
1 Visual enhancement using multiple audio streams in live music performance Rozenn Dahyot 1, Conor Kelly 1, and Gavin Kearney 2 1 School of Computer Science and Statistics, Trinity College Dublin, Ireland 2 Department of Electronic and Electrical Engineering, Trinity College Dublin, Ireland Correspondence should be addressed to Rozenn Dahyot (Rozenn.Dahyot@cs.tcd.ie) ABSTRACT The use of multiple audio streams from digital mixing consoles is presented for application to real-time enhancement of synchronised visual effects in live music performances. The audio streams are processed simultaneously and their temporal and spectral characteristics can be used to control the intensity, duration and colour of the lights. The efficiency of the approach is tested on rock and jazz pieces. The result of the analysis is illustrated by a visual OpenGL 3-D animation illustrating the synchronous audio-visual events occurring in the musical piece. 1. INTRODUCTION Visual effects such as stage lighting or fog machines are widely used in live music performances to enhance the emotion and mood of the music played. Such schemes are designed to visually immerse the audience in the feeling of the song. Video displays such as TV screens or video projectors are now standard facilities in small to large size venues and recent trends in art tend to design computer programs to automatically allow interactions between the music and the visual effects [1]. Typically in small to medium sized auditoriums, sound reinforcement for jazz and rock ensembles performing on stage involves the use of around 8 microphones, a mixing console and loudspeaker amplification. The microphone signals are pre-amplified and processed at the console and a stereo mix for Front of House amplification is generated. This stereo mix is fed to the lighting desk which allows control over several effects of the stage lights (colour, flash, intensity, direction, etc.). Often artificial intelligence is involved in the making of shows to assist in the work of sound and lighting engineers. For example, one popular automatic process is real-time beat detection implemented on a basic level on lighting desks [2]. The visual effects can then be synchronised to the music. Such algorithms tend to focus primarily on the low-frequency content of the stereo mix to infer tempos since the mid and upper frequency ranges are generally cluttered from the mix of sources. However, current lighting systems do not avail of the multiple audio streams available from digital mixing consoles such as the popular Tascam Digital Interface (TDIF) or Alesis Digital Audio (ADAT) protocols. We propose here to process in real-time a multi-channel audio stream from a digital mixing console to perform reliable lighting enhancement through temporal beat detection and frequency analysis. The advantage of such settings is that the musical content of each instrument is well separated, since, in well engineered performances the sound pressure level of a particular instrument contributes greater than the spill from the other instruments at its corresponding microphone. Thus, no processing for source separation is required for the different instruments. The temporal and spectral characteristics of these signals can then be analysed simultaneously to generate enhanced visual effects. Another advantage of using the separated sources is that the mid to high frequency components which are crucial in determination of signal attack are uncluttered. Thus a high audio resolution is of importance for the accurate detection of pitch and the temporal properties of higher frequency percussive instruments, such as hi-hats, as well as visual enhancement of spatial effects such as reverbs or delays on vocals or guitars for instance. We propose here to create a portable affordable system that will help automatically generate in real-time a visual artistic rendering of the music being played live in 1
2 a small or medium venue without the undesireable budget constraints that face many working artists. As an alternative to lightings, we illustrate our multi-stream music analysis by creating a real-time OpenGL animation that reacts to events in the music piece. Such a system could be used to increase the exposure of not yet well-known artists to international audience in the virtual world (e.g. by simultaneously performing in Second Life Our smart system has been tested on jazz and rock pieces. We show that real-time high resolution multistream music analysis can be performed, with reliable accuracy. We use various methods such as frequency spectrum analysis, beat detection amplitude analysis to get a feel for the mood and tempo of the song (see section 4). An animation is then created representing with more or less accuracy the members of the bands with their instruments on a stage. The lightings and the motion of the characters in the rendering changes in realtime according to the song (see section 5). Section 6 comments on the current performances of our system. 2. RELATED WORKS The work presented in this paper mixed different domains of computer science: digital music processing and computer graphics. In the following paragraphs, references are given to both areas of research Analysis of music Digital processing of music has attracted a lot of attention mainly due to the high commercial value for online sells of songs. A huge literature exists about features that are efficient to process music, and a good review can be found in [3]. Some of them, such as loudness, Fourier transform, energy band or median frequency have been used in our system and are presented in section 4. Beat detection significantly aids the classification of music genre, and is an elementary step for more thorough analysis of the music [4]. Further methods propose to infer the structure of popular music have been proposed by Maddage [4] and include music transition, voice detection and repeated pattern detection. Application of these methods can be found in music transcription, music summarisation and retrieval, and also in music streaming Visual music Perception of music. By stimulating a second sense along with music, visual effects have the ability to contribute to the communication that takes place between performers and their listeners [5]. Examples of visual contributions in music performances include facial expression or body gesture and movement of the performers, video projections, light and pyrotechnic shows amongst others. They amplify the emotion of the music and completely immerse the observer in the feeling of the song. Visuals & Graphics. In the following paragraphs, we report several computer-aided systems that have been proposed to generate visuals for music. Visuals can be created in many ways e.g. by films and lighshows and, as an illustration, the reader can visit the web exhibition visual music [6] that presents several visual expressions explored by artists to extend the perception of music. One important visual cue to music is the natural movement of the body expressed by performers [5] or by the listeners. For instance human foot taps are inferred from the perceived beat. Dancing is also a natural illustration of music. Denman et al. [7] have proposed to synchronise the beat of a song with the visual motion of a dancer in a video (performed for another song). Time-scale changes of the video feed are then performed to synchronise the detected beat of the new song with the movements of the dancer. The beat detection from a monophonic audio stream and the extraction of the motion from the video are performed offline. The created video can have application for generating visuals in nightclubs or for postproduction in music videos. Several applications using music to synchronise in realtime computer graphics animations have been proposed [8, 9, 1]. Applications can be found for entertainment or for learning music In [9], the motion curves is synchronise to music in computer animations. In [1], the graphics animation mimics the expressiveness of a drummer. However, on the contrary to [7], no audio analysis is performed as the relevant cues are already available in the complementary MIDI 1 stream of the soundtrack [8, 9, 1]. MIDI stores the events that would create the sound instead of the sound itself. This allows easy access to pitch, velocity, instrument and timing information. Using a MIDI file instead of a raw audio signal avoids the need of performing digital music processing 1 Musical Instrument Digital Interface Page 2 of 7
3 in realtime. Unfortunatly, this information is not always accessible in realtime from every instrument playing in a musical piece. 3. OVERVIEW OF OUR SYSTEM We consider live performance of small band Rock or Jazz with a few musicians and instruments. On stage, several microphones are placed close to each instrument. We assume the availability of multichannel audio streams from TDIF or ADAT interfaces from such consoles as the Yamaha 2R96 or the standalone analog to digital conversion capabilities of units such as the MOTU 248MK2. Figure 1 show an overview of the system. Several audio recordings coming from different microphones (mic1, mic2, etc.) are available for analysis. The final mix of the song is used only as a soundtrack in the final rendering. to process in real-time but would require computationally expensive routines to separate the contribution from each musician [11]. As an alternative, we propose to take the advantage of the direct out or bussing facilities available on most mixing consoles so that the seperated audio is already presented for analysis. This choice allows us to use simple and fast algorithms to extract reliably relevant music features but has the drawback of requiring the analysis of several audio streams in parallel. 4. MUSIC ANALYSIS Currently four audio channels are analysed simultaneously. These correspond most of the time to the microphones of the singer (voice), the drums, the guitarist and the bass. For the Jazz piece we analysed, the saxophone is selected instead of the guitar. In our simulation, independent audio streams are stored as mono WAV files sampled at 441Hz. Figure 2 shows an example of these different recordings for an extract of a rock song Fig. 1: Overview of our system. Fig. 2: 1 seconds of a rock song. From top to bottom: Pressure signals of singer voice, guitar, bass and drums. The main advantage of considering separate audio sources from each microphones instead of the mixed track resides in the fact that the different sources are well separated. In fact the closest instrument to each microphone is the one mainly audible on the corresponding audio stream little spill from the other instruments ranging in the order of 3 to 5dB less. Using only the mix to analyse the music would lessen the amount of data 4.1. Beat detection The beat detection algorithm is performed onto the drum audio stream. For the audio signal x(t), the loudness is computed for each window of 124 samples (ie = =.232s by: (n+1) l(n ) = x 2 (t) dt, starting at n = (1) n Page 3 of 7
4 The detection of the beat is performed by thresholding the loudness information. To be independent of background noise during the performance or the different loudness of the different sources, the threshold T is adaptative for each frame: T (n ) = γ 1 1 n n 1 x 2 (t) dt (2) where γ is a proportional coefficient set by hand at γ = 1.4 and the normalised integral corresponds to the average loudness in the preceding second before the window n. A beat is then detected when l(n ) > T (n ). Sometimes several successive temporal windows are detected above the threshold. Consequently only the first detected beat amongst a successive sequence of detected beats is actually labelled as a beat (i.e. the rule is that a beat is impulsive and cannot be detected in successive windows). Figure 3 shows the results of our beat detection performed on ten seconds of a rock song performed live. As can be noticed, the detected beat is sometimes one temporal window in advance of the actual peak in the loudness signal. This means that when a beat is detected, it is with =.232 second of accuracy. This temporal precision in the audio analysis is largely sufficient as the visual rendering is only changing every.4 second (i.e. the animation has 25 frames per second). computed every 23.2ms (or 124 samples) as follow: (n+1) X(n, f ) = x(t) exp( 2π ft) dt (3) n Using a adapted passband filter, each instrument is separated from any possible spills coming from other sources, and information such as the band energy of the instrument (or the voice) is recorded: f1 A(n ) = X(n, f ) d f (4) f where [ f, f 1 ] define the frequency band of the instrument. Without too much additional computational cost, the median frequency or the mean of the spectrum is also computed as follow [3]: f (n ) = f1 f X(n, f ) f d f f1 f X(n, f ) d f (5) These are the measures used in our system. Other informative features such as pitch can also be computed if the computation time remains low for the hardware used. 5. REAL-TIME ANIMATION Fig. 3: Result of beat detection performed on 1s of the audio track of drums for a rock song recorded in live session (cf. fig. 2). Red dot indicate detected beat and the blue curve corresponds to the loudness computed every Fourier Analysis For each audio stream, a Fast Fourier Transform (FFT) is A simulation of a stage complete with lighting effects is rendered on screen. This rendering is created and drawn using the OpenGL graphics library. Some of the various graphics methods used in the render include (see figure 4) 3D modelling, texture mapping and Tesselated objects as explained in the following paragraphs D Modeling The musicians and many of the stage props such as the drums, guitars, microphones, lights and light rig were modelled in 3D Studio Max as 3D models and then imported into opengl. Figure 5 shows a screen shot of the drums being modelled in 3D Studio Max. 3DS models are composed of many thousands of vertices and their texture coordinates. They are quite computational expensive to draw. For this reason, there is a trade off in the detail represented in the simulation and the speed at which it runs Texture mapping Texture mapping is the process of applying textures (stored as jpegs) to shapes drawn in order to add color Page 4 of 7
5 5.3. Tesselated objects OpenGL uses the Phong Illumination model to calculate lighting in scenes. This calculates light reflections at each vertex of an object and interpolates the light to surrounding polygons. As spot lights will be shining down onto the stage floor, it is required to display and reflect them realistically. To do this, the stage floor is drawn as a very fine mesh of vertices in a process known as tessellation. However, as it is a computationally expensive process, a trade off in between realism and computation has been found Animation to render the music feel Fig. 4: A screen shot of the render: Vocalist and guitarist are illuminated in the foreground, ambient lighting shines red in the background indicating an uptempo beat. Fig. 5: Modelling the drum kit in 3D Studio Max. and realism to the scene. The less complicated objects in the render such as the enclosing walls of the stage and the front of the stage floor can be presented with far fewer vertices and so are drawn with their static coordinates specified in the code. The texture mapping coordinates are also specified and so at render time, these textures are applied to the shapes to give them a realistic look. This is far more efficient than drawing 3DS models and so is used wherever possible. The visualisation of the information extracted from the music played is done in three ways: 1. Spot lighting. The most important of these is the concentration of spotlighting on any musician who is currently active. This uses information from the FFT performed on each musician s channel. The energies A are calculated and when a certain threshold level is breached, a spot light is shone on the musician. 2. Ambient lighting. Rather than concentrate on individual channels, ambient lighting s focus is on the behaviour of the song as a whole and so concentrates on the energies of all channels. It analyses the predominant frequencies of the FFT (i.e. f )and uses tempo information to attempt to provide ambient lighting in accordance with the mood of the song. The interpretation of the songs is based on the generalisation that lower frequencies and lower tempos indicate a more relaxed mood. This triggers low key colors such as purple or dark red. Brighter colors illuminate the stage when songs occupy higher frequency bands and have faster tempos to reflect the more exited performance. 3. Physical movement of the musicians on stage. Spot lighting will demonstrate that a musician s level has breached a certain threshold and is deemed to be playing or singing. Simple animation of the characters is performed give more information of exactly how loud their part is. This is shown by the speed at which the musicians arms and body move, the faster they move and the bigger a part in the overall mix they are playing. This is done using hierarchial Page 5 of 7
6 Dahyot et al. animation of the 3D models. The models seen in the render are actually made up of several models (body, head, legs etc. ) which are drawn together in OpenGL to simulate a person. These can be rotated and moved around each other to animate movement. This animation is controlled in accordance with the music analysis so their movement is directly linked to the music they are producing. Figure 6 shows two images from a recorded animation. A rock song is played where first only one guitar and drums are playing. Then a second guitar starts soloing. The yellow lighting on the musicians indicate if they are currently playing. The reddish lightings on the soloist illustrate the measure f computed in real-time by changing linearly from blue to red. For a better visualisation, a bar with a moving and changing colour spot indicates the value of f in its range in realtime. 6. PERFORMANCE AND OPTIMIZATION 6.1. Hardware The system as described in this paper is currently running on a standard laptop (Model: HP NX942) with Intel Core Duo CPU at 1.66GHz, 512mb of RAM memory and a graphics card ATi Radeon X16 256MB. The music analysis and the animation are created in real-time 25 frames per second for the video rendering Computational efficiency Computational efficiency of the code is of major issue for our system to work. As a lot of calculation is taking place for the analysis of multi-stream music, we need also a reasonable reaction time in the rendering to avoid desynchroning artefacts. Various methods are used in the graphics component of the project to ensure optimal performance. One such method is the use of hardware precaching of OpenGL display lists that allow some command to be precompiled onto the graphics card s memory and so removes the need for the CPU to perform repeated expensive calculations. This takes advantage of the dedicated memory and computation power of modern GPUs. In a direct comparison between the code with no hardware pre-caching and the code which makes use of display lists, a speed up (measured in the frame per second count) of roughly 4% was achieved. (a) 6.3. Perception of the animation Some results of the system are shown as videos at DemosMusic.html. The system has been successively tested using four simultaneous audio channels from rock and jazz bands, mainly in live performance situations but also with less noisy environments such as in studio recordings. The perceived animation is well synchronised to the beat, in particular the lights and the movement of the drummer. (b) Fig. 6: (a) The singer and one guitarist are not playing and are in the dark. (b) The guitarist is playing a solo, and using the median frequency f, the colour of the red lights varies from blue (low values of f ) to red (high values of f ). AES 31ST INTERNATIONAL CONFERENCE, London, England, 27 JUNE Page 6 of 7
7 7. CONCLUSION AND FUTURE WORK We have presented an innovative system using multichannel music recordings for real-time rendering. Using the computational power of a recent laptop, we have shown how to simultaneously perform music analysis and render a graphic animation expressing some aspect of the music being played. Both CPUs and GPUs abilities have being used to speed up the system. Future directions of this research will look at creating other animations illustrating better the music, such as using changes on facial expressions on one virtual face [5], or by animating a virtual dancer [8], or more generally to create more expressive animations. The music processing part of the system can also be improved by using prior information, for instance for the beat detection where currently no past information is used (i.e. beats are detected without the knowledge when the last beat was detected). The use of other informative audio features such as pitch will also be investigated. ACKNOWLEDGEMENTS Part of this work has been funded by the European Network of Excellence on Multimedia Understanding through Semantics, Computation and Learning MUS- CLE FP , [7] H. Denman and A. Kokaram, Dancing to a different tune, in 2nd IEE European Conference on Visual Media Production (CVMP), 3 Nov. - 1 Dec. 25, pp [8] D. Reidsma, A. Nijholt, R. Poppe, R. Rienks, and G. Hondorp, Virtual rap dancer: Invitation to dance, in CHI 6 extended abstracts on Human factors in computing systems. ACM, 26, pp [9] M. Cardle, L. Barthe, S. Brooks, and P. Robinson, Music-driven motion editing: Local motion transformations guided by music analysis, in 2th IEEE Eurographics UK Conference (EGUK), 22. [1] A. M. Wood-Gaines, Modelling expressive movement of musicians, Master s thesis, MSc Computing Science, Simon Fraser University, February [11] S. Choi, A. Cichocki, H. Park, and S.-Y. Lee, Blind source separation and independent component analysis: A review, Neural Information Processing - Letters and Reviews, vol. 6, no. 1, pp. 1 57, January REFERENCES [1] T. Winkler, Composing Interactive Music - Techniques and Ideas Using Max. MIT Press, [2] U. Sandström, Stage lighting Controls. Focal Press, [3] M. Davy and S. Godsill, Audio information retrieval: a bibliographical study, University of Cambridge, UK, Tech. Rep., November 21. [4] N. Maddage, Automatic structure detection for popular music, IEEE Multimedia, vol. 13, no. 1, pp , 26. [5] W. F. Thompson, P. Graham, and F. A. Russo, Seeing music performance: Visual influences on perception and experience, Semiotica, pp , 25. [6] Visual music, visualmusic/, Hirshhorn Museum, 25. Page 7 of 7
SOUND REINFORCEMENT APPLICATIONS
CHAPTER 6: SOUND REINFORCEMENT APPLICATIONS Though the Studio 32 has been designed as a recording console, it makes an excellent console for live PA applications. It has just as much (if not more) headroom
More informationPOST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS
POST-PROCESSING FIDDLE : A REAL-TIME MULTI-PITCH TRACKING TECHNIQUE USING HARMONIC PARTIAL SUBTRACTION FOR USE WITHIN LIVE PERFORMANCE SYSTEMS Andrew N. Robertson, Mark D. Plumbley Centre for Digital Music
More informationQuery By Humming: Finding Songs in a Polyphonic Database
Query By Humming: Finding Songs in a Polyphonic Database John Duchi Computer Science Department Stanford University jduchi@stanford.edu Benjamin Phipps Computer Science Department Stanford University bphipps@stanford.edu
More informationInteracting with a Virtual Conductor
Interacting with a Virtual Conductor Pieter Bos, Dennis Reidsma, Zsófia Ruttkay, Anton Nijholt HMI, Dept. of CS, University of Twente, PO Box 217, 7500AE Enschede, The Netherlands anijholt@ewi.utwente.nl
More informationHowever, in studies of expressive timing, the aim is to investigate production rather than perception of timing, that is, independently of the listene
Beat Extraction from Expressive Musical Performances Simon Dixon, Werner Goebl and Emilios Cambouropoulos Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-1010 Vienna, Austria.
More informationAutomatic music transcription
Music transcription 1 Music transcription 2 Automatic music transcription Sources: * Klapuri, Introduction to music transcription, 2006. www.cs.tut.fi/sgn/arg/klap/amt-intro.pdf * Klapuri, Eronen, Astola:
More informationMAutoPitch. Presets button. Left arrow button. Right arrow button. Randomize button. Save button. Panic button. Settings button
MAutoPitch Presets button Presets button shows a window with all available presets. A preset can be loaded from the preset window by double-clicking on it, using the arrow buttons or by using a combination
More informationAE16 DIGITAL AUDIO WORKSTATIONS
AE16 DIGITAL AUDIO WORKSTATIONS 1. Storage Requirements In a conventional linear PCM system without data compression the data rate (bits/sec) from one channel of digital audio will depend on the sampling
More informationA Matlab toolbox for. Characterisation Of Recorded Underwater Sound (CHORUS) USER S GUIDE
Centre for Marine Science and Technology A Matlab toolbox for Characterisation Of Recorded Underwater Sound (CHORUS) USER S GUIDE Version 5.0b Prepared for: Centre for Marine Science and Technology Prepared
More informationRobert Alexandru Dobre, Cristian Negrescu
ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Automatic Music Transcription Software Based on Constant Q
More informationDrum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods
Drum Sound Identification for Polyphonic Music Using Template Adaptation and Matching Methods Kazuyoshi Yoshii, Masataka Goto and Hiroshi G. Okuno Department of Intelligence Science and Technology National
More informationTHE DIGITAL DELAY ADVANTAGE A guide to using Digital Delays. Synchronize loudspeakers Eliminate comb filter distortion Align acoustic image.
THE DIGITAL DELAY ADVANTAGE A guide to using Digital Delays Synchronize loudspeakers Eliminate comb filter distortion Align acoustic image Contents THE DIGITAL DELAY ADVANTAGE...1 - Why Digital Delays?...
More information6.UAP Project. FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System. Daryl Neubieser. May 12, 2016
6.UAP Project FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System Daryl Neubieser May 12, 2016 Abstract: This paper describes my implementation of a variable-speed accompaniment system that
More informationECE 4220 Real Time Embedded Systems Final Project Spectrum Analyzer
ECE 4220 Real Time Embedded Systems Final Project Spectrum Analyzer by: Matt Mazzola 12222670 Abstract The design of a spectrum analyzer on an embedded device is presented. The device achieves minimum
More informationTASCAM DM-24. The DM-24 Basics. TEAC Professional Division. Digital Mixing console
TASCAM TEAC Professional Division DM-24 Digital Mixing console The DM-24 Basics DM-24 SIGNAL FLOW... 3 INPUTS... 3 RETURNS... 3 OPTIONS... 4 OUTPUTS... 5 AUX SENDS... 5 TRACKING OPTIONS... 5 Using AUX
More informationAN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY
AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY Eugene Mikyung Kim Department of Music Technology, Korea National University of Arts eugene@u.northwestern.edu ABSTRACT
More informationRECORDING AND REPRODUCING CONCERT HALL ACOUSTICS FOR SUBJECTIVE EVALUATION
RECORDING AND REPRODUCING CONCERT HALL ACOUSTICS FOR SUBJECTIVE EVALUATION Reference PACS: 43.55.Mc, 43.55.Gx, 43.38.Md Lokki, Tapio Aalto University School of Science, Dept. of Media Technology P.O.Box
More informationOBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES
OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES Vishweshwara Rao and Preeti Rao Digital Audio Processing Lab, Electrical Engineering Department, IIT-Bombay, Powai,
More informationni.com Digital Signal Processing for Every Application
Digital Signal Processing for Every Application Digital Signal Processing is Everywhere High-Volume Image Processing Production Test Structural Sound Health and Vibration Monitoring RF WiMAX, and Microwave
More informationIntroduction 3/5/13 2
Mixing 3/5/13 1 Introduction Audio mixing is used for sound recording, audio editing and sound systems to balance the relative volume, frequency and dynamical content of a number of sound sources. Typically,
More informationDetection of Panoramic Takes in Soccer Videos Using Phase Correlation and Boosting
Detection of Panoramic Takes in Soccer Videos Using Phase Correlation and Boosting Luiz G. L. B. M. de Vasconcelos Research & Development Department Globo TV Network Email: luiz.vasconcelos@tvglobo.com.br
More informationHENRY AWARDS EVALUATOR EXPECTATIONS
PERFORMANCE CATEGORIES (Ensemble, Lead/Supporting Actor/Actress) ACTING SINGING (IF APPLICABLE) MOVEMENT STAGE PRESENCE Basic understanding of character Can be heard and understood Ability to sing on pitch
More informationQUALITY OF COMPUTER MUSIC USING MIDI LANGUAGE FOR DIGITAL MUSIC ARRANGEMENT
QUALITY OF COMPUTER MUSIC USING MIDI LANGUAGE FOR DIGITAL MUSIC ARRANGEMENT Pandan Pareanom Purwacandra 1, Ferry Wahyu Wibowo 2 Informatics Engineering, STMIK AMIKOM Yogyakarta 1 pandanharmony@gmail.com,
More informationTHE importance of music content analysis for musical
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 15, NO. 1, JANUARY 2007 333 Drum Sound Recognition for Polyphonic Audio Signals by Adaptation and Matching of Spectrogram Templates With
More informationCalibration of auralisation presentations through loudspeakers
Calibration of auralisation presentations through loudspeakers Jens Holger Rindel, Claus Lynge Christensen Odeon A/S, Scion-DTU, DK-2800 Kgs. Lyngby, Denmark. jhr@odeon.dk Abstract The correct level of
More informationA prototype system for rule-based expressive modifications of audio recordings
International Symposium on Performance Science ISBN 0-00-000000-0 / 000-0-00-000000-0 The Author 2007, Published by the AEC All rights reserved A prototype system for rule-based expressive modifications
More informationinter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE
Copyright SFA - InterNoise 2000 1 inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering 27-30 August 2000, Nice, FRANCE I-INCE Classification: 7.9 THE FUTURE OF SOUND
More informationMusical Hit Detection
Musical Hit Detection CS 229 Project Milestone Report Eleanor Crane Sarah Houts Kiran Murthy December 12, 2008 1 Problem Statement Musical visualizers are programs that process audio input in order to
More informationDIGITAL COMMUNICATION
10EC61 DIGITAL COMMUNICATION UNIT 3 OUTLINE Waveform coding techniques (continued), DPCM, DM, applications. Base-Band Shaping for Data Transmission Discrete PAM signals, power spectra of discrete PAM signals.
More informationUsing the new psychoacoustic tonality analyses Tonality (Hearing Model) 1
02/18 Using the new psychoacoustic tonality analyses 1 As of ArtemiS SUITE 9.2, a very important new fully psychoacoustic approach to the measurement of tonalities is now available., based on the Hearing
More informationA Composition for Clarinet and Real-Time Signal Processing: Using Max on the IRCAM Signal Processing Workstation
A Composition for Clarinet and Real-Time Signal Processing: Using Max on the IRCAM Signal Processing Workstation Cort Lippe IRCAM, 31 rue St-Merri, Paris, 75004, France email: lippe@ircam.fr Introduction.
More informationNew recording techniques for solo double bass
New recording techniques for solo double bass Cato Langnes NOTAM, Sandakerveien 24 D, Bygg F3, 0473 Oslo catola@notam02.no, www.notam02.no Abstract This paper summarizes techniques utilized in the process
More informationMusic Radar: A Web-based Query by Humming System
Music Radar: A Web-based Query by Humming System Lianjie Cao, Peng Hao, Chunmeng Zhou Computer Science Department, Purdue University, 305 N. University Street West Lafayette, IN 47907-2107 {cao62, pengh,
More informationComputational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)
Computational Models of Music Similarity 1 Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST) Abstract The perceived similarity of two pieces of music is multi-dimensional,
More informationAudio-Based Video Editing with Two-Channel Microphone
Audio-Based Video Editing with Two-Channel Microphone Tetsuya Takiguchi Organization of Advanced Science and Technology Kobe University, Japan takigu@kobe-u.ac.jp Yasuo Ariki Organization of Advanced Science
More informationTechnical Guide. Installed Sound. Loudspeaker Solutions for Worship Spaces. TA-4 Version 1.2 April, Why loudspeakers at all?
Installed Technical Guide Loudspeaker Solutions for Worship Spaces TA-4 Version 1.2 April, 2002 systems for worship spaces can be a delight for all listeners or the horror of the millennium. The loudspeaker
More informationI. LISTENING. For most people, sound is background only. To the sound designer/producer, sound is everything.!tc 243 2
To use sound properly, and fully realize its power, we need to do the following: (1) listen (2) understand basics of sound and hearing (3) understand sound's fundamental effects on human communication
More informationSemi-automated extraction of expressive performance information from acoustic recordings of piano music. Andrew Earis
Semi-automated extraction of expressive performance information from acoustic recordings of piano music Andrew Earis Outline Parameters of expressive piano performance Scientific techniques: Fourier transform
More informationFigure 1: Feature Vector Sequence Generator block diagram.
1 Introduction Figure 1: Feature Vector Sequence Generator block diagram. We propose designing a simple isolated word speech recognition system in Verilog. Our design is naturally divided into two modules.
More informationDTS Neural Mono2Stereo
WAVES DTS Neural Mono2Stereo USER GUIDE Table of Contents Chapter 1 Introduction... 3 1.1 Welcome... 3 1.2 Product Overview... 3 1.3 Sample Rate Support... 4 Chapter 2 Interface and Controls... 5 2.1 Interface...
More informationInvestigation of Digital Signal Processing of High-speed DACs Signals for Settling Time Testing
Universal Journal of Electrical and Electronic Engineering 4(2): 67-72, 2016 DOI: 10.13189/ujeee.2016.040204 http://www.hrpub.org Investigation of Digital Signal Processing of High-speed DACs Signals for
More informationLX20 OPERATORS MANUAL
LX20 OPERATORS MANUAL CONTENTS SAFETY CONSIDERATIONS page 1 INSTALLATION page 2 INTRODUCTION page 2 FIRST TIME USER page 3 SYSTEM OPERATING LEVELS page 3 FRONT & REAR PANEL LAYOUT page 4 OPERATION page
More informationEMERGENT SOUNDSCAPE COMPOSITION: REFLECTIONS ON VIRTUALITY
EMERGENT SOUNDSCAPE COMPOSITION: REFLECTIONS ON VIRTUALITY by Mark Christopher Brady Bachelor of Science (Honours), University of Cape Town, 1994 THESIS SUBMITTED IN PARTIAL FULFILLMENT OF THE REQUIREMENTS
More informationSimple Harmonic Motion: What is a Sound Spectrum?
Simple Harmonic Motion: What is a Sound Spectrum? A sound spectrum displays the different frequencies present in a sound. Most sounds are made up of a complicated mixture of vibrations. (There is an introduction
More informationAutomatic Generation of Drum Performance Based on the MIDI Code
Automatic Generation of Drum Performance Based on the MIDI Code Shigeki SUZUKI Mamoru ENDO Masashi YAMADA and Shinya MIYAZAKI Graduate School of Computer and Cognitive Science, Chukyo University 101 tokodachi,
More informationDigital Strobe Tuner. w/ On stage Display
Page 1/7 # Guys EEL 4924 Electrical Engineering Design (Senior Design) Digital Strobe Tuner w/ On stage Display Team Members: Name: David Barnette Email: dtbarn@ufl.edu Phone: 850-217-9147 Name: Jamie
More informationWhite Paper : Achieving synthetic slow-motion in UHDTV. InSync Technology Ltd, UK
White Paper : Achieving synthetic slow-motion in UHDTV InSync Technology Ltd, UK ABSTRACT High speed cameras used for slow motion playback are ubiquitous in sports productions, but their high cost, and
More informationHugo Technology. An introduction into Rob Watts' technology
Hugo Technology An introduction into Rob Watts' technology Copyright Rob Watts 2014 About Rob Watts Audio chip designer both analogue and digital Consultant to silicon chip manufacturers Designer of Chord
More informationDepartment of Electrical & Electronic Engineering Imperial College of Science, Technology and Medicine. Project: Real-Time Speech Enhancement
Department of Electrical & Electronic Engineering Imperial College of Science, Technology and Medicine Project: Real-Time Speech Enhancement Introduction Telephones are increasingly being used in noisy
More informationPSYCHOACOUSTICS & THE GRAMMAR OF AUDIO (By Steve Donofrio NATF)
PSYCHOACOUSTICS & THE GRAMMAR OF AUDIO (By Steve Donofrio NATF) "The reason I got into playing and producing music was its power to travel great distances and have an emotional impact on people" Quincey
More informationAuthors: Kasper Marklund, Anders Friberg, Sofia Dahl, KTH, Carlo Drioli, GEM, Erik Lindström, UUP Last update: November 28, 2002
Groove Machine Authors: Kasper Marklund, Anders Friberg, Sofia Dahl, KTH, Carlo Drioli, GEM, Erik Lindström, UUP Last update: November 28, 2002 1. General information Site: Kulturhuset-The Cultural Centre
More informationMusic Source Separation
Music Source Separation Hao-Wei Tseng Electrical and Engineering System University of Michigan Ann Arbor, Michigan Email: blakesen@umich.edu Abstract In popular music, a cover version or cover song, or
More informationRadio for Everyone...
Radio for Everyone... P R O D U C T I O N O N A I R C O N S O L E Eight dual inputs Built in auto Silence detector 4 USB in/out stereo channels Play out USB control section included AES 3 digital program
More informationTempo and Beat Analysis
Advanced Course Computer Science Music Processing Summer Term 2010 Meinard Müller, Peter Grosche Saarland University and MPI Informatik meinard@mpi-inf.mpg.de Tempo and Beat Analysis Musical Properties:
More informationVoice Controlled Car System
Voice Controlled Car System 6.111 Project Proposal Ekin Karasan & Driss Hafdi November 3, 2016 1. Overview Voice controlled car systems have been very important in providing the ability to drivers to adjust
More informationHEAD. HEAD VISOR (Code 7500ff) Overview. Features. System for online localization of sound sources in real time
HEAD Ebertstraße 30a 52134 Herzogenrath Tel.: +49 2407 577-0 Fax: +49 2407 577-99 email: info@head-acoustics.de Web: www.head-acoustics.de Data Datenblatt Sheet HEAD VISOR (Code 7500ff) System for online
More informationCS229 Project Report Polyphonic Piano Transcription
CS229 Project Report Polyphonic Piano Transcription Mohammad Sadegh Ebrahimi Stanford University Jean-Baptiste Boin Stanford University sadegh@stanford.edu jbboin@stanford.edu 1. Introduction In this project
More informationComputer Coordination With Popular Music: A New Research Agenda 1
Computer Coordination With Popular Music: A New Research Agenda 1 Roger B. Dannenberg roger.dannenberg@cs.cmu.edu http://www.cs.cmu.edu/~rbd School of Computer Science Carnegie Mellon University Pittsburgh,
More informationLinrad On-Screen Controls K1JT
Linrad On-Screen Controls K1JT Main (Startup) Menu A = Weak signal CW B = Normal CW C = Meteor scatter CW D = SSB E = FM F = AM G = QRSS CW H = TX test I = Soundcard test mode J = Analog hardware tune
More informationAPPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC
APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC Vishweshwara Rao, Sachin Pant, Madhumita Bhaskar and Preeti Rao Department of Electrical Engineering, IIT Bombay {vishu, sachinp,
More informationMelody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng
Melody Extraction from Generic Audio Clips Thaminda Edirisooriya, Hansohl Kim, Connie Zeng Introduction In this project we were interested in extracting the melody from generic audio files. Due to the
More informationHidden melody in music playing motion: Music recording using optical motion tracking system
PROCEEDINGS of the 22 nd International Congress on Acoustics General Musical Acoustics: Paper ICA2016-692 Hidden melody in music playing motion: Music recording using optical motion tracking system Min-Ho
More informationA HIGHLY INTERACTIVE SYSTEM FOR PROCESSING LARGE VOLUMES OF ULTRASONIC TESTING DATA. H. L. Grothues, R. H. Peterson, D. R. Hamlin, K. s.
A HIGHLY INTERACTIVE SYSTEM FOR PROCESSING LARGE VOLUMES OF ULTRASONIC TESTING DATA H. L. Grothues, R. H. Peterson, D. R. Hamlin, K. s. Pickens Southwest Research Institute San Antonio, Texas INTRODUCTION
More informationAdvance Certificate Course In Audio Mixing & Mastering.
Advance Certificate Course In Audio Mixing & Mastering. CODE: SIA-ACMM16 For Whom: Budding Composers/ Music Producers. Assistant Engineers / Producers Working Engineers. Anyone, who has done the basic
More informationMultiband Noise Reduction Component for PurePath Studio Portable Audio Devices
Multiband Noise Reduction Component for PurePath Studio Portable Audio Devices Audio Converters ABSTRACT This application note describes the features, operating procedures and control capabilities of a
More informationPHYSICS OF MUSIC. 1.) Charles Taylor, Exploring Music (Music Library ML3805 T )
REFERENCES: 1.) Charles Taylor, Exploring Music (Music Library ML3805 T225 1992) 2.) Juan Roederer, Physics and Psychophysics of Music (Music Library ML3805 R74 1995) 3.) Physics of Sound, writeup in this
More informationMUSIC TRANSCRIBER. Overall System Description. Alessandro Yamhure 11/04/2005
Roberto Carli 6.111 Project Proposal MUSIC TRANSCRIBER Overall System Description The aim of this digital system is to convert music played into the correct sheet music. We are basically implementing a
More informationDAT335 Music Perception and Cognition Cogswell Polytechnical College Spring Week 6 Class Notes
DAT335 Music Perception and Cognition Cogswell Polytechnical College Spring 2009 Week 6 Class Notes Pitch Perception Introduction Pitch may be described as that attribute of auditory sensation in terms
More informationImplementation of an MPEG Codec on the Tilera TM 64 Processor
1 Implementation of an MPEG Codec on the Tilera TM 64 Processor Whitney Flohr Supervisor: Mark Franklin, Ed Richter Department of Electrical and Systems Engineering Washington University in St. Louis Fall
More informationSREV1 Sampling Guide. An Introduction to Impulse-response Sampling with the SREV1 Sampling Reverberator
An Introduction to Impulse-response Sampling with the SREV Sampling Reverberator Contents Introduction.............................. 2 What is Sound Field Sampling?.....................................
More informationTopic 10. Multi-pitch Analysis
Topic 10 Multi-pitch Analysis What is pitch? Common elements of music are pitch, rhythm, dynamics, and the sonic qualities of timbre and texture. An auditory perceptual attribute in terms of which sounds
More informationUNIVERSITY OF DUBLIN TRINITY COLLEGE
UNIVERSITY OF DUBLIN TRINITY COLLEGE FACULTY OF ENGINEERING & SYSTEMS SCIENCES School of Engineering and SCHOOL OF MUSIC Postgraduate Diploma in Music and Media Technologies Hilary Term 31 st January 2005
More informationStudy of White Gaussian Noise with Varying Signal to Noise Ratio in Speech Signal using Wavelet
American International Journal of Research in Science, Technology, Engineering & Mathematics Available online at http://www.iasir.net ISSN (Print): 2328-3491, ISSN (Online): 2328-3580, ISSN (CD-ROM): 2328-3629
More informationNON-UNIFORM KERNEL SAMPLING IN AUDIO SIGNAL RESAMPLER
NON-UNIFORM KERNEL SAMPLING IN AUDIO SIGNAL RESAMPLER Grzegorz Kraszewski Białystok Technical University, Electrical Engineering Faculty, ul. Wiejska 45D, 15-351 Białystok, Poland, e-mail: krashan@teleinfo.pb.bialystok.pl
More informationInstrument Recognition in Polyphonic Mixtures Using Spectral Envelopes
Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes hello Jay Biernat Third author University of Rochester University of Rochester Affiliation3 words jbiernat@ur.rochester.edu author3@ismir.edu
More informationA repetition-based framework for lyric alignment in popular songs
A repetition-based framework for lyric alignment in popular songs ABSTRACT LUONG Minh Thang and KAN Min Yen Department of Computer Science, School of Computing, National University of Singapore We examine
More informationTOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC
TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu
More informationBrowsing News and Talk Video on a Consumer Electronics Platform Using Face Detection
Browsing News and Talk Video on a Consumer Electronics Platform Using Face Detection Kadir A. Peker, Ajay Divakaran, Tom Lanning Mitsubishi Electric Research Laboratories, Cambridge, MA, USA {peker,ajayd,}@merl.com
More informationMeasurement of overtone frequencies of a toy piano and perception of its pitch
Measurement of overtone frequencies of a toy piano and perception of its pitch PACS: 43.75.Mn ABSTRACT Akira Nishimura Department of Media and Cultural Studies, Tokyo University of Information Sciences,
More informationFPFV-285/585 PRODUCTION SOUND Fall 2018 CRITICAL LISTENING Assignment
FPFV-285/585 PRODUCTION SOUND Fall 2018 CRITICAL LISTENING Assignment PREPARATION Track 1) Headphone check -- Left, Right, Left, Right. Track 2) A music excerpt for setting comfortable listening level.
More informationJOURNAL OF BUILDING ACOUSTICS. Volume 20 Number
Early and Late Support Measured over Various Distances: The Covered versus Open Part of the Orchestra Pit by R.H.C. Wenmaekers and C.C.J.M. Hak Reprinted from JOURNAL OF BUILDING ACOUSTICS Volume 2 Number
More informationVISUALIZING AND CONTROLLING SOUND WITH GRAPHICAL INTERFACES
VISUALIZING AND CONTROLLING SOUND WITH GRAPHICAL INTERFACES LIAM O SULLIVAN, FRANK BOLAND Dept. of Electronic & Electrical Engineering, Trinity College Dublin, Dublin 2, Ireland lmosulli@tcd.ie Developments
More informationToward a Computationally-Enhanced Acoustic Grand Piano
Toward a Computationally-Enhanced Acoustic Grand Piano Andrew McPherson Electrical & Computer Engineering Drexel University 3141 Chestnut St. Philadelphia, PA 19104 USA apm@drexel.edu Youngmoo Kim Electrical
More informationGuitar and Rock/Blues Vocalists
Addendum A, Page 1 to: Guitar and Rock/Blues Vocalists Guitar players and Rock/Blues vocalists share a similar part of the stage and as such, are similarly exposed to loud music. Some of the strategies
More informationFFT Laboratory Experiments for the HP Series Oscilloscopes and HP 54657A/54658A Measurement Storage Modules
FFT Laboratory Experiments for the HP 54600 Series Oscilloscopes and HP 54657A/54658A Measurement Storage Modules By: Michael W. Thompson, PhD. EE Dept. of Electrical Engineering Colorado State University
More informationDETECTING ENVIRONMENTAL NOISE WITH BASIC TOOLS
DETECTING ENVIRONMENTAL NOISE WITH BASIC TOOLS By Henrik, September 2018, Version 2 Measuring low-frequency components of environmental noise close to the hearing threshold with high accuracy requires
More informationPS User Guide Series Seismic-Data Display
PS User Guide Series 2015 Seismic-Data Display Prepared By Choon B. Park, Ph.D. January 2015 Table of Contents Page 1. File 2 2. Data 2 2.1 Resample 3 3. Edit 4 3.1 Export Data 4 3.2 Cut/Append Records
More informationCHAPTER 3 AUDIO MIXER DIGITAL AUDIO PRODUCTION [IP3038PA]
CHAPTER 3 AUDIO MIXER DIGITAL AUDIO PRODUCTION [IP3038PA] Learning Objectives By the end of this chapter, students should be able to: 1 State the function of the audio mixer in the sound studio. 2 Explain
More informationMULTIMIX 8/4 DIGITAL AUDIO-PROCESSING
MULTIMIX 8/4 DIGITAL AUDIO-PROCESSING Designed and Manufactured by ITEC Tontechnik und Industrieelektronik GesmbH 8200 Laßnitzthal 300 Austria / Europe MULTIMIX 8/4 DIGITAL Aim The most important aim of
More informationLiquid Mix Plug-in. User Guide FA
Liquid Mix Plug-in User Guide FA0000-01 1 1. COMPRESSOR SECTION... 3 INPUT LEVEL...3 COMPRESSOR EMULATION SELECT...3 COMPRESSOR ON...3 THRESHOLD...3 RATIO...4 COMPRESSOR GRAPH...4 GAIN REDUCTION METER...5
More informationVIDEO JUDGE SYSTEM SETUP & CAPTURE
VIDEO JUDGE SYSTEM SETUP & CAPTURE TABLE OF CONTENTS GENERAL OVERVIEW... 1 ABOUT THE COMPETITIONS... 1 PRIOR TO THE EVENT... 2 EQUIPMENT LIST... 2 ARRIVAL AT THE VENUE... 3 EQUIPMENT SETUP... 4 Camera
More informationVCE VET MUSIC INDUSTRY: SOUND PRODUCTION
Victorian Certificate of Education 2017 SUPERVISOR TO ATTACH PROCESSING LABEL HERE Letter STUDENT NUMBER VCE VET MUSIC INDUSTRY: SOUND PRODUCTION Aural and written examination Friday 17 November 2017 Reading
More informationConvention Paper Presented at the 139th Convention 2015 October 29 November 1 New York, USA
Audio Engineering Society Convention Paper Presented at the 139th Convention 215 October 29 November 1 New York, USA This Convention paper was selected based on a submitted abstract and 75-word precis
More informationAcoustic Instrument Message Specification
Acoustic Instrument Message Specification v 0.4 Proposal June 15, 2014 Keith McMillen Instruments BEAM Foundation Created by: Keith McMillen - keith@beamfoundation.org With contributions from : Barry Threw
More informationAphro-V1 Digital reverb & fx processor..
Aphro-V1 Digital reverb & fx processor.. Copyright all rights reserved 1998, 1999. Audio Mechanic & Sound Breeder page 1 Summary Specifications p 3 Introduction p 4 Main Interface p 5 LCD Display p 5 Interfaces
More informationVirtual Vibration Analyzer
Virtual Vibration Analyzer Vibration/industrial systems LabVIEW DAQ by Ricardo Jaramillo, Manager, Ricardo Jaramillo y Cía; Daniel Jaramillo, Engineering Assistant, Ricardo Jaramillo y Cía The Challenge:
More informationMajor Differences Between the DT9847 Series Modules
DT9847 Series Dynamic Signal Analyzer for USB With Low THD and Wide Dynamic Range The DT9847 Series are high-accuracy, dynamic signal acquisition modules designed for sound and vibration applications.
More informationMusic Complexity Descriptors. Matt Stabile June 6 th, 2008
Music Complexity Descriptors Matt Stabile June 6 th, 2008 Musical Complexity as a Semantic Descriptor Modern digital audio collections need new criteria for categorization and searching. Applicable to:
More informationPITZ Introduction to the Video System
PITZ Introduction to the Video System Stefan Weiße DESY Zeuthen June 10, 2003 Agenda 1. Introduction to PITZ 2. Why a video system? 3. Schematic structure 4. Client/Server architecture 5. Hardware 6. Software
More informationPiotr KLECZKOWSKI, Magdalena PLEWA, Grzegorz PYDA
ARCHIVES OF ACOUSTICS 33, 4 (Supplement), 147 152 (2008) LOCALIZATION OF A SOUND SOURCE IN DOUBLE MS RECORDINGS Piotr KLECZKOWSKI, Magdalena PLEWA, Grzegorz PYDA AGH University od Science and Technology
More information