A SIMPLE ACOUSTIC ROOM MODEL FOR VIRTUAL PRODUCTION AUDIO. R. Walker. British Broadcasting Corporation, United Kingdom. ABSTRACT

Similar documents
The Cocktail Party Effect. Binaural Masking. The Precedence Effect. Music 175: Time and Space

A few white papers on various. Digital Signal Processing algorithms. used in the DAC501 / DAC502 units

How to Obtain a Good Stereo Sound Stage in Cars

Hugo Technology. An introduction into Rob Watts' technology

Investigation of Digital Signal Processing of High-speed DACs Signals for Settling Time Testing

Digital Audio: Some Myths and Realities

Studies for Future Broadcasting Services and Basic Technologies

Measurement of overtone frequencies of a toy piano and perception of its pitch

StepArray+ Self-powered digitally steerable column loudspeakers

RECORDING AND REPRODUCING CONCERT HALL ACOUSTICS FOR SUBJECTIVE EVALUATION

RECOMMENDATION ITU-R BT (Questions ITU-R 25/11, ITU-R 60/11 and ITU-R 61/11)

Acoustic concert halls (Statistical calculation, wave acoustic theory with reference to reconstruction of Saint- Petersburg Kapelle and philharmonic)

A typical example: front left subwoofer only. Four subwoofers with Sound Field Management. A Direct Comparison

D. BARD, J. NEGREIRA DIVISION OF ENGINEERING ACOUSTICS, LUND UNIVERSITY

Aphro-V1 Digital reverb & fx processor..

DESIGNING OPTIMIZED MICROPHONE BEAMFORMERS

Laboratory Assignment 3. Digital Music Synthesis: Beethoven s Fifth Symphony Using MATLAB

A New "Duration-Adapted TR" Waveform Capture Method Eliminates Severe Limitations

THE DIGITAL DELAY ADVANTAGE A guide to using Digital Delays. Synchronize loudspeakers Eliminate comb filter distortion Align acoustic image.

DH400. Digital Phone Hybrid. The most advanced Digital Hybrid with DSP echo canceller and VQR technology.

TEPZZ A_T EP A1 (19) (11) EP A1 (12) EUROPEAN PATENT APPLICATION. (51) Int Cl.: H04S 7/00 ( ) H04R 25/00 (2006.

TEPZZ 94 98_A_T EP A1 (19) (11) EP A1 (12) EUROPEAN PATENT APPLICATION. (43) Date of publication: Bulletin 2015/46

4 MHz Lock-In Amplifier

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE

Multiband Noise Reduction Component for PurePath Studio Portable Audio Devices

Musical Acoustics Lecture 15 Pitch & Frequency (Psycho-Acoustics)

Open loop tracking of radio occultation signals in the lower troposphere

UNIVERSITY OF DUBLIN TRINITY COLLEGE

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE

RECOMMENDATION ITU-R BT.1201 * Extremely high resolution imagery

PRACTICAL APPLICATION OF THE PHASED-ARRAY TECHNOLOGY WITH PAINT-BRUSH EVALUATION FOR SEAMLESS-TUBE TESTING

FPFV-285/585 PRODUCTION SOUND Fall 2018 CRITICAL LISTENING Assignment

White Paper JBL s LSR Principle, RMC (Room Mode Correction) and the Monitoring Environment by John Eargle. Introduction and Background:

SREV1 Sampling Guide. An Introduction to Impulse-response Sampling with the SREV1 Sampling Reverberator

PSYCHOACOUSTICS & THE GRAMMAR OF AUDIO (By Steve Donofrio NATF)

Practical Application of the Phased-Array Technology with Paint-Brush Evaluation for Seamless-Tube Testing

Witold MICKIEWICZ, Jakub JELEŃ

The interaction between room and musical instruments studied by multi-channel auralization

2. AN INTROSPECTION OF THE MORPHING PROCESS

ECE438 - Laboratory 4: Sampling and Reconstruction of Continuous-Time Signals

Clock Jitter Cancelation in Coherent Data Converter Testing

2 MHz Lock-In Amplifier

Notes on Digital Circuits

Methods to measure stage acoustic parameters: overview and future research

PS User Guide Series Seismic-Data Display

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE

OBJECT-AUDIO CAPTURE SYSTEM FOR SPORTS BROADCAST

FX Basics. Time Effects STOMPBOX DESIGN WORKSHOP. Esteban Maestre. CCRMA Stanford University July 2011

Mixing in the Box A detailed look at some of the myths and legends surrounding Pro Tools' mix bus.

Getting Started with the LabVIEW Sound and Vibration Toolkit

Simple Harmonic Motion: What is a Sound Spectrum?

Noise Detector ND-1 Operating Manual

ON THE INTERPOLATION OF ULTRASONIC GUIDED WAVE SIGNALS

MIE 402: WORKSHOP ON DATA ACQUISITION AND SIGNAL PROCESSING Spring 2003

Using the BHM binaural head microphone

THE PSYCHOACOUSTICS OF MULTICHANNEL AUDIO. J. ROBERT STUART Meridian Audio Ltd Stonehill, Huntingdon, PE18 6ED England

Stabilising stereo images Michael Gerzon

SPATIAL LIGHT MODULATORS

CLASSROOM ACOUSTICS OF MCNEESE STATE UNIVER- SITY

Chapter 6: Real-Time Image Formation

Space Optimisation and Space Optimisation

Room acoustics computer modelling: Study of the effect of source directivity on auralizations

White Paper Measuring and Optimizing Sound Systems: An introduction to JBL Smaart

EFFECTS OF REVERBERATION TIME AND SOUND SOURCE CHARACTERISTIC TO AUDITORY LOCALIZATION IN AN INDOOR SOUND FIELD. Chiung Yao Chen

Department of Electrical & Electronic Engineering Imperial College of Science, Technology and Medicine. Project: Real-Time Speech Enhancement

Proceedings of Meetings on Acoustics

Rec. ITU-R BT RECOMMENDATION ITU-R BT * WIDE-SCREEN SIGNALLING FOR BROADCASTING

IP Telephony and Some Factors that Influence Speech Quality

Proceedings of Meetings on Acoustics

DTS Neural Mono2Stereo

DELTA MODULATION AND DPCM CODING OF COLOR SIGNALS

Operation Manual OPERATION MANUAL ISL. Precision True Peak Limiter NUGEN Audio. Contents

Notes on Digital Circuits

Calibrate, Characterize and Emulate Systems Using RFXpress in AWG Series

NanoGiant Oscilloscope/Function-Generator Program. Getting Started

Collection of Setups for Measurements with the R&S UPV and R&S UPP Audio Analyzers. Application Note. Products:

ACTIVE SOUND DESIGN: VACUUM CLEANER

Journal of Theoretical and Applied Information Technology 20 th July Vol. 65 No JATIT & LLS. All rights reserved.

Introduction to Data Conversion and Processing

CM3106 Solutions. Do not turn this page over until instructed to do so by the Senior Invigilator.

TV Synchronism Generation with PIC Microcontroller

Synthesized Clock Generator

Build Applications Tailored for Remote Signal Monitoring with the Signal Hound BB60C

Binaural Measurement, Analysis and Playback

Techniques for Extending Real-Time Oscilloscope Bandwidth

UNITED STATES PATENT AND TRADEMARK OFFICE BEFORE THE PATENT TRIAL AND APPEAL BOARD

Color Reproduction Complex

AE16 DIGITAL AUDIO WORKSTATIONS

RECOMMENDATION ITU-R BT Studio encoding parameters of digital television for standard 4:3 and wide-screen 16:9 aspect ratios

PCM ENCODING PREPARATION... 2 PCM the PCM ENCODER module... 4

Ch. 1: Audio/Image/Video Fundamentals Multimedia Systems. School of Electrical Engineering and Computer Science Oregon State University

USB-TG124A Tracking Generator User Manual

SC26 Magnetic Field Cancelling System

Cathedral user guide & reference manual

Using the new psychoacoustic tonality analyses Tonality (Hearing Model) 1

ENGINEERING COMMITTEE

VTX V25-II Preset Guide

Intelligent Monitoring Software IMZ-RS300. Series IMZ-RS301 IMZ-RS304 IMZ-RS309 IMZ-RS316 IMZ-RS332 IMZ-RS300C

OPERA APPLICATION NOTES (1)

A detailed discussion of echo cancellation methods and the testing results follows.

Transcription:

A SIMPLE ACOUSTIC ROOM MODEL FOR VIRTUAL PRODUCTION AUDIO. R. Walker British Broadcasting Corporation, United Kingdom. ABSTRACT The use of television virtual production is becoming commonplace. This paper describes the principles of a simple, experimental audio system, developed to accompany a video virtual production system. The acoustic model included the direct sound, six first-order room boundary reflections, reverberation, source directivity, air and surface absorption and obstruction or reflection of the direct sound by one internal, flat object within the room, with diffraction effects around the obstacle. The model was implemented using a commercial dsp development platform and was designed for 5.0 multichannel reproduction. A simplified, two-channel stereophonic version was also implemented. The system used three, 56000 dsp processors, each running at 40 MHz clock rate. The audio sample rate was 48 khz. The final system update rate, for a complete recalculation of all of the directional sound components including obstruction or reflection, was about 400 Hz. The system included manual controls for testing and development, as well as a TCP/IP network interface for control by the video processor. INTRODUCTION In television production, the use of syntheticallygenerated visual studio features is becoming commonplace. It has obvious potential for saving the costs of constructing and moving studio sets and can also allow studios to appear to be of a different size to their actual size. The technique is generally known as Virtual Production (VP). SYNTHETIC VISUAL IMAGES The creation of a VP visual image involves two main components. The studio (or, more accurately, the live performance space, since it may not be anything like a conventional studio) must exist in reality. It provides the space in which the live action can take place. The second main component is a computer model of the virtual studio. In many cases, there will be elements common to the real and virtual spaces. Objects that are located inside the space, rather than on the perimeter surfaces, will usually have to exist in both, at least in order to provide cues for the performers. The link between the real and virtual studios is provided by a computer system, which generates the appropriate synthetic view of the virtual studio, from inputs defining the location and orientation of the cameras and the positions of moving objects. THE HUMAN HEARING SYSTEM The human hearing system is not well enough understood to provide a satisfactory model on which to base the creation of completely accurate and foolproof artificial direction, distance and environment cues. It is well adapted to (and therefore highly sensitive at) obtaining spatial cues from the acoustic early reflection pattern. It is also sensitive to the overall acoustic impression of the whole space. These aspects of hearing have been widely studied. For very close sources, of the order of 2 m distance, a typical early reflection pattern is short, with close-spaced arrival times. The temporal discrimination for those short times, less than 5 ms, is very poor (almost zero). The reflected sounds interfere with the direct component to cause effects that are generally perceived in the frequency domain - as disturbances to the spectral balance. That may be irrelevant if the source spectrum is unknown, but is still likely to cause perceptible effects if the spectral balance changes significantly. After about 5 ms relative to the arrival of the direct sound, the hearing system becomes sensitive to individual reflections, though they are not perceived as discrete echoes. After about 50 ms, all of the remaining sound energy is summed into the general reverberation pattern, representing the surrounding space on a

global scale. There is virtually no sensitivity to the detail of the later reflection pattern with the exception of echoes. If an individual reflection has a significantly higher level than the general reverberation at that time, it is perceived as a discrete echo. ACOUSTICS OF ROOMS AND ACOUSTICAL COMPLEXITY Sound emanating from a source travels outwards from the source at the characteristic velocity of sound ( 340 m/s). The relative intensities in particular directions are governed by the radiation pattern of the source, which is quite likely to be non-uniform. Because of the spreading loss, the sound intensity in a particular direction decreases at a rate of 6 db per doubling of distance. In any partially or fully enclosed space, that uniform spreading proceeds for only a short time, until part of the sound wave strikes some acoustically significant object. What then happens is always complicated. Sound propagates as a wave function and demonstrates all of the properties usually associated with the interactions of waves and objects reflection, refraction and absorption. What happens when a sound wave meets a discontinuity in the medium depends on the acoustic properties of the boundary materials and the size of the discontinuity in relation to the wavelength of the sound wave. Over the normal audio frequency span, wavelengths range from about 15 mm to 7 m. That nicely encompasses most sizes of objects within rooms, and even the room itself. Thus, the interactions between sound waves and the room and its contents cover the whole gamut of reflection and refraction effects, as well as absorption. It is that complexity which renders real sound fields impractical to treat analytically. Numerical methods, like Finite Element Analysis or Ray Tracing, are also limited to fairly simple approximations. In a typical room, there is usually at least the floor surface within about 2m of the source. Therefore, from a maximum of about 6 ms onwards, the sound field (even outdoors) contains components which have interacted with some surfaces or objects. After 30 ms in a small room the sound wavefront will have travelled in every direction to the boundaries of the room and will have interacted at least once with every object contained therein. ACOUSTICAL SIMPLIFICATIONS It is obviously impractical to create even a moderately accurate objective model of anything but the simplest acoustic space. It may well be unnecessary anyway. For the purposes of improving the subjective quality of the sound in a VP production, a relatively simple synthesis is sufficient to produce a convincing audio illusion. The most obvious simplification would be simply to add artificially generated reverberation to the clean sound. Done in a sensitive manner, that might provide most of the illusion required for many applications. However, it would not be an automatic process and would fail if either the source or the microphone moved close to or behind a large surface (virtual or real). A more realistic, but still much simplified model would be to synthesise the early sound from the direct and first-order boundary surface reflections and a few internal objects and to model the reverberation as a separate process, based on the global room parameters. SIMPLE VIRTUAL ROOM MODEL For the creation of a simple room model, a number of assumptions had to be made. The first one was whether the model should be two-dimensional (2D) or three-dimensional (3D). It is clear that the practical limitations of broadcasting production and home reproduction systems will, for some time, constrain the listener s experience to a 2D acoustic space. Budget constraints will mean that, for many years, the best audio reproduction technology that can be anticipated will be the existing 3/2 (5.0/5.1) multichannel system, ITU (1). Even that may take some time to become widespread. The second main issue was the likely listening environment. Though systems for reproducing virtual sound spaces on headphones (for example, using manipulation of HRTF responses) are available, it is not likely that they will form the principal means of reproduction for the majority of the audience. Such systems also have difficulties with the variability of individual responses. For this work, a more general system, based on a multiplicity of spaced loudspeakers around the listening area and conventional amplitude panning, was assumed - even though it is known that simple panning is significantly defective for image presentations to the sides and rear of the listener. This paper describes the development of a simple room acoustic synthesis system, based on a

horizontal, 2-D arrangement of loudspeakers, corresponding to the 3/2 layout of ITU Rec. 775 (1). The system was also limited to modelling a rectangular room shape, in order to ease the calculations of acoustic response. It will become clear later in this paper that a priori concerns about the need for fast recalculation of the acoustic parameters were shown to have been justified, thus making more complex models impractical at the present time anyway. The development of the model was stopped at a somewhat arbitrary stage. Most parameters were not optimised and many additional features could have been incorporated. It was necessary to stop at some point in order to carry out evaluations, which could only be done using a fully functional system, with the accompanying pictures, even if the parameters were known to be imperfect. THE ACOUSTIC ROOM MODEL The acoustic model was based on current understandings of the behaviour of the human hearing system. The three psycho-acoustic parameters described in the previous section were mapped onto three different aspects of physical room acoustics. The hearing process can be summarised as being sensitive to three distinct features - (a) the direct sound (or first arrival), (b) a number of discrete, delayed arrivals as a result of reflections from nearby surfaces (c) the overall, diffuse reverberation. Direct and early sound The geometric modelling of the direct sound was essentially trivial and self-evident, though the time delay in the direct sound path was not modelled (very rarely will it be necessary to model any overall delay in real productions). In an enclosure, the boundary surfaces (walls, floor and ceiling), together with any large objects inside the enclosure, will cause a number of discrete reflections of the sound. For the purposes of the experimental system, the space was assumed to be a right rectangular prism, which resulted in six, first-order early reflections and made the calculation of the discrete image locations trivial. No second-order early reflections were included. The 3-D model was modified to fit the 2-D reproduction paradigm by mapping the reflections from the walls, floor and ceiling onto the horizontal plane at the correct angle and distance relative to the listener. All calculations were carried out in the 3-D space and the mapping only applied to the final result. All of the seven discrete sound source images were mapped onto the reproduction loudspeaker layout by amplitude panning, using sine-function panning with constant sound power, Blumlein (2). The same panning law was applied to images to the sides and rear of the listener, where the large angular loudspeaker spacing and the poor acuity of the human hearing mechanism meant that the results were questionable in principle and worked very poorly in practice (though the listener is not actually constrained always to face forwards). Both the direct sound and the early reflections are potentially subject to several different kinds of frequency-domain filtering, for example the selective effects of air absorption over larger distances and the possibility of a non-uniform source directivity. Reverberant sound Any complete enclosure will form a reverberant space. The statistical properties of that space can easily be calculated from basic acoustic principles. The reverberation was assumed to be controlled by boundary surface absorption, uniformly distributed on all surfaces, together with air absorption. The required reverberation time was an input control parameter. The only significant, general feature of reverberation is that it is (theoretically) constant in level throughout a room. Thus, the reverberation amplitude and statistical time-domain properties need only to be calculated once for each new room. REFINEMENTS AND FILTER RESPONSES Filters The signals had to be filtered to model a number of acoustic propagation effects. In the simple model, all of the filters were implemented as a combination of a wide-band attenuation and a first-order, IIR low-pass with a single sample of delay, Fig. 1. The filter effects modelled included :- Short-term air absorption Wall reflection. Source directivity, Dunn and Farnsworth (3). Reverberation characteristics. (source spectrum and long-term air absorption).

Internal objects The basic acoustic model represented only the interior surfaces of the empty room. The representation of a multiplicity of objects within the room would have been relatively simple in principle, but the complexity of the model would very rapidly become unmanageable. The adverse effects on the calculation speed would also have been substantial. However, it was clearly necessary to model at least one internal object. Such an object could either obstruct the direct sound or add an additional reflection to the six from the boundary surfaces. The obstruction model also included an approximation to the effects of diffraction around the object. That was necessary because of the severe high-frequency attenuation which occurs in the shadow zone. When an obstruction was detected, the direct sound was switched off and replaced by two new sources in the directions of the edges of the object and at distances corresponding to the total indirect path length. Movement As originally implemented, the simple model produced results for static geometries which were, objectively, of high audio quality. However, when the geometry was changed dynamically, clicks could be heard with speech programme and more severe distortion with music programme (though music is not a very likely signal type as a source in VP). The reason for the distortion was the quantisation of the time delays in the model by the audio sampling intervals ( V 7KDW FUHDWHG ODUJH and clearly audible phase discontinuities. For example, at a modest source velocity of 1 m/s, and an system update speed of, say, 10 ms, a 4kHz source would suffer a step phase discontinuity of about 40. It was therefore necessary to quantise the distances/time delays to a much finer resolution by interpolation and to implement a crossfade mechanism to change from one discrete set of parameters to another. IMPLEMENTATION Hardware The experimental hardware consisted of a Lake DSP Huron development system. It had four, 40 MHz Motorola 56002 processors on a single card. It also had an I/O card with input ADCs and output DACs. The dsp calculations were carried out using 24-bit fixed point processing. The audio I/O was 16-bit. The system audio sample rate was 48 khz, which gave (theoretically) 416 programme steps per sample period. The host computer into which the Huron system was installed was an industry-standard IBM PCtype computer. It used a Pentium 160 as the main processor, had 16 Mbyte of RAM and the system ran under Windows 3.11. The dsp software was written in assembly language and assembled using the Motorola 56000 assembler. The control system was written in C/C++ and was compiled using Microsoft Visual C. System Overview The signal processing requirements far exceeded the capabilities of a single dsp chip. Therefore, the system had to be partitioned to divide the tasks. Fig. 2 shows the split. The three modules that resulted were the panner, the early processor and the reverberation generator. The audio input signal was applied to both early and reverberation modules. The latter also had a second input for effects or ambience. The early processor generated the direct sound, six room surface reflection signals and two internal object signals. It provided appropriate time delay and filtering for all of those outputs, but no amplitude control. The reverberation processor provided four mutually incoherent reverberation outputs, with appropriate filtering and no amplitude control. Though five reverberation signals might have been better, it was thought that four would be adequate and, in any case, the dsp processor did not have the capacity to process five outputs. It meant that one of the loudspeaker signals (the centre front one) did not have any reverberation component. The centre front loudspeaker of the 3/2 layout is also somewhat unsymmetrically placed compared with the other four, which do make a roughly symmetrical, square arrangement around the listener. In practice, the absence of reverberation in the centre loudspeaker was barely detectable, even quite close to it. The panner module summed all of these 13 inputs, in appropriate proportions, to create the five loudspeaker drive signals. In all, 96 control signals were required from the host to the dsp systems for all of the parameters.

NETWORK CONTROL The ultimate objective for the acoustic room simulator was for control by a video Virtual Production engine. There had, therefore, to be some system of remote control. A standard TCP/IP network interface was implemented. It was compatible with the sort of workstations used for the video system and was intended to give the maximum flexibility. TCP/IP messages typically take the form of relatively small blocks of information. The control message structure was therefore based on short packets, actually C/C++ structures. The control messages also included a time parameter to permit actions or changes to be scheduled by clock time rather than as they arrive over the network. WINDOWS CONTROL For the purposes of development and demonstration, a Windows control interface was developed. Though complete control flexibility would clearly not be necessary in a final product, some aspects of the system will always require manual control. Even if all of the geometrical data were to be derived from the video model, setting the reverberation time would always be a strictly acoustic input PERFORMANCE Overall quality The static noise and distortion (THDN) performance of the complete model were around - 78 db relative to zero level (22 Hz - 22 khz unweighted). Most of the noise and distortion arose from the fixed-point, 24-bit quantisation limit but was judged to be adequate (at least for an experimental system). The actual noise floor of the basic dsp system, measured using a simple mixer example supplied by the manufacturer, was not measurably different (±0.5dB). That showed that the large amount of signal processing (at 24-bit resolution) did not significantly affect the 16-bit output. Subjectively, the output sounded too reverberant. The relative output level of the reverberation was confirmed by measurement to be objectively correct, so that the effect had to be a psychoacoustic one. It is well known that the sound quality from a microphone located at a hypothetical viewer s location does not sound right - it picks up to much reverberance and is quite likely to be lacking in spatial information. There is also a significant difference between actually listening in a real space and listening to a simulation of the same thing even if that simulation is simply the sound picked up by a microphone at the same place. In the final system, the question would be reduced to the subjectively optimum ratio of clean to reverberant sound. It would be an entirely trivial matter to change the model to adjust that ratio, but it would need trial productions and subjective tests to establish the correct value. For proper subjective assessment, that must be done with matching pictures and sound together. Update speed Table 1 shows the results of timing tests (for a Pentium 166 system) :- Action Time for one cycle, ms Update frequency, Hz Crossfade 0.22 4440 Update early 1.90 526 response a Update early 2.26 442 response b Update early 2.48 403 response c Update reverb 0.86 n/a only d TABLE 1 - NOTES Results of speed tests on experimental system. a - Includes checking for potential obstruction or reflection due to the internal object. b - includes processing a reflection c - includes processing an obstruction d - Obtained by subtraction. In the present model the update of the reverberation processor was never carried out alone. This indicates what might be achieved if an improved model with multiple room capability were to be developed. The crossfade rate was especially important because it had a profound effect on the representation of moving objects. However, at 4.4 khz, the crossfade rate was achieving nearly 1/10th of the audio sample rate. At such speeds, it

becomes arguable about exactly what is being updated and when. A better method, for the future, would be to implement the interpolation directly in the dsp processors, on a sample-by-sample basis. The choice of crossfade rate depended on the conflicting requirements of movement speed, source type and desired audio quality. For a maximum source speed of 1 m/s, the distortion was completely inaudible on all speech programme and many types of music programme. For a 4 khz test tone (the most critical frequency range), the distortion was clearly audible. SUMMARY AND CONCLUSIONS This paper has attempted to summarise the implementation of a simple audio processor for creating virtual sound fields. The experimental system incorporated models of the three main aspects of the acoustic sound field the direct sound, the first-order surface reflections from the room boundary and the uniform, diffuse late sound which decays slowly to form the reverberation. The model was based on a simple, rectangular room in order to simplify the geometrical calculations and was designed for a standard, fivechannel sound reproduction system. Many acoustic refinements were included to model aspects of sound propagation such as air absorption and absorption at room boundaries. Also included were variable reverberation characteristics and a model of the directionality of the human voice. calculated and that the measured noise and distortion performances were indistinguishable from those of the basic dsp system, at least for a static geometry. The sound quality during geoemetry changes was lower, but still adequate for human speech sources and most types of music. Subjective tests, using experimental programmes, will have to be carried out to assess the operating parameters and to optimise the control functions. A full version of this paper can be found in Walker (4). REFERENCES 1. ITU-R Recommendation BS.775-1, Multichannel stereophonic sound system with and without accompanying picture. (Geneva, 1992-1994). 2. Blumlien, A.D. British Patent Specification 394,325 (Directional effects in sound systems), JAES, 6, pp 91-98, 1958 3. Dunn, H.K. and Farnsworth, D.W., Exploration of pressure field around the human head during speech. J.Acoust. Soc. Am., 10, 184-199, 1939. 4. Walker, R. A simple acoustic room model for Virtual Production. 106 th AES Convention, Munich, May 1999, Preprint #4937. ACKNOWLEGEMENTS This paper is published by permission of the British Broadcasting Corporation. Although the room model was otherwise empty, one internal object was included, if required, in order to model obstruction and reflection by large sections of the studio set. That had, perforce, to include a model of the diffraction of sound around an obstacle. In order to permit the model to be controlled by an external process, the system included a TCP/IP network interface, a simple command language and the ability to receive and interpret messages. For the purposes of experimental control and demonstration, the system also included a comprehensive manual control system, using MS Windows dialogue boxes. The system appeared to perform adequately. Objective tests showed that it was behaving as

1 sample Delay β 1-β Fig. 1. Basic filter arrangement Control inputs from controlling PC (96 in total) Main input 26 46 Early sound module Direct Image1 Image2 Image3 Image4 Image5 Image6 Object1 Object2 Panning module Loudspeaker drive outputs L C R LS RS Ambience input Reverb. module 24 Reverb1 Reverb2 Reverb3 Reverb4 Control inputs from controlling PC (96 in total) Fig. 2. Final dsp partition.