longterm archiving experience

Similar documents
NTSC/PAL. Network Interface Board for MPEG IMX TM. VTRs BKMW-E2000 TM

The experience of RAI DigiMaster Project, in progress..

Videotape to digital files solutions

Digital Terrestrial HDTV Broadcasting in Europe

R&S VENICE The ingest and playout platform.

Signal Ingest in Uncompromising Linear Video Archiving: Pitfalls, Loopholes and Solutions.

FREE TV AUSTRALIA OPERATIONAL PRACTICE OP- 59 Measurement and Management of Loudness in Soundtracks for Television Broadcasting

Technical requirements for the reception of TV programs, with the exception of news and public affairs programs Effective as of 1 st January, 2018

Delivery of Spots to

Archiving: Experiences with telecine transfer of film to digital formats

TECHNICAL MEDIA SPECIFICATION ON THE FILE BASED SUBMISSION OF MATERIALS TO BE AIRED

Supervision of Analogue Signal Paths in Legacy Media Migration Processes using Digital Signal Processing

PWS-100TD1 Tape Digitizing Station Simple-to-use tape digitizing server for efficient migration of legacy videotape archives

Verifying Digitized Files

AN MPEG-4 BASED HIGH DEFINITION VTR

IASA TC 03 and TC 04: Standards Related to the Long Term Preservation and Digitisation of Sound Recordings

Dietrich Schüller. Safeguarding audiovisual information for future generations. Inforum 2016 Prague May 2016

Quality Control Experiences from a Large-Scale Film Digitisation Project

Ad Benchmark s Insertion Workflow. Increase revenues and flexibility with advanced regionalization for broadcast ads and programs

Standard Definition. Commercial File Delivery. Technical Specifications

P1: OTA/XYZ P2: ABC c01 JWBK457-Richardson March 22, :45 Printer Name: Yet to Come

WHITE PAPER THE FUTURE OF SPORTS BROADCASTING. Corporate. North & Latin America. Asia & Pacific. Other regional offices.

Metadata for Enhanced Electronic Program Guides

TEN.02_TECHNICAL DELIVERY - INTERNATIONAL

TECHNICAL REQUIREMENTS Commercial Spots

ATSC Standard: A/342 Part 1, Audio Common Elements

Preservation-Worthy Digital Video, or How to Drive Your Library into Chapter 11. Jerome P. McDonough Elmer Bobst Library, New York University

USING LIVE PRODUCTION SERVERS TO ENHANCE TV ENTERTAINMENT

MDPI Film Processing Harder, Better, Faster, Stronger. Brian Wheeler, Library Technologies Digital Library Brown Bag Series #dlbb April 18, 2018

IASA-TC 06 Video Preservation Guidelines

Digital disaster recovery for audiovisual collections: testing the theory

TECHNICAL SUPPLEMENT FOR THE DELIVERY OF PROGRAMMES WITH HIGH DYNAMIC RANGE

Implementation of MPEG-2 Trick Modes

Understanding PQR, DMOS, and PSNR Measurements

(12) Patent Application Publication (10) Pub. No.: US 2007/ A1

Media Delivery Technical Specifications for VMN US Network Operations

One of the challenges in a broadcast facility is

AN EXPLORATION OF THE BENEFITS OF MIGRATION TO DIGITAL BROADCASTING

IBM Linear Tape File System Applications and considerations for broadcast adopters

Composite Video vs. Component Video

SONY SHOWCASES COMPREHENSIVE LINE-UP OF DIGITAL ACQUISITION AND PRODUCTION EQUIPMENT AT NAB

SMPTE 334M DATA DOES HAVE A LIFE BEFORE EMISSION. JIM CARRUTHERS PhD, PEng CEO NORPAK CORPORATION. norpak corporation 19 Oct 00

Case study TeleZüri goes tapeless with

Automated QC for File-based Media. November 2016

Digital Cinema Specifications 71 Locarno Festival

PROGRAM INFORMATION. If your show does not meet these lengths, are you willing to edit? Yes No. English French Aboriginal Language: Other:

TECHNICAL STANDARDS FOR DELIVERY OF TELEVISION PROGRAMMES TO

Case Study: Can Video Quality Testing be Scripted?

Research & Development. White Paper WHP 318. Live subtitles re-timing. proof of concept BRITISH BROADCASTING CORPORATION.

Vision Standards Bring Sharper View to Medical Imaging

COPYRIGHT 2011 AXON DIGITAL DESIGN B.V. ALL RIGHTS RESERVED

BBC PSB UHD HDR WCG HLG DVB - OMG!

NEW APPROACHES IN TRAFFIC SURVEILLANCE USING VIDEO DETECTION

Videotape Transfer. Why Transfer?

ITU-T Y.4552/Y.2078 (02/2016) Application support models of the Internet of things

Telecommunication Development Sector

MPEG-2 MPEG-2 4:2:2 Profile its use for contribution/collection and primary distribution A. Caruso L. Cheveau B. Flowers

Why We Measure Loudness

A COMPARATIVE ANALYSIS OF TAPE TECHNOLOGIES FOR MID-RANGE SYSTEMS AND SERVER APPLICATIONS

Advanced Coding and Modulation Schemes for Broadband Satellite Services. Commercial Requirements

SERIES H: AUDIOVISUAL AND MULTIMEDIA SYSTEMS Infrastructure of audiovisual services Coding of moving video

DELIVERY SPECIFICATIONS. TAPE and FILE DELIVERY

The Norwegian Digital Radio Archive - 8 years later, what happened? Svein Arne Brygfjeld, National Library of Norway

TR 038 SUBJECTIVE EVALUATION OF HYBRID LOG GAMMA (HLG) FOR HDR AND SDR DISTRIBUTION

Alpha channel A channel in an image or movie clip that controls the opacity regions of the image.

EBU R The use of DV compression with a sampling raster of 4:2:0 for professional acquisition. Status: Technical Recommendation

AT780PCI. Digital Video Interfacing Products. Multi-standard DVB-T2/T/C Receiver & Recorder & TS Player DVB-ASI & DVB-SPI outputs

Next Generation Software Solution for Sound Engineering

DIGITAL BROADCAST TEST AND MONITORING SOLUTIONS

White Paper : Achieving synthetic slow-motion in UHDTV. InSync Technology Ltd, UK

Alchemist XF Understanding Cadence

New Technologies for Premium Events Contribution over High-capacity IP Networks. By Gunnar Nessa, Appear TV December 13, 2017

CONSOLIDATED VERSION IEC Digital audio interface Part 3: Consumer applications. colour inside. Edition

ADS Basic Automation solutions for the lighting industry

Implementation of an MPEG Codec on the Tilera TM 64 Processor

FLEXIBLE SWITCHING AND EDITING OF MPEG-2 VIDEO BITSTREAMS

Understanding Compression Technologies for HD and Megapixel Surveillance

Cost models for digitisation and storage of audiovisual archives

DVB-T2 Transmission System in the GE-06 Plan

THINKING ABOUT IP MIGRATION?

NAS vs. SAN: Storage Considerations for Broadcast and Post- Production Applications

Signum BlackHive. Generation II. Broadcast Production System and video server. The new system generation signum.blackhive

MPEG Solutions. Transition to H.264 Video. Equipment Under Test. Test Domain. Multiplexer. TX/RTX or TS Player TSCA

Will Widescreen (16:9) Work Over Cable? Ralph W. Brown

The Lincoln TX-2 Input-Output System*

Altera's 28-nm FPGAs Optimized for Broadcast Video Applications

METADATA CHALLENGES FOR TODAY'S TV BROADCAST SYSTEMS

Frame Compatible Formats for 3D Video Distribution

EndNote: Keeping Track of References

ITU-T Y Functional framework and capabilities of the Internet of things

PulseCounter Neutron & Gamma Spectrometry Software Manual

ST2110 Why Is It So Important?

High Definition Television. Commercial File Delivery. Technical Specifications

TECHNICAL SPECIFICATIONS. Television commercials

ENGINEERING COMMITTEE Energy Management Subcommittee SCTE STANDARD SCTE

THE MPEG-H TV AUDIO SYSTEM

MULTI-CHANNEL CALL RECORDING AND MONITORING SYSTEM

TR 040 EVALUATION OF IMF FOR BROADCASTERS

Subtitle Safe Crop Area SCA

Scan. This is a sample of the first 15 pages of the Scan chapter.

Transcription:

QC and longterm archiving experience Roman Meszmer ORF During the digitization of legacy archive material at ORF, the Austrian public-service broadcaster, hundreds of thousands of video cassettes will need to be converted to digital file formats over the next decade. To guarantee the sustainable usability of this file-based material, thoroughly-planned guidelines for Quality Control (QC) have to be worked out. Although some QC tools already exist, their detailed parameter settings as well as the exact requirements to test against, have not yet been determined and will be the main focus of the EBU QC group. Besides giving some details of a migration project at ORF, this article discusses different proposals concerning the understanding of QC from an archival viewpoint, along with file format collateral effects and, finally, some data-tape library issues. Long-term digital archiving means ingesting material and, in special cases, also the migration of analogue and digital tape content to files as well as maintaining the availability and manageability of the generated files over decades and even centuries! From that perspective, quality control seems to have a different relevance compared to the everyday use of programme material. What seems sufficiently high quality for edit and playout at present could lead to major problems in terms of inacceptable errors during material handling in the future if standard compliance and audio or video consistency, as well as metadata correctness, are not ensured during the archiving process. However, one of the most unpleasant jobs of archivists is weighing up the storage and handling costs against possible future value. Thus, special care must be taken to optimise every task of the archiving and de-archiving process, especially the time-consuming ones. One essential part to achieving a perfectly working archive which is usable for a long time to come is the planning and realization of a well-conceived QC workflow which is the main focus of the EBU QC group. A short history of media formats Fig. 1 outlines the history of the video carriers mostly used in ORF. Starting with black-and-white silent movie material from the 1890s (which is still usable if prepared specially) right through to highquality HDTV file formats, there has been a long history of change. In the 1960s, the first magnetic two-inch video tapes were introduced. They brought a new quality in terms of material editing and processing... but also a drawback in preservability because of the deterioration of the magnetization over time. The development goal of media carriers was to decrease their size, which goes hand in hand with reducing their life span and thus shortening the recycling interval. Around 1975, the oneinch tape was introduced, followed in 1979 by the UMATIC cassette format, which represented the all-time bottom of the quality level. EBU TECHNICAL REVIEW 2011 Q3 1 / 8

After the introduction of analogue BETACAM, a big step-up in quality was achieved by switching to the digital formats BETA DIGI and BETA IMX. The latter format is now used intensively within the recycling processes of SD material. Because the IMX carrier uses the same coding as an existing file format, no generation loss will occur when finally migrating the tapes to files. For the first time, these carriers have been used with both the 4:3 and 16:9 video formats. Figure 1 Changing media types over time The launch of HDTV required new cassette types. Several television companies finally decided to switch to XDCAM-HD using the MXF wrapper [1]. From the archivist s viewpoint, this format has the advantage of the smallest HD bandwidth and therefore the lowest storage demand, but it also has the disadvantage of carrying relatively low video quality (LGOP, 8bit only ) and more difficult handling than I-frame-only formats (in editing for example). As there is a demand for different video resolutions from higher ones such as 4k material, to very low ones such as current internet-quality material, and also new ones such as 3DTV or interactive production formats a lot of new codec versions (and wrappers) can be expected over the next decades. IMX migration During the last decade, lots of analogue magnetic video media such as MII have been migrated to the IMX (cassette) format during an attempt to recycle the media. The IMX tape format allows the recording of video and audio signals on behalf of a codec which can now be used as file essence without any loss of quality. At ORF, around 360,000 hours of IMX along with DIGI BETA material (see Fig. 2) has to be transferred (migrated) into IMX files. Many of the collection tapes have several productions stored on one tape, which gives a total of more than a million files. A short calculation shows the amount of data involved. IMX D10 as well as XDCAM-HD 4:2:2, using a 50 Mbit/s video codec (with audio and data overhead concerned) Figure 2 Part of ORF s Digi-Beta and IMX archive practically results in about 60/8 = 7.5 MByte/s = 2.7x10 10 Byte/h for each tape. For the complete archive of 360,000 hours, the storage capacity comes to a total of 9.7x10 15 Bytes/h which rounds up to 10 PB (Petabytes). As the focus of the group lies on quality control, material handling is not provided in full detail. Only the QC-relevant workflow items are mentioned here. The migration process has been planned to start with selecting those tapes which are of most interest to the current programming schedules, as some content appears to be more relevant than others. During this selection process, a daily production list with cassette (barcode) numbers and their related technical metadata will be prepared. The final QC reports have to be checked against this EBU TECHNICAL REVIEW 2011 Q3 2 / 8

list to be sure that the migration process was correct. Errors such as wrong aspect ratio or loss of audio channels could be detected by this procedure. The next step concerns the cleaning of the tape. The cleaning machines report information about the quality of the tape, which could be valuable when attempting to migrate very old cassettes. Whether this type of information should be collected and handled by QC or not, is still a matter of debate. This kind of information, as well as the following reports, will have to be presented in a common form such as XML. The main ingest process will be realized by several Sony E-VTR devices. Cassettes are ingested by robotics or also by hand. Implementations show that VTR robots have only a minor cost advantage over manual ingestion, so their implementation for the ORF archive process has not yet been decided. During this process step, some tape-related metadata information has to be gathered. Because of recording in more than real time, a monitoring of the video signal is not possible. The E- VTR-generated files contained in a (problematic) MXF header are stored in cache storage together with already-collected QC reports on the tape, in XML form. The E-VTR-generated MXF container does not accord with the MXF standard and therefore has to be corrected using one or more already-developed solutions for that purpose. A wrapper check is applied to guarantee the correct syntax of the MXF container. As there is no possibility of live video monitoring when ingesting faster than real time over E-VTR, a tailored software has to be applied which can analyse the generated video file for characteristic audio and video errors. The Joanneum Institute in Graz, Austria, for example, offers a monitoring system which, up to a certain point, is able to find most of these errors: the associated software delivers easy-to-configure XML reports as well as a human-readable display of the reported file segments with audio and video representation. Together with the now standardized MXF file, the XML report is sent to archive personnel to be sample-checked with the help of the generated QC file and compared against the daily production list. A detailed refusal procedure for defective tapes or otherwise-failed ingests has yet to be determined. The cassette has to be ingested a second time, the material has to be restored by editing or the cassette has to be declared non-restorable and as a consequence has to be disposed of. QC placing QC is meant to report non-compliances of files to a given set of certain parameters. Common tools provide very detailed reports. Their interpretation is not easy to handle, because nearly every media coder has its own specialities (errors, problems and non-compliances against the common standard). Although there are error summaries and first attempts to abstract from the high number of error issues such as red and green lamps (see Fig. 3) signalling a kind of go/no-go state they very often show a red signal if the analyser has not been configured correctly: a general quality statement about the usability of a given file for archiving is not possible. Figure 3 Error summary Also, there arises the question about the actual location for QC. While this question is yet to be worked out, there are two basic approaches: The first proposal for the archivist, the more pleasant one considers the QC on the production side. All QC is done after the ingest process, there are no more faulty files in the workflow, and no problems arise when archiving or de-archiving files. However, this proposal demands a consequent and deep QC at every ingest process, which could lead to non-acceptable delays. Also for the production side, the QC is of minor importance. If files can be edited in edit suites and played out without errors over the (current) playout stations, there is no interest in an additional workflow item which only decreases working speeds. EBU TECHNICAL REVIEW 2011 Q3 3 / 8

This leads to the second proposal, where deep QC rests mainly on the archive side. Again the delays play a role but they do not hinder the production process directly. On the other hand, QC would be performed unnecessarily because all of the files would be checked automatically, even though they would already have passed some lighter QC during production and ingest. Considering the huge amount of files which have to be archived, there are also costs in terms of the QC hardware and software involved. Both proposals are not optimal per se. To find a feasible solution, a combination of both versions has to be developed. Besides, some technical issues such as watch folder technology or a decision for event-triggered QC have to be discussed. QC variants and depths Different places in the workflow will require different QC approaches. IMX migration touches several very different task areas for QC. Some important examples should be mentioned here: tape quality check; ingest process; check on analogue errors; loudness [4]; metadata; wrapper (syntax) check; audio channels (Dolby-E, ). All items should be checked against some not-yet-defined common standard (also called house standard ). Additionally there are several file formats involved. Although the current ORF production format is XDCAM-HD 4:2:2 at 50 Mbit/s in the form of RDD9 (in two flavours), legacy formats also still play an important role (with declining importance over time). The lion s share of archive files is represented by D10/IMX files which, since one year, are generated out of analogue material and put into a temporary storage. After solving some problems by updating older versions of our ingest system, our QC tools (and interpreting specialists) rated the generated file quality as acceptable for archiving (though still not perfect). Other file formats delivered by older edit systems are less comfortable to deal with. Some of them even generate MXF files but they do not stick to the standard. To get these files into a correct form, severe converting effort has to be accomplished. Another issue concerns high-quality codecs such as proprietary Apple ProRes or Avid DNxHD. As raw material they are archived in their original form, but for use in the common production environment they have to be converted into XDCAM-HD. The QC workflow concerning those codecs has not yet been established; the current emphasis lies on the standard production workflow. Up to now, barely one of the available ingest systems generates a truly open-standard compliant file. No wonder that troubles are daily fare in a heterogeneous ingest / playout system. Hopefully the Abbreviations AAF Advanced Authoring Format BWF (EBU) Broadcast Wave Format HDTV, HD High-Definition Television IETF Internet Engineering Task Force http://www.ietf.org/ IMX (Sony) Interoperable Material exchange MXF QC SDTV, SD VTR XML Material exchange Format Quality Control Standard-Definition Television Video Tape Recorder extensible Markup Language, IETF RFC 2119 EBU TECHNICAL REVIEW 2011 Q3 4 / 8

efforts within the EBU group to find a common reliable and automated QC can bring more transparency into these problem areas and finally lead to error-free playout, which also works reliably in the years to come. Why stick with the D10/IMX format? Sticking to a single standard codec and wrapper as a production format can only last a limited period of time. Thus, archiving material in its original codec (e.g. the D10/IMX format for recent SD material) is the only way to conserve its full quality and avoid generation losses due to repeated codec conversions. Although it seems inviting to convert the whole archive into the prevailing working codec (e.g. XDCAM-HD), a lossy re-coding procedure (D10 to XDCAM) as well as a reduction in video resolution must be taken into account. Converting from 4:3 to 16:9 for general use would mean converting into pillar box format. If resolution loss cannot be accepted for a special production, recourse to SD material and choosing the segment of interest at full resolution would be advisable. Unwrapping Another proposal for holding media files in the archive, without the hassle of retaining non-audio/ video information, could be to extract the video and audio streams (MPEG and BWF [3] respectively) from the MXF wrapper using currently-available tools such as MXF splitter and storing the essences separately. Incompatibilities or even errors in the wrapper information could be avoided at the cost of a more complex handling of files, because of the need to recombine several audio channels with the video signal synchronously and to hold several files as one item in the library. This method could be very efficient if a change of wrappers were to be decided in the future and all essence data could be wrapped into this successor wrapper. The AAF-derived [1] MXF-wrapper still seems to cope with current and foreseeable-future tasks but special requirements such as 3D or internet formats could arise which possibly could not be supported by MXF. Checksum issues The purpose of registering checksums for each file was discussed thoroughly. Certainly, there are some costs involved in the calculation time. As has been shown, the processing time for the two main candidates MD5 and SHA-1 lies well under real time and can thus be ignored. At least three arguments for checksum generation can be proposed: to guarantee the originality of the material; to check that the cycling process did not insert new errors; to verify that no further long-lasting deep QC is necessary, once QC has been done and the checksum has been registered. In several cases the originality of the material can be of importance. Modifications to the material after the archiving process can be detected easily and reliably. No audio comment can be modified or deleted, nor can parts of the video be removed or even new material inserted. During the cycling process, checksums are a very useful tool to guarantee that the file which has to be transferred is equivalent to the original. On detection of a mismatch between the registered and the currently-calculated checksum, another source of the material has to be chosen which hopefully exists in the form of a backup copy. Finally, calculation effort can be avoided if an already-executed deep QC has been done and registered into the archive database. Depending on the implementation of the general QC concept for EBU TECHNICAL REVIEW 2011 Q3 5 / 8

gathering a report or determining the applicability for a certain purpose, only a checksum has to be calculated and compared against the database entry. If the values match, the old QC report can be used. Tape library concerns Two main concerns of QC have to be standard compliance and correct metadata within the archived files because especially when using tape libraries there exists no easy way to correct the whole archive in background processes! The reason for it lies in the mechanical and logical limitation of tape robotics. They are designed for a standard archiving process; several files are written into the library concurrently, according to the dimensioning of the robots, and can be retrieved on demand providing only a limited number of files per hour. There are two problem groups involved with modifying the whole archive. The first refers to the mechanical layout of the robot system (arms, bearing ), which usually is not designed for such an intensive de-archiving process and the very limited number of tape read/write cycles. The second group refers to the fragmentation which occurs on the library tapes after deleting. The intended tape library is unable to delete files to gain space effectively, such as on disks; files can only be marked as non-existent. New space on the tape or even a new tape will be chosen to write the corrected file back. Thus, a relatively high percentage of the files would have to be copied onto new tapes in a defragmentation process, which is expensive and additionally wears out the robotic system unnecessarily. The next efficient chance, without an additional write attempt for a file repair, is opened during the cycling process. This infinite background process, with incidentally triggered copy activity, would lead to an unwanted mixture of corrected and non-corrected files over a period of decades. The alternative restoration during the de-archiving process would delay the delivery process substantially. Following the repair process, an additional QC over the whole essence would have to be carried out. The repair process and the QC together lead to a delay of more than real time depending on the type of error (the file has to be read once, the corrections have to be made, the file has to be stored, read from the QC process again, tested and proofed). Negative example Fig. 4 represents a regularly-occurring frame of a video in a clip which was reported as being perfect in terms of video codec. Some combinations of small MXF errors lead to totally disrupted video frames when played out by a commonly-used video server. Detecting such errors occasionally after ingesting lots of material into your archive database leaves you in a disastrous situation. Only an appropriate QC combined with a repair or denial strategy can help to avoid this fate. Figure 4 Problems during playout Conclusions QC in long-term archiving is very different from QC for everyday production: the archive can contain material from both known and unknown sources, with indeterminate quality. Even an upstream CMS cannot guarantee that the specified standard production format is in place. Production systems may very well leave ready-prepared files unprocessed and forward them to the archive system. EBU TECHNICAL REVIEW 2011 Q3 6 / 8

Furthermore, in addition to the transfer of correspondents material via the internet or user-generated content the opportunity to receive viewers input in various formats there is also the socalled archive exchange. In this way, material is received in various wrappers and codecs and often has to be tested and converted. One is well advised to set clearcut interface definitions between archive systems, the outside world and the production area, DIGITAL ARCHIVES Roman Meszmer is a senior design engineer and has worked in the Broadcast Production Systems department of ORF for 26 years. In addition to offering technical support to ORF correspondents worldwide, and planning the TV studios mostly in the ORF Centre, his main focus currently lies on filebased issues encountered while setting up the file-based ORF video archive. Mr Meszmer received his Master s Degree in computer science (informatics diploma) in 2009 from FernUniversitaet in Hagen, Germany, with emphasis on network theory. In 2005 he became active within the EBU Multichannel Audio Workshop and is currently involved with the EBU QC group. and to see to it that these are also adhered to. One of the key cost-generating factors in the future will be the handling of files despite the clear setting of a company standard codec. It can be reported from experience that, in addition to the integration of high and low bandwidth codecs into the production system, already diverse dialects of the actual determined codec / wrapper combination cause predictable but also non-deterministic failures. In order to maintain essences in the best quality possible over the long term, the content must be held in the original codec. Alternatively, it may be appropriate to archive uncompressed essences something which, for financial reasons, can only be practical for a relatively small amount of highquality material. The decision to exclusively provide the archive in the respective production format is short-sighted and leads to the essence quality continuously declining over time, due to small incompatibilities in the changeover from these old codecs to the current codec. Fortunately, in the scope of the QC Group, the EBU is focussing on analysing the available devices and software for a suitable Quality Control and offers a platform for the comprehensive elaboration of necessary key parameters involving their use. On this basis, a framework is created which will guarantee that the original quality of the archive material is maintained in the long term, both in the case of forward and backward archiving. References [1] B. Gilmer: File Interchange Handbook Elsevier Inc., 2004, pp 1-30. [2] B. Devlin and J. Wilkinson: The MXF book Elsevier Inc., 2006. [3] EBU doc Tech 3285: Specification of the broadcast wave format EBU, Geneva, 2011. [4] EBU doc Tech 3343: Practical guidelines for EBU R128 (Loudness) EBU, Geneva, 2011. This version: 22 September 2011 EBU TECHNICAL REVIEW 2011 Q3 7 / 8

Published by the European Broadcasting Union, Geneva, Switzerland ISSN: 1609-1469 Editeur Responsable: Lieven Vermaele Editor: E-mail: Mike Meyer tech@ebu.ch The responsibility for views expressed in this article rests solely with the author EBU TECHNICAL REVIEW 2011 Q3 8 / 8