Lip-Sync Investigation and Management

Similar documents
Audio Watermarking (NexTracker )

Synchronization Issues During Encoder / Decoder Tests

METADATA CHALLENGES FOR TODAY'S TV BROADCAST SYSTEMS

COPYRIGHT 2011 AXON DIGITAL DESIGN B.V. ALL RIGHTS RESERVED

TIME-COMPENSATED REMOTE PRODUCTION OVER IP

Multi-language audio in Dolby E. A description of how to encode multiple COPYRIGHT 2011 AXON DIGITAL DESIGN B.V. ALL RIGHTS RESERVED

THE MPEG-H TV AUDIO SYSTEM

Research & Development. White Paper WHP 318. Live subtitles re-timing. proof of concept BRITISH BROADCASTING CORPORATION.

quantumdata 980 Series Test Systems Overview of Applications

TECHNICAL MEDIA SPECIFICATION ON THE FILE BASED SUBMISSION OF MATERIALS TO BE AIRED

Standard Definition. Commercial File Delivery. Technical Specifications

White Paper. Video-over-IP: Network Performance Analysis

Case Study: Can Video Quality Testing be Scripted?

EBU Digital AV Sync and Operational Test Pattern

METHODS TO ELIMINATE THE BASS CANCELLATION BETWEEN LFE AND MAIN CHANNELS

FREE TV AUSTRALIA OPERATIONAL PRACTICE OP- 59 Measurement and Management of Loudness in Soundtracks for Television Broadcasting

Date <> Time-of-day <> Frequency <> Phase

ENGINEERING COMMITTEE Digital Video Subcommittee SCTE STANDARD SCTE

SPG8000A Master Sync / Clock Reference Generator Release Notes

Cablecast SX. Setup Guide. c Tightrope Media Systems For Cablecast version Build 206

TECHNICAL SPECIFICATIONS FOR THE DELIVERY OF CLOSE TO TRANSMISSION TELEVISION PROGRAMMES TO THE

Contents. xv xxi xxiii xxiv. 1 Introduction 1 References 4

A NEW METHOD FOR RECALCULATING THE PROGRAM CLOCK REFERENCE IN A PACKET-BASED TRANSMISSION NETWORK

Introduction How to operate Cable Test Limit mode Log & EDID Troubleshooting. HDMI Diagnostics and Troubleshooting

ATSC vs NTSC Spectrum. ATSC 8VSB Data Framing

A/V Cinema Scaler Pro II

AMERICAN NATIONAL STANDARD

Audio Watermarking (SyncNow ) Audio watermarking for Second Screen SyncNow with COPYRIGHT 2011 AXON DIGITAL DESIGN B.V. ALL RIGHTS RESERVED

SG4424 HDTV Slave Sync Generator User Guide

Verifying Digitized Files

High Definition Television. Commercial File Delivery. Technical Specifications

CONSOLIDATED VERSION IEC Digital audio interface Part 3: Consumer applications. colour inside. Edition

QIP7232 P2. Hybrid QAM/IP High-definition Set-top. Quick Start Guide

A new technique to maintain sound and picture synchronization

TelePresence Cisco TelePresence Synch with Edge95MXP - Troubleshooting

OPERATION MANUAL. USF-1013DEMUX Digital Audio Demultiplexer. 2 nd Edition. Software Version Higher

Neuron. Network attached processing for broadcast infrastructures

TECHNICAL SUPPLEMENT FOR THE DELIVERY OF PROGRAMMES WITH HIGH DYNAMIC RANGE

One of the challenges in a broadcast facility is

FS3. Quick Start Guide. Overview. FS3 Control

IQDEC01. Composite Decoder, Synchronizer, Audio Embedder with Noise Reduction - 12 bit. Does this module suit your application?

SOC Single Channel H264 + Audio Encoder module

BTV Tuesday 21 November 2006

UltraPioneerAVR3 HSPI User s Guide A HomeSeer HS3 plug-in to enable control of Pioneer A/V Receivers that support the IP & RS232 control system.

USING LIVE PRODUCTION SERVERS TO ENHANCE TV ENTERTAINMENT

The use of Time Code within a Broadcast Facility

Enabling and Enriching Broadcast Services by Combining IP and Broadcast Delivery. Mike Armstrong, James Barrett & Michael Evans

FLEXIBLE SWITCHING AND EDITING OF MPEG-2 VIDEO BITSTREAMS

AT780PCI. Digital Video Interfacing Products. Multi-standard DVB-T2/T/C Receiver & Recorder & TS Player DVB-ASI & DVB-SPI outputs

Making Digital AV Signals Work Roy A. Hermanson Jr., CTS-I/D Regional Applications Specialist

Delivery of Spots to

BER MEASUREMENT IN THE NOISY CHANNEL

IPTV delivery of media over networks managed end-to-end, usually with quality of service comparable to Broadcast TV

V pro8 QUICK START GUIDE

decodes it along with the normal intensity signal, to determine how to modulate the three colour beams.

Dolby MS11 Compliance Testing with APx500 Series Audio Analyzers

SoundField UPM-1 Stereo to 5.1 Converter

3G/HD/SD dual channel audio embedder/de-embedder

MULTIMEDIA TECHNOLOGIES

Understanding and Managing Conversion Delays in Live Production Systems

Ultra 4K Tool Box. Version Release Note

Ultra TQ V3.4 Update 4KTQ Ultra TQ Update 1

TV Synchronism Generation with PIC Microcontroller

Case Study Monitoring for Reliability

Operating Instructions

Media Delivery Technical Specifications for VMN US Network Operations

HEDD Bridge B1 - Dante

PixelNet. Jupiter. The Distributed Display Wall System. by InFocus. infocus.com

Eventide Inc. One Alsan Way Little Ferry, NJ

h t t p : / / w w w. v i d e o e s s e n t i a l s. c o m E - M a i l : j o e k a n a t t. n e t DVE D-Theater Q & A

QRF5000 MDU ENCODER. Data Sheet

Mobile DTV Viewer. User Manual. Mobile DTV ATSC-M/H DVB-H 1Seg. Digital TV ATSC DVB-T, DVB-T2 ISDB-T V 4. decontis GmbH Sachsenstr.

Cablecast Server. Setup Guide. c Tightrope Media Systems For Cablecast version Build 74

BM-A1-E16SHD V2.2. Manual BM-A1-E16SHD. 16 Channel Digital Audio Monitor. User s Guide. Page 1

PRO-HDMI2HD. HDMI to SDI/3G-HD-SD Converter. User Manual. Made in Taiwan

New Technologies for Premium Events Contribution over High-capacity IP Networks. By Gunnar Nessa, Appear TV December 13, 2017

TECHNICAL STANDARDS FOR DELIVERY OF TELEVISION PROGRAMMES TO

Thor Broadcast SDI-DVBT-IP & SDI-DVBT-IPLL Product Lines

SingMai Electronics SM06. Advanced Composite Video Interface: HD-SDI to acvi converter module. User Manual. Revision 0.

IP LIVE PRODUCTION UNIT NXL-IP55

BM- AV1- E16SHD Manual BM-AV1-E16SHD. 16 Channel Digital Audio Monitor. User s Guide. Version /01/2013. Version 2.

4K UHDTV: What s Real for 2014 and Where Will We Be by 2016? Matthew Goldman Senior Vice President TV Compression Technology Ericsson

Forward TS Product Line

3Gb/s, HD, SD embedded domain watermarking encoder based on Kantar technology A Synapse product COPYRIGHT 2016 AXON DIGITAL DESIGN BV

ISO Digital Forensics- Video Analysis

BASE-LINE WANDER & LINE CODING

Tutorial on the Grand Alliance HDTV System

BrightEye NXT Frame Synchronizers

SPG700 Multiformat Reference Sync Generator Release Notes

Pablo Rio, Pablo PA. V2.0 rev 13 New Feature List. If you have any questions please contact Damon Hawkins

Scalable Media Systems using SMPTE John Mailhot November 28, 2018 GV-EXPO

Professional Media. over IP Networks. An Introduction. Peter Wharton Happy Robotz. Introduction to Video over IP

1 HDMI YPbPr HD Digital TV Compact Modulator QAM ATSC DVB-T 1080p/60

By David Acker, Broadcast Pix Hardware Engineering Vice President, and SMPTE Fellow Bob Lamm, Broadcast Pix Product Specialist

3Gb/s, HD, SD embedded domain Dolby E/D/D+ decoder and to Dolby E encoder with audio shuffler and optional audio description processor

Kramer Electronics, Ltd.

BLOCK CODING & DECODING

PWS-100TD1 Tape Digitizing Station Simple-to-use tape digitizing server for efficient migration of legacy videotape archives

MPEG Solutions. Transition to H.264 Video. Equipment Under Test. Test Domain. Multiplexer. TX/RTX or TS Player TSCA

Technical requirements for the reception of TV programs, with the exception of news and public affairs programs Effective as of 1 st January, 2018

Transcription:

Lip-Sync Investigation and Management Operational Guide Document created by: Carys Hughes Design Engineer Date created: May 2018

The content of this document is for general information purposes only. Whilst Sky UK Limited make reasonable efforts to ensure the information is correct, we make no representation, warranty or guarantee, whether express or implied, that the content of this document is accurate, complete or up to date. Any reliance you place on the information contained herein is therefore strictly at your own risk. In no event will we be liable for any loss or damage arising from your use of the information contained within this document. With contribution from: Mark Ross-Smith Vladislav Korotkov Adrian Hurt Kevin McCue Matthew Brading Sam Sale Dolby Laboratories Inc Copyright Sky UK Limited 2018 1

Table of Contents 1 Introduction and main principles 3 2 Frequently encountered lip-sync issues 3 3 A/V latency diagrams 4 4 Tools and troubleshooting 6 4.1 Baseband (SDI) 6 4.1.1 Lip-sync measurement tools within these areas 6 4.2 Analysing file based workflows 7 4.3 Transport Stream Analysis 8 5 QC and Monitoring Environments 9 5.1 Considerations for monitoring chains 9 5.2 Assessment within consumer style environments 9 6 Managing lip-sync within an UHD context 10 6.1 Progressive frames and Dolby E 10 6.2 Where could Dolby E header alignment affect lip-sync? 12 6.3 Dolby Atmos - Dolby Digital Plus Encode 12 6.4 Historical difficulties when measuring UHD lip-sync 12 7 Test Pattern Analysis Methods 13 7.1 Hitomi Matchbox 13 7.2 Sync-One2 13 7.3 Transport stream analysis method using the Hitomi Matchbox pattern 14 7.4 NLE, file based analysis using Sync-One2 15 8 Comments and Errors 16 Copyright Sky UK Limited 2018 2

1 Introduction and main principles The purpose of this document is to draw attention to areas within our chains and workflows where we may be vulnerable to introducing lip-sync errors and to set a precedent for the appropriate level of investigation that should be established for rectifying these errors. Support teams must be mindful of components within a chain/workflow that are prone to introducing lip-sync offsets and should be able to identify: The size and direction of the offset and which component is affected Consistency in the size and occurrence of any offset What causes and/or exacerbates the offset When considering an A/V chain/workflow, any individual component s processing delay should be accounted for within the unit itself, or compensated for before the next monitoring point, to ensure that all monitoring is reliable and that the station output has known, correct A/V sync. A/V offsets should be tracked within a standard diagram format. This is outlined later on in this document and enables thorough tracking through a chain. These standardised diagrams can then be used for future reference and for updating documentation where appropriate. An agreed and proven piece of synchronous test content is essential for the reliable testing of all chains and workflows. 2 Frequently encountered lip-sync issues Examples of the primary issues encountered when managing lip-sync include but are not limited to: Out of sync audio and video components within live feeds - whether being played out internally, or externally sourced. Errors introduced during extra signal processing tasks, such as audio de-embedding/reembedding, video up/down conversion. Errors introduced during encoding and decoding processes, whether live or file-based. Misjudged subjective lip-sync assessment, due to monitoring chains that contain their own unaccounted-for A/V processing delays. Some feeds or file-based content also contain Dolby E encoded audio. This audio is frame based and misalignment of the audio components can cause additional problems. This will be addressed later on in the document. Copyright Sky UK Limited 2018 3

3 A/V latency diagrams The diagram format outlined in this section should always be used when tracking lip-sync. This diagram format should also be present in any High-Level Diagrams that introduce A/V equipment/processing into TX chains/vod workflows. Swim lanes are used to separate each process in the chain/workflow A Key should be used to: Describe components present within the diagram (If relevant) indicate the layout of the components being transported Describe processing blocks Describe how the latency was observed Indicate where measurements have been taken for a particular process/component Indicate where delays have been inferred by doing measurements up/downstream of the process/component Indicate where delay has been configured Clearly show where A/V operate in different domains (embedded/de-embedded) The figures in each block are used to show the offset added by that hardware to each component - even if 0. Copyright Sky UK Limited 2018 4

Tables similar to the examples included below should be used to track each audio component against video: GDR26 U4T100 Buffer U4T100 emb/delay ISM Result Quads 1,2,3,4 Video 0 20 400 0 420 PCM Stereo Audio 0 20 400 0 420 Table 1 - PCM delay against video. All figures in milliseconds. GDR26 U4T100 Buffer DDP94 DE>DD U4T100 ISM Result Quads 1,2,3,4 Video 0 20 - emb/delay 400 0 420 Dolby Digital BAU Audio 0 20 370 31 0 420 Table 2 Dolby Digital delay against video. All figures in milliseconds. GDR26 DP591 U4T100 Buffer U4T100 emb/delay ISM Result Quads 1,2,3,4 Video 0-20 400 0 420 Dolby Digital Plus BAU Audio 0 245 0 174 0 419 Table 3 Dolby Digital Plus delay against video. All figures in milliseconds. Results shown in red indicate when an error may be observed in the resulting calculation. It is worth noting here that each measurement tool will have a limited measurement resolution (i.e a VALID Reader might have a result granularity of 1 ms and so it can t be any more accurate than +/- 1ms). This means that we do need to allow for a degree of measurement rounding error and the size of that allowance will depend on the resolution of the measurements being given. That being said, accuracy in measurement mustn t be ignored - tolerance is allowed for the resolution of measurement equipment, but be aware that accumulated, tolerated offsets throughout a chain have the potential to combine to become unacceptably large. For that reason, unexpected offsets should be scrutinized at each point in the chain. Copyright Sky UK Limited 2018 5

4 Tools and troubleshooting 4.1 Baseband (SDI) Monitoring points within MCR and TX environments should be able to accurately represent the lip-sync of incoming OB feeds and transmission playout. Management of lip-sync within monitoring chains will be covered later on in this document. If incoming feeds contain lip-sync errors, then it would be necessary for correction and troubleshooting to take place within the OB. When this is not possible, or no further realignment can be performed at the OB, it may be necessary for delay to be applied within MCR in order to pass synchronous audio and video through to the TX chains. Within our linear TX chains; encoding, decoding, and transcoding times should be accounted for correctly for main, reserve and forced (Dolby Digital Plus for UHD with Dolby Atmos, Dolby Digital and PCM) 2 channel audio paths. 4.1.1 Lip-sync measurement tools within these areas Test signal generators/reader pairs are used to measure the lip-sync of incoming feeds from OB sources. For HD sources, a VALID generator is used, while for UHD we would use an Hitomi Matchbox system. The Matchbox can decode Dolby E, Dolby Digital Plus and Dolby Digital. A test pattern is output from a Matchbox or VALID Generator at the OB and measured for lip-sync errors within MCR using the receiver unit in our facility. OB Sky VALID/ Matchbox Generator VALID/ Matchbox Reader Copyright Sky UK Limited 2018 6

4.2 Analysing file based workflows Any process where essence or its container is altered has the potential to introduce lip-sync errors. At Sky UK There are lots of content preparation processes that have the potential to affect the sync of audio and video. These include but are not limited to: Ingest, TX Packaging (Conform and Transcode), Audio normalisation in the MAM, OTT and VOD Transcodes, Vantage transcodes for EVS and NLE Editing. A known good test file is used to test our workflows this enables us to track the introduction of lip-sync errors reliably through our workflows. The file contains a Sync-One2 test pattern, with sync pulse and flash frames, which can then be reliably used for lip-sync measurement. Depending on the output format, the sync of audio and video can be verified using one of a number of methods. Most reliable would be an analysis tool, but the MAM support and QC teams also use NLEs (Non Linear Editors such as Adobe Premiere Pro) to view waveforms and verify whether sync points match up with the correct video frames in our test pattern. For further instructions for NLE based assessment, see Test Pattern Analysis Methods. For Dolby E analysis, an additional audio decode is required to do this type of comparison and it is essential to know whether the decode tool adds any additional offset, so that this can be accounted for in calculations. E.g. 40ms audio delay is introduced by SoundCode, as it is designed to mimic the Dolby 572 when decoding Dolby E, with the introduction of a frame of audio delay. When assessing a workflow, the general warning sign would be that the output sync doesn't match the input. We must follow the workflow through and check the output of each process in order to find a cause. Example As much as 90ms of audio delay was introduced into TX packaged assets at 2 points in the chain. a) Audio and Video was found to be out of sync within the original content, held in the MAM - Caused when performing audio normalisation. In this case an updated version of code for the encoders has solved the problem. This could affect either stereo or Dolby audio, depending on which had required normalization. b) Additional audio delay was introduced into the Dolby E audio in TX Packaged assets by the transcoder used to prepare playout files this was acknowledged by the vendor but with an anomaly in their stated figure - still under investigation. It is also important to identify where more complex processes may introduce further error - e.g multipart transcodes where the transcoder must seek through multiple assets parts. Copyright Sky UK Limited 2018 7

4.3 Transport Stream Analysis Reliable analysis tools, with the ability to decode Dolby audio can be used to measure the offset of each audio element within a Transport Stream, with respect to video as long as you can work out where the audio should be positioned. The only way to accurately do this is to use a test asset, containing an audio sync pulse, with a trackable video element, such as a flash frame. In our case, we have used either the Sync-One2 or Hitomi test patterns. When these patterns are encoded, passed through a workflow and wrapped within a transport stream, whether for VOD or DTH the TS can be opened and measured within an analysis tool such as Harmonic s A/VCodec Analyser. Once the file is open in the analyser, the first sync point should be located (in the case of Hitomi, this is when the white line moves around to the 12 o clock position). We know that that point of the pattern should occur at the same time as an audio sync pulse and a mute in the L channel audio. The position of the presentation timestamps given for the video frames can be located in the audio timeline within the analyser, and compared to the position of the expected audio pulse/mute at that point in the pattern. For further instructions regarding the use of the Hitomi pattern for transport stream analysis, see Test Pattern Analysis Methods. Copyright Sky UK Limited 2018 8

5 QC and Monitoring Environments 5.1 Considerations for monitoring chains As previously covered, components within any A/V chain (be it Delivery, Monitoring or TX) may have native audio and video processing delays that contribute to the overall observed lip-sync at an end device, when the overall audio and video latencies do not match. Native audio and video processing delays should already be compensated for within each component. When the output of a device is found to still contain lip-sync errors this should be addressed, in order to allow us to judge accurately and eliminate end-to-end lip-sync sync error within the guidelines of the EBU (European Broadcasting Union) 1 : End to end (ie. OB to Consumer) Sound before picture 40 ms Sound after picture 60 ms Per stage in chain Sound before picture 5 ms Sound after picture 15 ms When measuring lip-sync errors within a monitoring setup, we must account for any delay that may be introduced by components in our chain, so that they do not affect our judgement of end-to-end lip-sync. Examples could include: Cards/Devices used to decode video or audio Cards/interfaces used to de-embed audio Display Latency This equipment may have a known delay (for example a 40ms Dolby E decode delay) or this may need to be measured. It is always to best to confirm the figure through measurement, so that any 1-2ms second discrepancy can be accounted for. It will also help confirm whether equipment is behaving as expected. 5.2 Assessment within consumer style environments Some of our assessment environments contain viewing and listening set ups that more closely resemble a consumer s own viewing environment. An example of this would be our VOD QC suites, where we generally use consumer displays, playback devices and AVRs. The consumer environment may contain other elements that could contribute to observed lip-sync. Consumer panels offer a range of processing options that are designed to optimise content presentation. When performing assessment of lip-sync, it is essential that any additional processing that may incur further A/V delay is disabled. These could include but are not limited to: Motion Smoothing/TruMotion Super Resolution Edge Enhance MPEG Noise Reduction Real Cinema If audio is decoded by an AVR or a playback device, then the decode time of that unit will also need to considered, as it may introduce a significant error and impair the judgement of the QC operator. 1 EBU Recommendation R37-2007 https://tech.ebu.ch/docs/r/r037.pdf Copyright Sky UK Limited 2018 9

6 Managing lip-sync within an UHD context 6.1 Progressive frames and Dolby E Unlike SD and HD, UHD content is currently broadcast in 50 progressively scanned frames per second. In the baseband domain, that is currently transported over four 3G-SDI links (Quad 3G SDI). Stereo audio sources are provided as PCM, while Dolby E is provided as an audio source when surround audio formats are required. Within this baseband domain, we must be aware that: Our station reference, whether Black and Burst, or Tri Level Sync, is still 25Hz Any Dolby E audio sources are transported as 25 audio frames per second A 40ms Dolby E frame bridges two 50P video frames (20ms in length), or two interlaced fields. The A/V requirements in the context of baseband and Dolby E can be roughly broken down as follows: Video should be locked to station reference Audio and Video should be in sync Dolby E headers should occur with a predictable Frame/Field. In our case, with a 25Hz reference used with both interlaced and progressive feeds, it makes sense for us to align Dolby E headers to start in phase with the station reference, and so with the first frame/field. If the Dolby E header was to be aligned to the wrong progressive video frame/interlaced field (mid sync pulse) and line, this could lead to switching and cutting mid audio frame, causing unwanted audio artefacts, such as mutes or splats. For this reason, Dolby E sources should have their headers phase aligned with station reference. In order to align audio and video frames appropriately, we may need to push/pull audio frames into alignment and in the case of multiple Dolby E streams being present in the source, may have to prioritise the lip-sync of a specific audio pair in order to achieve full alignment of all headers. Tektronix WFM can now be used to perform Dolby E alignment measurement with respect to our 25Hz reference. We can see where the header is with respect to a 25Hz system reference and note the field/frame and line. For further information regarding line positions please defer to the relevant Dolby documentation. 2 2 Preferred Alignment for Dolby E in HD Systems https://www.dolby.com/us/en/technologies/dolby-e-preferred-alignment.html Copyright Sky UK Limited 2018 10

Once the Tektronix unit is set to time Dolby sync to external reference, we can use the below view to see the positioning of the Dolby E header: Screenshot taken from Tektronix UI In this example, we can see that the Dolby E Frame is located on Line 32 of Field 1 within our 25Hz reference: Screenshot taken from Tektronix UI The Axon U4T100 (UHD Toolbox Card) can be used for frame alignment and delay insertion etc. It has a 20ms input alignment buffer and is used in MCR to retime the audio against the video. It is also used to combine the various audio formats and embed with UHD video, in sync, within SDI. An example of use: The TX UHD chain has 400ms (20 frames) of video delay and as the video and audio is combined, additional audio delays are applied with respect to each of the existing audio encoding/transcoding times (Dolby E-> Dolby Digital, PCM down-mix generation, external DP591 Dolby Atmos encode time). The resulting output is all then in sync with the video (while also considering Dolby E header alignment). Copyright Sky UK Limited 2018 11

6.2 Where could Dolby E header alignment affect lip-sync? Incoming UHD OB feeds MCR must ensure that the Dolby E pairs on an incoming feed are aligned to the correct field/frame once decoded and handled with respect to system reference. The U4T100 gives the option to align to a Dolby E pair of choice - useful if the OB are unable to apply further correction. We usually default to Dolby E on pair 5&6. Dolby E and ED2 pairs should all be in alignment with each other. The act of realigning audio so that the frame headers occur together may well have an impact on lip-sync. 6.3 Dolby Atmos - Dolby Digital Plus Encode Dolby Digital Plus is currently used to deliver Dolby Atmos as part of our UHD service in the UK. Dolby DP591 The A/V sync of this unit drifts when incoming content is switched badly e.g router cuts through a Dolby E frame, or outside of the Vertical switching standards defined in RP168. The delay is incremented with each bad switch from our SAM router. This is currently fixed by disconnecting and reconnecting the input. The problem can be mitigated using frame synchronisers in line with the router output. The UHD 2 Staging System doesn't have post router synchronisers, so it happens there more frequently. Downstream, we must account for a Dolby Atmos, Dolby Digital Plus Encode time of 245ms. 6.4 Historical difficulties when measuring UHD lip-sync Previously we have needed to down-convert progressive UHD quadrants in order to measure using VALID, which results in the audio being early by an additional 6-7ms within the measurement chain. In order to measure Dolby Digital and Dolby Digital Plus, the audio first had to be decoded. This was done using the Tektronix WFM (contains a CAT552 card) the decoding time of the WFM was measured to be 34ms. This is no longer a problem as the Hitomi Matchbox can handle a 1080P quadrant, so as long as a full representation of the test pattern is contained within a single quadrant an additional decode will no longer be a consideration when measuring Dolby E sources. Copyright Sky UK Limited 2018 12

7 Test Pattern Analysis Methods At present, there are two primary test patterns used within Sky for lip-sync assessment: 7.1 Hitomi Matchbox This pattern is generally used to assess the lip-sync of incoming OB feeds, but once encoded, can also be used for transport stream analysis. Our file based Matchbox pattern contains a period of channel idents and announcements where normal tone pattern is not running. The result is that sometimes the video frame containing the sync point at the 12 o clock position is not accompanied by audio. Analysis cannot be performed until the next sync point to occur with audio. The visual sync point should occur at the same time as a mute in the L channel audio. When using a progressive Matchbox pattern for UHD, the 1080P pattern is duplicated in each of the UHD quadrants. 7.2 Sync-One2 The Sync-One2 pattern is useful as it has a series of easily predictable sync points. A 10 second countdown is followed by a series of beeps and flash frames, repeated at 1 second intervals. This can be used along with the actual Sync-One2 handheld measurement tool, to give a reading for lipsync error in a viewing environment. The pattern is also suitable for file based assessments. Copyright Sky UK Limited 2018 13

7.3 Transport stream analysis method using the Hitomi Matchbox pattern Both the Hitomi Matchbox and Sync-One2 patterns can be used for transport stream analysis of either linear stream captures or streams generated in file based transcodes. When analyzing Dolby audio formats, there may be some variations in the format naming conventions: Dolby AC3 An alternative name used to refer to Dolby Digital Dolby E-AC3/Enhanced AC3 Also used to refer to Dolby Digital Plus When analyzing this pattern in a transport stream using A/V Codec Analyser, the easiest method for use is as follows: Identify the frame containing the sync point (as shown below) and note the presentation timestamp (the bottom number shown below the frame thumbnail). Screenshot taken from Harmonic A/V Codec Analyser, containing a frame of Sky UK s implementation of the Hitomi Matchbox test pattern Click on the audio element requiring analysis in our case we have selected Dolby Enhanced AC3. Once in the audio analysis tab, select the Audio/Video synchronization option, so that the desired frame is located within the audio timeline. Also ensure that the Video timestamps are selected for display. Locate the required timestamp number and measure from the red dotted line at that point, to the beginning of the left channel audio mute. This can be done by clicking on the red dotted line and dragging the cursor to the beginning of audio mute. In the below example, we have measured a 0.008s offset, with the mute occurring before the sync point (audio early). Copyright Sky UK Limited 2018 14

Screenshot taken from Harmonic A/V Codec Analyser A/V Codec Analyser can also be used to analyse MXF, MOV and MP4 contained media. PCM, MPEG II, AAC, Dolby Digital and Dolby Digital Plus audio are all supported, but this tool is not able to decode Dolby E. 7.4 NLE, file based analysis using Sync-One2 When analyzing master files, it is also useful to be able to assess the alignment of audio and video using a non-linear edit timeline. This method can also be used to assess the alignment of Dolby E against video but this process also requires a preliminary decode of the Dolby E audio, so that it can then be inserted into the NLE timeline as decoded PCM. Once the video and audio are present on the timeline, the granularity of the timeline should be altered until the audio waveform can be clearly seen, with audio time units should visible so that lip-sync can be measured with appropriate granularity (in this case in audio samples). In the example below, the sync frame has been located and the audio beep can be seen to begin at exactly the same position on the timeline. Screenshot taken from Adobe Premiere Pro CC, showing a frame of the Sync-One2 test pattern Copyright Sky UK Limited 2018 15

8 Comments and Errors If you spot any errors or have any comments or suggestions for future revision of this document please contact Carys Hughes or Vlad Korotkov at one of the following addresses: carys.hughes@sky.uk vladislav.korotkov@sky.uk Copyright Sky UK Limited 2018 16

Material contained within this document is protected by copyright. It may not be copied, reproduced, republished, posted, broadcast or transmitted in any way except for your own internal use. Dolby and the double-d symbol are trademarks of Dolby Laboratories. TEKTRONIX and TEK are registered trademarks of Tektronix, Inc. Adobe, the Adobe logo, and Adobe Premiere are either registered trademarks or trademarks of Adobe Systems Incorporated in the United States and/or other countries. Harmonic [all other Harmonic products mentioned] are trademarks, registered trademarks or service marks of Harmonic Inc. in the United States and other countries. Other company and product names mentioned within this document may be trademarks or service marks of their respective owners. Copyright Sky UK Limited 2018 17