Frame Interpolation and Motion Blur for Film Production and Presentation GTC Conference, San Jose

Similar documents
White Paper : Achieving synthetic slow-motion in UHDTV. InSync Technology Ltd, UK

Understanding Compression Technologies for HD and Megapixel Surveillance

Research Topic. Error Concealment Techniques in H.264/AVC for Wireless Video Transmission in Mobile Networks

High Efficiency Video coding Master Class. Matthew Goldman Senior Vice President TV Compression Technology Ericsson

Avivo and the Video Pipeline. Delivering Video and Display Perfection

Lecture 23: Digital Video. The Digital World of Multimedia Guest lecture: Jayson Bowen

4K UHDTV: What s Real for 2014 and Where Will We Be by 2016? Matthew Goldman Senior Vice President TV Compression Technology Ericsson

Digital Media. Daniel Fuller ITEC 2110

InSync White Paper : Achieving optimal conversions in UHDTV workflows April 2015

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur

Neat Video noise reduction plug-in for Premiere (Mac)

How Does H.264 Work? SALIENT SYSTEMS WHITE PAPER. Understanding video compression with a focus on H.264

Neat Video noise reduction plug-in for Final Cut (Mac)

Torsional vibration analysis in ArtemiS SUITE 1

ATI Theater 650 Pro: Bringing TV to the PC. Perfecting Analog and Digital TV Worldwide

Chapter 10 Basic Video Compression Techniques

Contents. xv xxi xxiii xxiv. 1 Introduction 1 References 4

17 October About H.265/HEVC. Things you should know about the new encoding.

Neat Video noise reduction plug-in for Pinnacle Studio

MPEG has been established as an international standard

CM3106 Solutions. Do not turn this page over until instructed to do so by the Senior Invigilator.

Digital Representation

An Overview of Video Coding Algorithms

By David Acker, Broadcast Pix Hardware Engineering Vice President, and SMPTE Fellow Bob Lamm, Broadcast Pix Product Specialist

Neat Video noise reduction plug-in for AVX hosts (Win)

OVE EDFORS ELECTRICAL AND INFORMATION TECHNOLOGY

Information Transmission Chapter 3, image and video

Neat Video noise reduction plug-in for After Effects (Mac)

Neat Video noise reduction plug-in for Premiere (Win)

Interlace and De-interlace Application on Video

Optimization of Multi-Channel BCH Error Decoding for Common Cases. Russell Dill Master's Thesis Defense April 20, 2015

A Novel Study on Data Rate by the Video Transmission for Teleoperated Road Vehicles

Neat Video noise reduction plug-in for Vegas

Using enhancement data to deinterlace 1080i HDTV

Video Transmission. Thomas Wiegand: Digital Image Communication Video Transmission 1. Transmission of Hybrid Coded Video. Channel Encoder.

Video coding standards

Intra-frame JPEG-2000 vs. Inter-frame Compression Comparison: The benefits and trade-offs for very high quality, high resolution sequences

Film Sequence Detection and Removal in DTV Format and Standards Conversion

Impact of scan conversion methods on the performance of scalable. video coding. E. Dubois, N. Baaziz and M. Matta. INRS-Telecommunications

Fast MBAFF/PAFF Motion Estimation and Mode Decision Scheme for H.264

Case Study: Can Video Quality Testing be Scripted?

Format Conversion Design Challenges for Real-Time Software Implementations

Hardware Implementation for the HEVC Fractional Motion Estimation Targeting Real-Time and Low-Energy

Motion Video Compression

Advanced Video Processing for Future Multimedia Communication Systems

h t t p : / / w w w. v i d e o e s s e n t i a l s. c o m E - M a i l : j o e k a n a t t. n e t DVE D-Theater Q & A

Bit Rate Control for Video Transmission Over Wireless Networks

High Quality Digital Video Processing: Technology and Methods

Neat Video noise reduction plug-in for OpenFX-compatible hosts (Mac)

Understanding IP Video for

Composite Video vs. Component Video

MULTI-STATE VIDEO CODING WITH SIDE INFORMATION. Sila Ekmekci Flierl, Thomas Sikora

Audio and Video II. Video signal +Color systems Motion estimation Video compression standards +H.261 +MPEG-1, MPEG-2, MPEG-4, MPEG- 7, and MPEG-21

Bring out the Best in Pixels Video Pipe in Intel Processor Graphics

Understanding IP Video for

Introduction to image compression

ELEC 691X/498X Broadcast Signal Transmission Fall 2015

4K & DVB-S2X HOW OPERATORS CAN BE COST-EFFECTIVE. Market Trend. Introduction. 4K & DVB-S2X. How Operators Can Be Cost-effective

Pre-processing of revolution speed data in ArtemiS SUITE 1

Real-time SHVC Software Decoding with Multi-threaded Parallel Processing

Constant Bit Rate for Video Streaming Over Packet Switching Networks

A Study of Encoding and Decoding Techniques for Syndrome-Based Video Coding

Neat Video noise reduction plug-in for OpenFX-compatible hosts (Linux)

Color Image Compression Using Colorization Based On Coding Technique

HEVC: Future Video Encoding Landscape

Video Over Mobile Networks

Monitor and Display Adapters UNIT 4

RedEye Analog ConvNet Image Sensor Architecture for Continuous Mobile Vision

Nattress Standards Conversion V2.5 Instructions

Chapter 3 Fundamental Concepts in Video. 3.1 Types of Video Signals 3.2 Analog Video 3.3 Digital Video

AUDIOVISUAL COMMUNICATION

New-Generation Scalable Motion Processing from Mobile to 4K and Beyond

The H.263+ Video Coding Standard: Complexity and Performance

GTC 2013 Presentation Amit Gulati Gopi Lakshminarayanan Dolby Laboratories, Inc.

ECE3296 Digital Image and Video Processing Lab experiment 2 Digital Video Processing using MATLAB

Efficient Implementation of Neural Network Deinterlacing

Content storage architectures

The Lecture Contains: Frequency Response of the Human Visual System: Temporal Vision: Consequences of persistence of vision: Objectives_template

Professor Laurence S. Dooley. School of Computing and Communications Milton Keynes, UK

Video Processing Applications Image and Video Processing Dr. Anil Kokaram

1. INTRODUCTION. Index Terms Video Transcoding, Video Streaming, Frame skipping, Interpolation frame, Decoder, Encoder.

International Journal for Research in Applied Science & Engineering Technology (IJRASET) Motion Compensation Techniques Adopted In HEVC

Multimedia Systems Video I (Basics of Analog and Digital Video) Mahdi Amiri April 2011 Sharif University of Technology

MPEG + Compression of Moving Pictures for Digital Cinema Using the MPEG-2 Toolkit. A Digital Cinema Accelerator

Video 1 Video October 16, 2001

Alcatel-Lucent 5910 Video Services Appliance. Assured and Optimized IPTV Delivery

Keep your broadcast clear.

Compressed-Sensing-Enabled Video Streaming for Wireless Multimedia Sensor Networks Abstract:

Lossless Compression Algorithms for Direct- Write Lithography Systems

3.0 Next Generation Digital Terrestrial Broadcasting

FRAME RATE CONVERSION OF INTERLACED VIDEO

Midterm Review. Yao Wang Polytechnic University, Brooklyn, NY11201

Advanced Computer Networks

Audio Watermarking (NexTracker )

Yong Cao, Debprakash Patnaik, Sean Ponce, Jeremy Archuleta, Patrick Butler, Wu-chun Feng, and Naren Ramakrishnan

SWITCHED INFINITY: SUPPORTING AN INFINITE HD LINEUP WITH SDV

Implementation of an MPEG Codec on the Tilera TM 64 Processor

A NEW METHOD FOR RECALCULATING THE PROGRAM CLOCK REFERENCE IN A PACKET-BASED TRANSMISSION NETWORK

Chapter 2. Advanced Telecommunications and Signal Processing Program. E. Galarza, Raynard O. Hinds, Eric C. Reed, Lon E. Sun-

Getting Images of the World

Transcription:

Frame Interpolation and Motion Blur for Film Production and Presentation 2013 GTC Conference, San Jose Keith Slavin, isovideo LLC (slides 20 to 22 by Chad Fogg) 1

What we have today 24 frames/sec is too low to avoid judder on fast moving camera pans with detail To avoid judder (a perception of uneven motion), low frame rates require some combination of: low light levels, low camera motion, out-of-focus backgrounds, or motion blur (larger shutter angles). Existing deblur algorithms assume global camera motion and are heavily iterative (slow) No algorithms known today will undo local motion blur nor (with noise and Cramer-Rao bounds analysis) should we assume they will ever exist! Films made today don't look good on the best display systems of today definitely not near limits of human perception, and not future proof Europe is also subjected to a 4% speedup from 24 -> 25, requiring damaging audio pitch conversion for some viewers isovideo GPU Technology Conference, 2013 2

What we have available Today Motion-Compensated Frame Rate Conversion Motion vector quality from 24 to higher frame rates is strongly scene dependent Obtaining high quality motion vectors from occlusion and revelation is an open problem Motion aliasing is common in repetitive man-made objects and wagon-wheels Frame-to-frame motion where an object does not overlap itself is very problematic Up-converted 24 looks smooth-and-blurry, with effective shutter angles >> 360 And the Consequences Are... Most people end up viewing movies with duplicated frames, and many with additional 3:2 judder on fixed 60HZ refresh displays. This talk is about judder, frame rates, blur, motion estimation, interlacing & deinterlacing, and also their effects on compression. isovideo GPU Technology Conference, 2013 3

Judder Perception For a display refresh frequency and refresh period F R,=1/T R presented an average of n times, given by: and a video frame-rate F v, we expect each video frame to be F R F v = T v T R =n=a p+b q Where p,q represent the two nearest distinct integer repeat rates, given by p= n,and q= p+1 and a,b are the relative weightings, such that This gives: a+b=1 a=1+ p n and b=n p The repeat values have associated presentation times T p = p T R and T q = q T R, where T p < T q. We associate an overall weighted judder score for these time periods: Judder=J Tp a+ J Tq b Where J T is a judder measure as a function of presentation time. We expect that: J Tp <J Tq isovideo GPU Technology Conference, 2013 4

Experiment involved: 1) using our Legato MCFRC to convert to many frame-rates, no added blur 2) using refresh rates that are an exact multiple of frame rates. 120Hz: 0.0083, 0.0166, 0.0250, 0.0333, 0.0417, 0.0500, 0.0583, 0.0666 The perceptual judder function was found empirically to be a sigmoid function of the frame presentation time J T : J T = 1 q+1, where q=e(t center T )/T gain where for crowd-run clip at viewing distance 1.5 x picture height: T center 42 ms (about 24Hz) T gain 6.25 ms (about 160Hz) and presentation time T is integer multiple of monitor refresh T refresh = 1/F refresh. T center is proportional to a global measure of object velocity for a scene. isovideo GPU Technology Conference, 2013 5

Sigmoid Judder Perception Function Crowd-run, 1.5 x picture height T center = 0.042 seconds, T gain =0.00625 J T T presentation = m T refresh isovideo GPU Technology Conference, 2013 6

Example 1: if F R =60, F v =25, then n=2.4, p=2, q=3, a=0.6, b=0.4. We obtain two presentation times: T p =2/60=0.033, T q =3/60=0.050 seconds. From the graph, J 0.033 = 0.19, J 0.050 = 0.80, so: Judder=0.6 J 0.033 +0.4 J 0.050 =0.6 0.19+0.4 0.80=0.434 Example 2: if F R =60, F v =24, then n=2.5, p=2, q=3, a=0.5, b=0.5. The same presentation times apply, so: Judder=0.5 J 0.033 +0.5 J 0.050 =0.5 0.19+0.5 0.80=0.495 The above shows simple linear interpolation from one point on the sigmoid to another. The set of points is determined by the monitor refresh rate. Example 3: if F R =60, F v =50, then n=1.2, p=1, q=2, a=0.8, b=0.2. The presentation times are: 1/60=0.0167, and 2/60=0.033 seconds. Judder=0.8 J 0.0167 +0.2 J 0.033 =0.8 0.015+0.2 0.19=0.05 isovideo GPU Technology Conference, 2013 7

Judder Perception & Blur Judder perception is discordance between natural (smooth) and perceived motion in our vision system Shutter angle 120º may see judder 360º shutter - no judder-but may look blurry at low frame rates Upsampled - Shutter angle 720º No judder but blurry isovideo GPU Technology Conference, 2013 8

Judder Perception a Summary: Judder perception is discordance between natural (smooth) and perceived motion in our vision system Perception varies between people Perception is proportional to speed of object motion Perception decreases with motion blur Perception increases with amount of detail/texture/edges in an object Perception increases with light levels and contrast Perception increases with solid angle of moving object to eye Perception increases with longer presentation times of individual images Perception increases for any larger integer multiples of refresh time in a rendering cadence Frequencies are not an issue this is all about presentation time isovideo GPU Technology Conference, 2013 9

Where Movie Production needs to go Large subtended angle, bright, high refresh rate displays are becoming common Sufficient movie information needs to be captured to meet human vision limitations Improved cameras and increased compute power allow new solutions Movies need to be near human perception limits Solution: high frame rates min. blur from cameras, none from graphics/effects Motion-compensated frame rate conversion then works well Simulated motion blur can be added for those who like 24 frames/sec isovideo GPU Technology Conference, 2013 10

Reference: High Frame Rates Solve all Conversion Problems HFR reduces MCFRC problems associated with occlusions. HFR reduces the likelihood of aliasing in the reference. HFR uses faster shutter speeds, with much less motion blur. Motion deblurring is never required. Reduced motion blur improves edge detail - allows MCFRC to work better. Motion blur is small enough that it doesn't adversely MCFRC algorithms. In down-conversion, simulated motion blur can be added. New, fast technologies allows high quality derivatives to be created at will. For post-production, the product is the reference work. Derivatives can be automatically generated - some might prefer to adjust blur scene-by-scene. Experiments are on-going to add blur automatically based on scene analysis. isovideo GPU Technology Conference, 2013 11

GPU acceleration of Motion-Compensated Frame-Rate Conversion Legato-cinema is our CUDA-based MCFRC product, with simulated motion blur. Without blur: 90 frames/sec output rates for 1080p50 to 60 conversions. Motion blur is implemented by upsampling to a higher frame rate and averaging groups of frames. Blur typically slows output to around 20 frames/sec. Simulated shutter angles are used to control motion blur familiar paradigm for the movie industry. Estimate of input shutter angle can be used to control oversampling. Smaller input angles (higher oversampling) is visually safer, but mostly just slows conversions. The output angle controls the added output motion blur as expected. Motion deblur will probably never be supported! isovideo GPU Technology Conference, 2013 12

System Issues 16-bit CUDA processing improved SNR and simpler workflow Dynamic GPU resource allocation: multi-gpu systems avoid bottlenecks. Frame-grained parallelism achieves efficient conversions in multi-user systems and conversion pipelines. Our lossless 2:1 super-fast compression tool can be used to help preserve quality over many operations, while doubling storage bandwidth and halving file sizes. Relatively low CPU usage allows CPU intensive tools like x264 encoding to be in a processing pipeline. On our 3.8GHz over-clocked Intel 3930K reference machine with Samsung SSD 830, and VDPAU, we have been able to smoothly display 3840x2160 clips at 50 frames/sec. isovideo GPU Technology Conference, 2013 13

Deinterlacing Same raw video bandwidth, each frame two fields, double temporal sample rate Doubling temporal sampling can reduce the perception of judder, but... Sampling is spatially damaged by discarding alternate odd/even lines Tearing occurs from motion, so at some stage, deinterlacing for progressive displays is required Computation grows exponentially for an asymptotic improvement as more input samples become involved in reconstruction Excellent deinterlacing quality is computationally very expensive OK results are possible for 1080i if the output is spatially low-pass filtered but why not use 720p? Human vision limitations don't sit too close Spatial damage means modern compression algorithms (H.264 and HEVC) can do better with the same frame rate vs field rate at the same SNR isovideo GPU Technology Conference, 2013 14

A compressed interlaced transmission system can be replaced with: deinterlace compress decompress reinterlace lossy compression decompression + Noise Interlaced (blue) samples on left noisy channel to right Noise from (a) influence of interpolated (red) samples on compression, and (b) lossy compression Deinterlacing provides progressive video at the end of transmission for future-proof system integration Deinterlacing provides progressive video for archives where the original interlaced can be extracted with an improved SNR vs compressing raw interlaced directly. isovideo GPU Technology Conference, 2013 15

TV Distribution & Broadcast piecemeal replacement of interlaced capture/production/distribution systems is possible. May take a long time. No technical barriers remain. Motives for migration include: Lower bit rates lower costs Progressive systems or better deinterlacers improved distributed image quality Better control of final quality (no deinterlacers in the wild ) Lower transmission bandwidth/channel in future ATSC (>=2.0) broadcast Better integration with Internet and computer-based display systems. Better access to portable devices (which can't/don't deinterlace) Simpler production and editing Simpler conversion between formats (scaling, frame-rate-conversion, etc) isovideo GPU Technology Conference, 2013 16

Demeler Deinterlacer CUDA-based motion estimation is particularly effective for most of picture area CUDA: diagonal interpolation - improves results in some situations CUDA: any failures in motion estimation and diagonal interpolation (resulting in combing) are detected and patched Faster than real-time performance is possible with two GTX 580s or GTX 690s. Demeler has low flicker and no output filtering Low Flicker Field Pass-Through (LFFPT) LFFPT lossless compressed deinterlaced archives can recover the original interlaced video. LFFPT an average 15% bandwidth reduction when used before H.264 or HEVC compression, instead of compressing interlaced directly. isovideo GPU Technology Conference, 2013 17

Moving an Interlaced File from Interlaced -> Deinterlaced Archive START: Interlaced TFF-1 from archive Deinterlaced-2 (assume TFF) Interlaced TFF-2 B A A=B? No Stop! Yes Verify Deinterlaced-2 TFF assumption & then archive Deinterlaced-2 file (reversible lossless, or lossy?) isovideo GPU Technology Conference, 2013 18

Updating a Deinterlaced Archive (new Deinterlacer) Interlaced TFF-1 Deinterlaced-2 (TFF implied) Interlaced TFF-2 B A A=B? No Stop! START with Deintrlaced-1 From archive Replace Deinterlaced-1 with Deinterlaced-2 in archive Yes isovideo GPU Technology Conference, 2013 19

0 Interlaced vs. deinterlaced (HEVC) Measure PSNR Interlace source video 1920x540x60(i) HEVC Encoder Field sequence 1920x540x60(i) (Level 4.1) Decoder Interlaced output Pre-deinterlace HEVC Encoder Decoder 1920x1080x60p Frame sequence 1920x1080x60p (Level 4.2) Progressive output Measure PSNR fine detail is preserved, and input fields passed through unchanged.

1 Results so far.. On the diverse but challenging test sequence set chosen.. deinterlaced HEVC coded frame sequences average -15% (lower bit-) rates than HEVC coded field sequences (fixed QP=22,27,32,37, HM 8.0). Range is -39% to +32% Pre-encoding deinterlaced AVC coded sequences average lower rate than AVC MBAFF coded frame sequences (-18%). Range [-40%,+22%] Bdrate() suggests deinterlacing prior to encoding is better than deinterlacing after decoding.

Upgrading to HEVC LTO Hard drive SSD Satellite IRD or file server feed File storage Decoder (e.g. 40 Mbps) Video Processing (color; LPF;..) 1080i (SDI) Logo insertion & mixing deint Encoder 1080p@60frames/sec 5 mbps mux modulator To RF Only necessary change from AVC & MPEG-2 to HEVC: drop-in encoder replacement with deinterlacer Audio; other programs in same multiplex isovideo GPU Technology Conference, 2013 22

Viarte Professional Quality Standards-Conversion/Transcoding Server Simple deployment Viarte is file-based and mountable as a shared drive, Scalable to multiple servers, Configurable drag-and-drop triggers one or more conversions, Faster-than-realtime full-hd throughput via i) load-balanced multi-gpu acceleration and ii) an intelligent optimization (that speeds up throughput by up to 250%). Bit-rate reduction achieved by customizing frame rates and images sizes for distribution to mobile networks, while maintaining or improving picture quality. isovideo GPU Technology Conference, 2013 23