Motion Blur Reduction for High Frame Rate LCD-TVs

Similar documents
LCD Motion Blur Reduced Using Subgradient Projection Algorithm

AN OVERVIEW OF FLAWS IN EMERGING TELEVISION DISPLAYS AND REMEDIAL VIDEO PROCESSING

High Quality Digital Video Processing: Technology and Methods

UNIVERSAL SPATIAL UP-SCALER WITH NONLINEAR EDGE ENHANCEMENT

OPTIMAL TELEVISION SCANNING FORMAT FOR CRT-DISPLAYS

An Overview of Video Coding Algorithms

Keep your broadcast clear.

Module 3: Video Sampling Lecture 17: Sampling of raster scan pattern: BT.601 format, Color video signal sampling formats

FRAME RATE CONVERSION OF INTERLACED VIDEO

The Lecture Contains: Frequency Response of the Human Visual System: Temporal Vision: Consequences of persistence of vision: Objectives_template

Chapter 10 Basic Video Compression Techniques

CM3106 Solutions. Do not turn this page over until instructed to do so by the Senior Invigilator.

PREDICTION OF PERCEIVED QUALITY DIFFERENCES BETWEEN CRT AND LCD DISPLAYS BASED ON MOTION BLUR

New-Generation Scalable Motion Processing from Mobile to 4K and Beyond

TOWARDS VIDEO QUALITY METRICS FOR HDTV. Stéphane Péchard, Sylvain Tourancheau, Patrick Le Callet, Mathieu Carnec, Dominique Barba

DISPLAY AWARENESS IN SUBJECTIVE AND OBJECTIVE VIDEO QUALITY EVALUATION

Module 3: Video Sampling Lecture 16: Sampling of video in two dimensions: Progressive vs Interlaced scans. The Lecture Contains:

AUDIOVISUAL COMMUNICATION

Research and Development Report

A New Standardized Method for Objectively Measuring Video Quality

Interlace and De-interlace Application on Video

TERRESTRIAL broadcasting of digital television (DTV)

Dynamic Scan Clock Control in BIST Circuits

RECOMMENDATION ITU-R BT.1201 * Extremely high resolution imagery

Objective video quality measurement techniques for broadcasting applications using HDTV in the presence of a reduced reference signal

Research Topic. Error Concealment Techniques in H.264/AVC for Wireless Video Transmission in Mobile Networks

Lecture 2 Video Formation and Representation

Understanding PQR, DMOS, and PSNR Measurements

AN IMPROVED ERROR CONCEALMENT STRATEGY DRIVEN BY SCENE MOTION PROPERTIES FOR H.264/AVC DECODERS

ZONE PLATE SIGNALS 525 Lines Standard M/NTSC

ALIQUID CRYSTAL display (LCD) has been gradually

Research and Development Report

Principles of Video Compression

IC FOR MOTION-COMPENSATED DE-INTERLACING, NOISE REDUCTION, AND PICTURE-RATE CONVERSION

RECOMMENDATION ITU-R BT (Questions ITU-R 25/11, ITU-R 60/11 and ITU-R 61/11)

Rec. ITU-R BT RECOMMENDATION ITU-R BT PARAMETER VALUES FOR THE HDTV STANDARDS FOR PRODUCTION AND INTERNATIONAL PROGRAMME EXCHANGE

DELTA MODULATION AND DPCM CODING OF COLOR SIGNALS

White Paper : Achieving synthetic slow-motion in UHDTV. InSync Technology Ltd, UK

Reduced complexity MPEG2 video post-processing for HD display

A Unified Approach to Restoration, Deinterlacing and Resolution Enhancement in Decoding MPEG-2 Video

Chapter 3 Fundamental Concepts in Video. 3.1 Types of Video Signals 3.2 Analog Video 3.3 Digital Video

White Paper. Uniform Luminance Technology. What s inside? What is non-uniformity and noise in LCDs? Why is it a problem? How is it solved?

NON-UNIFORM KERNEL SAMPLING IN AUDIO SIGNAL RESAMPLER

Development of Simple-Matrix LCD Module for Motion Picture

Deinterlacing An Overview

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video

Audio and Video II. Video signal +Color systems Motion estimation Video compression standards +H.261 +MPEG-1, MPEG-2, MPEG-4, MPEG- 7, and MPEG-21

A SVD BASED SCHEME FOR POST PROCESSING OF DCT CODED IMAGES

Video coding standards

InSync White Paper : Achieving optimal conversions in UHDTV workflows April 2015

Information Transmission Chapter 3, image and video

Video compression principles. Color Space Conversion. Sub-sampling of Chrominance Information. Video: moving pictures and the terms frame and

DETECTION OF SLOW-MOTION REPLAY SEGMENTS IN SPORTS VIDEO FOR HIGHLIGHTS GENERATION

CHAPTER 8 CONCLUSION AND FUTURE SCOPE

MPEG has been established as an international standard

Using enhancement data to deinterlace 1080i HDTV

HIGH QUALITY GEOMETRY DISTORTION TOOL FOR USE WITH LCD AND DLP PROJECTORS

Sauter Components

SUBJECTIVE QUALITY EVALUATION OF HIGH DYNAMIC RANGE VIDEO AND DISPLAY FOR FUTURE TV

CERIAS Tech Report Preprocessing and Postprocessing Techniques for Encoding Predictive Error Frames in Rate Scalable Video Codecs by E

SERIES J: CABLE NETWORKS AND TRANSMISSION OF TELEVISION, SOUND PROGRAMME AND OTHER MULTIMEDIA SIGNALS Measurement of the quality of service

Advanced Computer Networks

A Novel Approach towards Video Compression for Mobile Internet using Transform Domain Technique

Adaptive Key Frame Selection for Efficient Video Coding

MULTI-STATE VIDEO CODING WITH SIDE INFORMATION. Sila Ekmekci Flierl, Thomas Sikora

Compression of digital hologram sequences using MPEG-4

+ Human method is pattern recognition based upon multiple exposure to known samples.

Appendix D. UW DigiScope User s Manual. Willis J. Tompkins and Annie Foong

Research Article. ISSN (Print) *Corresponding author Shireen Fathima

Wipe Scene Change Detection in Video Sequences

Upgrading E-learning of basic measurement algorithms based on DSP and MATLAB Web Server. Milos Sedlacek 1, Ondrej Tomiska 2

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur

Browsing News and Talk Video on a Consumer Electronics Platform Using Face Detection

Chapter 2 Circuits and Drives for Liquid Crystal Devices

technical note flicker measurement display & lighting measurement

Case Study: Can Video Quality Testing be Scripted?

MPEG-2. ISO/IEC (or ITU-T H.262)

Chapter 2. Advanced Telecommunications and Signal Processing Program. E. Galarza, Raynard O. Hinds, Eric C. Reed, Lon E. Sun-

Adaptive decoding of convolutional codes

ESI VLS-2000 Video Line Scaler

Region Adaptive Unsharp Masking based DCT Interpolation for Efficient Video Intra Frame Up-sampling

Colour Reproduction Performance of JPEG and JPEG2000 Codecs

Chapter 6: Real-Time Image Formation

OVE EDFORS ELECTRICAL AND INFORMATION TECHNOLOGY

An Efficient Low Bit-Rate Video-Coding Algorithm Focusing on Moving Regions

Region Based Laplacian Post-processing for Better 2-D Up-sampling

RECOMMENDATION ITU-R BT Methodology for the subjective assessment of video quality in multimedia applications

Perceptual Analysis of Video Impairments that Combine Blocky, Blurry, Noisy, and Ringing Synthetic Artifacts

Extraction Methods of Watermarks from Linearly-Distorted Images to Maximize Signal-to-Noise Ratio. Brandon Migdal. Advisors: Carl Salvaggio

DVG-5000 Motion Pattern Option

GLOBAL DISPARITY COMPENSATION FOR MULTI-VIEW VIDEO CODING. Kwan-Jung Oh and Yo-Sung Ho

RECOMMENDATION ITU-R BT Studio encoding parameters of digital television for standard 4:3 and wide-screen 16:9 aspect ratios

OPERATING GUIDE. HIGHlite 660 series. High Brightness Digital Video Projector 16:9 widescreen display. Rev A June A

Topic 4. Single Pitch Detection

United States Patent: 4,789,893. ( 1 of 1 ) United States Patent 4,789,893 Weston December 6, Interpolating lines of video signals

Features 1 Harris and other corners

Improving Color Text Sharpness in Images with Reduced Chromatic Bandwidth

Neat Video noise reduction plug-in for Premiere (Mac)

PAL uncompressed. 768x576 pixels per frame. 31 MB per second 1.85 GB per minute. x 3 bytes per pixel (24 bit colour) x 25 frames per second

Reproducible Quality Analysis of Deinterlacing and Motion Portrayal for Digital TV Displays

Transcription:

Motion Blur Reduction for High Frame Rate LCD-TVs F. H. an Heesch and G. de Haan Philips Research Laboratories High Tech Campus 36, Eindhoen, The Netherlands Abstract Today s LCD-TVs reduce their hold time to preent motion blur. This is best implemented using frame rate upconersion with motion compensated interpolation. The registration process of the TV-signal, by film or ideo camera, has been identified as a second motion blur source, which becomes dominant for TV displays with a frame rate of 00 Hz or higher. In order to justify any further hold time reduction of LCDs, this second type of motion blur, referred to as camera blur, needs to be addressed. This paper presents a real-time camera blur estimation and reduction method, suitable for TV-applications. I. INTRODUCTION At the introduction of LCDs in the TV domain, motion blur was identified as one of the most important picture quality aspects that required improement to compete with the then dominant CRT displays. From the many motion blur reduction methods that hae since been proposed [], Motion Compensated Frame Rate Conersion (MC-FRC) has been the most successful. The MC-FRC method has the adantage that the hold time of a display system can be reduced, without negatie picture quality side-effects such as large area flicker and loss of light efficiency. As a result, 20Hz, 240Hz and een 480Hz LCDs with MC-FRC hae been proposed [2]. The improements in motion blur reduction for such high frame rate displays is limited, howeer, due to the presence of motion blur that originated during registration [3]. The combination of both motion blur sources, i.e. display blur and camera blur, has been inestigated in [3], erified in [4], and is illustrated in Fig.. Camera blur and display blur can both be described as a temporal aeraging filter that, due to motion tracking, is perceied as a spatial aeraging along the motion ector [3]. There are, howeer, two important differences for TV-applications leading to different motion blur reduction algorithms. First, display blur can be known from the display properties, while camera blur needs to be estimated from the TV s input picture. Second, display blur is caused by eye moement relatie to the display, while camera blur is caused by camera motion relatie to the scene. Because of these differences, camera blur reduction requires a post-processing of the TV-signal, while display blur reduction is a pre-processing method. In this paper, the implementation of a real-time camera blur reduction method is discussed. Because it attempts to inert the perceied motion blur, this filter method is referred to as Motion Compensated Inerse Filtering (MCIF). We will show in Section II, that the filter characteristics of motion blur Display blur (ΔT ) 0.75 0.50 0.25 0 0.25 0.50 0.75 Camera blur (TAW) combined blur perceied as: 0.50 0.75 Fig.. Motion blur perceied from ideo that is displayed on an LCD is a combination of display blur (ertical ais) and camera blur (horizontal ais). Both blur sources are relatie to the picture delay, ΔT. The measurement points hae been obtained from a perception test [4]. are straightforwardly deried from theory, but that a practical implementation for remoing this blur is not easily robust. We will describe the signal theory in Section II and from the theory it follows that, for the inerse filter, motion estimation and motion blur estimation are required pre-processing steps. This will be described in Section II-A. In Section II-B, we will discuss two filter implementations and results are shown in Section III. Conclusions are drawn in Section V. II. CAMERA BLUR REDUCTION The perceied motion blur that is caused by an LCD has been described in [5] and can be analyzed in the frequency domain by describing the perceied picture: I f p (f,f t )=I f a (f,f t )A f d (f, f ) A f p(f,f t ), () with Ia f the reconstructed picture on the display, A f d, the spatio-temporal display aperture, A f p, the aperture of the Human Visual System (HVS),, the motion ector corresponding to tracking and f and f t, the spatial and temporal frequencies, respectiely. The leel of perceied motion blur is determined by analyzing the attenuation of the high spatial frequencies of Ia f for f t =0. The reconstructed picture, Ia f, can be epressed to include camera blur, using a spatio-temporal registration aperture: I f a (f,f t )=I f c (f,f t )A f c (f, f ), (2) with Ic f the sampled picture during registration and A f c the registration aperture. In order to directly compare Ic f with the

I (,t) a -2 - b(,t) without blur due to Ac,t 0 2 (a) Camera blur for b(, t) =5 piels. AC(I (,t)) a I (,t) a -2-0 2 (b) The deriatie of I a along the motion ector. (,t) correct range? (,t) consistent? I a(,t) teture? 0 2 3 4 5 6 7 8 9 (c) The AC of (b), ideal (black) and measurement eamples (lines). Fig. 2. The position of the minimum of the AC of the image deriatie corresponds to the blur length. reconstructed picture, Ia f, and to focus on motion blur, we hae ignored spatial scaling and the influence of the spatial camera aperture and will describe the temporal camera aperture by the Temporal Aperture Width (TAW), a t, as reported in [6]: {, if t 2a A c,t (t) = t A f c,t(f t )=a t sinc(πf t a t ), 0, if t > 2a t with sinc(t) = sin(t) t and denoting the Fourier pair. In order to reduce the attenuation caused by camera blur, an enhancement filter can be constructed that inerts the temporal registration aperture, A f c,t( f ), i.e. we can write for the amplitude response of the enhancement: Hc f (f ) = A f = c,t( f ) sinc(π( f )a t ). (3) The amplitude response of this filter has infinite gains and is unstable in the presence of noise. Therefore, we approimate the filter response by a filter that boosts high spatial frequencies along the local motion ector. Furthermore, the amplitude response scales with the TAW. In Section II-A, we will eplain how the TAW is estimated from the input picture, before discussing the filter design in Section II-B. A. Camera blur estimation The estimation of camera blur has been mainly inestigated for image processing in the area of astronomy and still picture processing [7], [8], [9]. In [7], a method is proposed that estimates the blur characteristics from the spatial cepstral domain using multiple sub-images to attenuate signal characteristics in faor of the blur characteristics. The cepstral domain method was replaced in [8] with a method that estimates the blur etent, b(, t), i.e. the length of the blur along the motion ector, using the autocorrelation of the signal s directional intensity deriatie. This method was etended in [9] to be able to cope with motion gradients, net to uniform motions. In this section, we will present a TAW estimation algorithm based on the camera blur estimation from [8] that is suitable for arbitrary motions. & AC[I a ((,t))] argmin(ac) AC(,t) sort & aerage TAW Fig. 3. Camera blur estimation is performed on selected locations. Candidates are discarded based on motion ector size and consistency and based on local image detail (teture). The TAW follows from multiple measurements of the location of the minimum of the AC of the picture deriatie along the motion ector. One of the main limitations of the methods presented in [8] and [9] is a reliable motion estimate. For ideo, howeer, a motion estimator can more reliably determine a motion ector at each position in the image. This simplifies the camera blur estimation, as it allows to normalize the estimated blur etent to a blur time that directly corresponds to the TAW, according to: a t = b(, t) ΔT, (4) (, t) with ΔT the picture period. Furthermore, the TAW can be epected constant within a picture sequence, allowing for locally selectie measurements and the eploitation of temporal consistency to improe robustness by combining multiple measurements. The camera blur estimation algorithm determines the autocorrelation (AC) along the motion ector at selected locations for the signal deriatie. The position of the first minimum in the AC corresponds to the blur length, as eplained by Fig. 2. Combining measurements improes the robustness of the TAW estimate, but further robustness is obtained by discarding measurements that are likely to yield erroneous results, such as measurements at: high speeds: The ideo signal is attenuated as a function of motion. Faster motions result in larger blur etents, while noise can be epected constant. As a result, the SNR, and therefore the reliability of the TAW estimate, reduces for increasing speeds. ery low speeds: For low speeds, the influence of the

spatial camera aperture cannot be ignored. The influence of focus blur, scaling and digital compression would typically lead to an oer-estimate of the camera blur. flat areas: A low SNR reduces the reliability of the camera blur estimate. Therefore, flat areas are detected using a local actiity metric and remoed from the blur etent measurement. motion edges: The epression of camera blur as stated in Eq. 3 does not hold when locally the motion ector is not constant. In this case, the estimates at edges in the motion ector field are unreliable, een if the motion estimates themseles are accurate. The locations are found by erifying the consistency of the motion along its ector. Locations that fail the consistency check are discarded. The robustness is further improed by using the resemblance of the autocorrelation function with respect to the ideal shape, shown in Fig. 2(c), as a confidence measure. For eample by aeraging the results for those measurements meet a certain confidence threshold. This is schematically shown in Fig. 3. It has been demonstrated that, for HD TV-signals [4], this method is quite resilient to both analog and MPEG noise, and can cope with low contrasts and focal blur. The method fails to proide an accurate estimate only in case of ery little motion, or in case of ery high noise leels in the picture or motion ector field. In these cases, howeer, it is unlikely that camera blur reduction will yield a significant improement in picture quality. B. Inerse filtering The camera blur estimate, a t, and the motion estimate, (, t), enable the implementation of the sharpness enhancement filter described by Eq. 3. We refer to this method as inerse filtering, although it is not the goal to obtain the best approimation to the theoretical inerse filter. Instead, we attempt to inert the perceied reduction in sharpness. The infinite gains of the amplitude response complicate a practical implementation of the inerse filter and an approimation is required. An implementation that uses a FIR filter, known as Motion Compensated Inerse Filtering (MCIF), has been proposed in [5] and [0] for display blur reduction. For display applications, MCIF filtering applies a spatial filter for each piel in the picture along the motion ector that corresponds to the motion tracking of the HVS. Therefore, a D FIR-filter is rotated and scaled to align with the motion ector. In practice, such filtering can be implemented by fetching the luminance of I a (, t) at an interpolated piel grid along this scaled and rotated ector using e.g. bilinear interpolation. For camera blur, the FIR filter also scales linearly with the TAW as illustrated in Fig. 4. In this section, we will discuss MCIF for camera blur reduction using two inerse filter design methods: Luminance Transient Improement (LTI) [] and Trained Filtering (TF) [2]. I c() Δ Δ2 Fig. 4. Filtering along the motion ector requires fetching of intensity alues on a non-integer grid along the (scaled) motion ector. The scaled motion ector, indicated by the white arrow, determines the position of the sample positions, indicated by the black dots. The intensity alues at these locations are calculated from the nearest intensities. This is illustrated for the bottom left sample position. In case of bilinear interpolation, its intensity is calculated from the surrounding 4 piel locations (white dots), using the distances Δ and Δ 2 to proportionally weight the intensities. The LTI method, known from static sharpness enhancement, can be eplained as a two-step process. First, a sharpness enhancement filter reduces camera blur at the cost of oershoots at edges. Second, the oershoots at edges are eliminated by clipping the filter output to the local intensity etremes, as illustrated for a D signal (along the motion ector) in Fig. 5. Note that the method does not attempt to approimate an ideal inerse filter, instead it enhances mid-frequencies, limiting the enhancement of noise. This filter method was found to perform well at edges, but yields only a modest sharpness enhancement in tetured areas. Also, this method was found to be still sensitie to noise. Improed performance and robustness is obtained by linking a noise estimator to the filter gain and by making the filter coefficients adaptie to the local picture structure. A second filter design strategy uses a TF method to find Mean Square Error (MSE) optimal inerse filters. With a training set of degraded and original ideo data, containing motion blurred and motion-blur free content, respectiely, the filter that obtains the minimum difference between the two data sets can be determined. This requires a classification method to sort data similarity, an energy function that defines a b(,t) (a) Fetching piels. b I (,t) a Intensity 0 a b(,t) * h C blurred edge highpass added clipped (b) Filtering and clipping. Fig. 5. The LTI method: piels are fetched at line segment [a-b] at an interpolated grid along the motion ector,. The FIR filter is scaled to match the blur length. A highpass filter is added to the blurred edge (red line), yielding oershoots at edges (blue line). Clipping to the local intensity etremes of the blurred edge reduces the oershoots (green line). b

filter window reorder DR = 00000000 0 class code (a) I (,t) a filter table filter coefficients (b) Fig. 6. The TF method: (a) The filter window consists of a line along the motion ector and a bo around the center piel containing 7 sample locations. The intensities within this window are fetched from an interpolated grid. (b) A 9-bit class code is constructed by comparing the sampled intensities with the window aerage, yielding 7 bits, together with a 2-bit code for the DR. The class code is the inde of the filter table. (a) (b) a quantifiable difference metric, and a representatie data set of sufficient size. The classification encodes local picture structure using a window aligned along the scaled motion ector. In order to constrain the number of classes to a practical number for training, the number of piels in the window is limited and the structure is typically encoded with one bit per piel, e.g. 0 and correspond to a lower or higher luminance alue compared to the window aerage, respectiely. The classification process for camera blur reduction has been thoroughly researched in [3]; an etensie search of all possible symmetric filter structures using up to 9-bits for the classification, as illustrated in Fig. 6(a), using one bit per piel and an additional two bits to encode the Dynamic Range (DR), yields an MSE-optimal filter structure as illustrated in Fig. 6(b). For this ealuation, a test set of 45 HD images was used. The robustness to noise is obtained with the encoding of the dynamic range, while the content adaptation is encoded by the filter structure. The disadantage of this method is the large filter table that is required to store all filter coefficients. Furthermore, a re-training is required for eery change in (the tuning of) the classification. Special care is required for the borders of moing objects, i.e. occlusion areas. The inerse filter should take occlusion areas into account and discriminate between coering and uncoering areas, because the motion blur model, described by Eq. 3 does not hold here. Visible artifacts at occlusion areas can be alleiated by filling in the occlusion data from a preious and net picture, or by gradually limiting the enhancement at occlusion areas. III. RESULTS The MCIF filter using the LTI and TF methods hae been applied to the eample HD picture shown in Fig. 7(a) containing a partial panning motion of 30 piels per frame. The HD picture has been registered with a TAW of a t =0.5. The filtering methods hae been applied to the luminance signal (c) Fig. 7. (a) A picture from the Trainloop sequence, containing a horizontal motion of 30 piels per frame, registered with a TAW of a t =0.5. (b) A close-up of the luminance component, illustrating the camera blur. (c), (d) The same close-up after applying the LTI and TF methods, respectiely. only. The filter outputs are illustrated in Fig. 7(c) and Fig. 7(d). In addition, the camera blur reduction methods hae been applied to a test set and ealuated with an informal subjectie assessment on a high frame LCD-TV. From this assessments, we found the LTI method to be most robust to noise, but limited in reconstructing details in tetured regions. The trained filter method was found to perform best for recreating details, without creating isible oershoots at edges, although a tuning for seeral types of picture degradations was found more cumbersome. Further work is required to quantify the quality improement of these methods and to determine which is most suitable for camera blur reduction in LCD-TV applications. IV. DISCUSSION To achiee the best perceied picture quality when implementing camera blur reduction for LCD-TVs, other picture quality improements that might be present in between a TV s picture reception and rendering must be taken into account. In particular, the combination of MCIF with spatial sharpness enhancement (SSE) [] and MC-FRC (for display blur reduction) cannot be considered independently. MC-FRC and MCIF both use motion ector estimates and hae to tackle isible artifacts that can appear when moing objects occlude. In addition, the eecution order of (d)

MC-FRC and MCIF need to be considered. Applying MCIF after MC-FRC results in higher computational requirements, while applying MCIF before MC-FRC, influences the frame interpolation of MC-FRC. When combining MCIF with SSE, in general, care has to be taken that the sharpness enhancements of both processing steps do not oerlap. In particular for low speeds, MCIF can influence the spatial sharpness, causing an oer-sharpening in combination with SSE. From informal eperiments, we found that isible artifacts that result from combining MCIF with SSE or MC-FRC are highly implementation specific and, therefore, need to be addressed on a case-by-case basis. V. CONCLUSIONS For ideo and film content, motion blur can only be reduced on high frame rate LCD-TVs by reducing camera blur. The blur characteristics of camera blur are described by the TAW which needs to be estimated from the TV-signal. MCIF is required for camera blur reduction. A camera blur estimation method, two filter design strategies and a system ealuation hae been implemented. The TF method was found to perform best in recreating details, while the LTI method was found to be most robust to noise. To implement MCIF for TV applications, the combination with SSE and MC-FRC need to be considered to optimize picture quality and required computational resources. REFERENCES [] F. H. an Heesch and M. A. Klompenhouwer, Video processing for optimal motion portrayal, in Proc. of the IDW, 2006, pp. 993 996. [2] Hyun-G et al., 45.4: High Frequency LVDS Interface for F-HD 20Hz LCD TV, SID Symposium Digest of Technical Papers, ol. 39, no., pp. 68 684, 2008. [3] M. A. Klompenhouwer, 22.2: Dynamic Resolution: Motion Blur from Display and Camera, SID Symposium Digest of Technical Papers, ol. 38, no., pp. 065 068, 2007. [4] W. Yarde, Estimating motion blur on liquid crystal displays, Master s thesis, Eindhoen Uniersity of Technology, 2009. [5] M. A. Klompenhouwer and L. J. Velthoen, Motion blur reduction for liquid crystal displays: Motion compensated inerse filtering, in Proc. of the SPIE, ol. SPIE-5308, 2004, pp. 690 699. [6] F. H. an Heesch and M. A. Klompenhouwer, 26.4: The Temporal Aperture of Broadcast Video, SID Symposium Digest of Technical Papers, ol. 39, no., pp. 370 373, 2008. [7] T. M. Cannon, Blind deconolution of spatially inariant image blurs with phase, ASSP, ol. 24, no., pp. 58 63, February 975. [8] Y. Yitzhaky and N. S. Kopeika, Identification of the Blur Etent from Motion Blurred Images, Graphical Models and Image Processing, ol. 59, no. 5, pp. 30 320, 997. [9] H. J. Trussell and S. Fogel, Restoration of spatially ariant motion blurs in sequential imagery, Image Processing Algorithms and Techniques II, ol. 452, no., pp. 39 45, 99. [0] C. Dolar, LCD-Modelle und ihre Anwendung in der Videosignalerarbeitung, Ph.D. dissertation, Technische Uniersität Dortmund, 200. [] J. A. P. Tegenbosch et al., Improing nonlinear up-scaling by adapting to the local edge orientation, Visual Communications and Image Processing 2004, ol. 5308, no., pp. 8 90, 2004. [2] M. Zhao, Video Enhancement Using Content-adaptie Least Mean Square Filters, Ph.D. dissertation, Eindhoen Uniersity of Technology, 2006. [3] G. Kwintenberg, Motion-Blur Reduction for Liquid Crystal Displays, Master s thesis, Eindhoen Uniersity of Technology, 2009.