Improving Color Text Sharpness in Images with Reduced Chromatic Bandwidth

Similar documents
The Development of a Synthetic Colour Test Image for Subjective and Objective Quality Assessment of Digital Codecs

Colour Reproduction Performance of JPEG and JPEG2000 Codecs

Processing. Electrical Engineering, Department. IIT Kanpur. NPTEL Online - IIT Kanpur

DELTA MODULATION AND DPCM CODING OF COLOR SIGNALS

!"#"$%& Some slides taken shamelessly from Prof. Yao Wang s lecture slides

Minimizing the Perception of Chromatic Noise in Digital Images

Murdoch redux. Colorimetry as Linear Algebra. Math of additive mixing. Approaching color mathematically. RGB colors add as vectors

Color Spaces in Digital Video

[source unknown] Cornell CS465 Fall 2004 Lecture Steve Marschner 1

Fundamentals of Multimedia. Lecture 3 Color in Image & Video

Television History. Date / Place E. Nemer - 1

Multimedia. Course Code (Fall 2017) Fundamental Concepts in Video

Chrominance Subsampling in Digital Images

UNIVERSAL SPATIAL UP-SCALER WITH NONLINEAR EDGE ENHANCEMENT

Ch. 1: Audio/Image/Video Fundamentals Multimedia Systems. School of Electrical Engineering and Computer Science Oregon State University

Calibration Best Practices

ECE438 - Laboratory 4: Sampling and Reconstruction of Continuous-Time Signals

Vannevar Bush: As We May Think

An Overview of Video Coding Algorithms

Multimedia Systems Video I (Basics of Analog and Digital Video) Mahdi Amiri April 2011 Sharif University of Technology

Understanding Human Color Vision

DCI Requirements Image - Dynamics

Lecture 2 Video Formation and Representation

RECOMMENDATION ITU-R BT (Questions ITU-R 25/11, ITU-R 60/11 and ITU-R 61/11)

Audio and Video II. Video signal +Color systems Motion estimation Video compression standards +H.261 +MPEG-1, MPEG-2, MPEG-4, MPEG- 7, and MPEG-21

ECE 5765 Modern Communication Fall 2005, UMD Experiment 10: PRBS Messages, Eye Patterns & Noise Simulation using PRBS

Chapter 3 Fundamental Concepts in Video. 3.1 Types of Video Signals 3.2 Analog Video 3.3 Digital Video

A Perceptual Distortion Metric for Digital Color Video

Using the NTSC color space to double the quantity of information in an image

High Quality Digital Video Processing: Technology and Methods

A Color Scientist Looks at Video

A Comparative Study of Color and Contrast Enhancement for Still Images and Consumer Video Applications

Toward Better Chroma Subsampling By Glenn Chan Recipient of the 2007 SMPTE Student Paper Award

Man-Machine-Interface (Video) Nataliya Nadtoka coach: Jens Bialkowski

TERMINOLOGY INDEX. DME Down Stream Keyer (DSK) Drop Shadow. A/B Roll Edit Animation Effects Anti-Alias Auto Transition

Interface Practices Subcommittee SCTE STANDARD SCTE Measurement Procedure for Noise Power Ratio

Chapter 4 Color in Image and Video. 4.1 Color Science 4.2 Color Models in Images 4.3 Color Models in Video

Appendix D. UW DigiScope User s Manual. Willis J. Tompkins and Annie Foong

decodes it along with the normal intensity signal, to determine how to modulate the three colour beams.

Preventing Illegal Colors

Colour Matching Technology

The Lecture Contains: Frequency Response of the Human Visual System: Temporal Vision: Consequences of persistence of vision: Objectives_template

Assessing and Measuring VCR Playback Image Quality, Part 1. Leo Backman/DigiOmmel & Co.

Spatial-frequency masking with briefly pulsed patterns

Research and Development Report

BTV Tuesday 21 November 2006

Video Signals and Circuits Part 2

RECOMMENDATION ITU-R BT Studio encoding parameters of digital television for standard 4:3 and wide-screen 16:9 aspect ratios

Introduction & Colour

ELEC 691X/498X Broadcast Signal Transmission Fall 2015

Motion Video Compression

LCD and Plasma display technologies are promising solutions for large-format

Dither Explained. An explanation and proof of the benefit of dither. for the audio engineer. By Nika Aldrich. April 25, 2002

Masking in Chrominance Channels of Natural Images Data, Analysis, and Prediction

Politecnico di Torino HIGH SPEED AND HIGH PRECISION ANALOG TO DIGITAL CONVERTER. Professor : Del Corso Mahshid Hooshmand ID Student Number:

N T I. Introduction. II. Proposed Adaptive CTI Algorithm. III. Experimental Results. IV. Conclusion. Seo Jeong-Hoon

To discuss. Types of video signals Analog Video Digital Video. Multimedia Computing (CSIT 410) 2

ZONE PLATE SIGNALS 525 Lines Standard M/NTSC

Color Nonuniformity in Projection-Based Displays: Analysis and Solutions

CERIAS Tech Report Preprocessing and Postprocessing Techniques for Encoding Predictive Error Frames in Rate Scalable Video Codecs by E

An Alternative Architecture for High Performance Display R. W. Corrigan, B. R. Lang, D.A. LeHoty, P.A. Alioshin Silicon Light Machines, Sunnyvale, CA

Impact of scan conversion methods on the performance of scalable. video coding. E. Dubois, N. Baaziz and M. Matta. INRS-Telecommunications

By David Acker, Broadcast Pix Hardware Engineering Vice President, and SMPTE Fellow Bob Lamm, Broadcast Pix Product Specialist

DVM-3000 Series 12 Bit DIGITAL VIDEO, AUDIO and 8 CHANNEL BI-DIRECTIONAL DATA FIBER OPTIC MULTIPLEXER for SURVEILLANCE and TRANSPORTATION

INTERNATIONAL JOURNAL OF ELECTRONICS AND COMMUNICATION ENGINEERING & TECHNOLOGY (IJECET)

CSE Data Visualization. Color. Jeffrey Heer University of Washington

Essence of Image and Video

Analysis of WFS Measurements from first half of 2004

PCM ENCODING PREPARATION... 2 PCM the PCM ENCODER module... 4

Reducing tilt errors in moiré linear encoders using phase-modulated grating

The Art and Science of Depiction. Color. Fredo Durand MIT- Lab for Computer Science

Supplementary Course Notes: Continuous vs. Discrete (Analog vs. Digital) Representation of Information

Perceptual Analysis of Video Impairments that Combine Blocky, Blurry, Noisy, and Ringing Synthetic Artifacts

Lab 5 Linear Predictive Coding

ANTENNAS, WAVE PROPAGATION &TV ENGG. Lecture : TV working

Wide Color Gamut SET EXPO 2016

Dan Schuster Arusha Technical College March 4, 2010

CHAPTER 3 COLOR TELEVISION SYSTEMS

1. Broadcast television

5.1 Types of Video Signals. Chapter 5 Fundamental Concepts in Video. Component video

APPLICATION NOTE AN-B03. Aug 30, Bobcat CAMERA SERIES CREATING LOOK-UP-TABLES

AN ARTISTIC TECHNIQUE FOR AUDIO-TO-VIDEO TRANSLATION ON A MUSIC PERCEPTION STUDY

Multirate Digital Signal Processing

DCI Memorandum Regarding Direct View Displays

DRAFT. Proposal to modify International Standard IEC

High Value-Added IT Display - Technical Development and Actual Products

Audiovisual Archiving Terminology

White Paper. Uniform Luminance Technology. What s inside? What is non-uniformity and noise in LCDs? Why is it a problem? How is it solved?

Evaluation of the Color Image and Video Processing Chain and Visual Quality Management for Consumer Systems

Region Adaptive Unsharp Masking based DCT Interpolation for Efficient Video Intra Frame Up-sampling

Self-Calibrating Wide Color Gamut High Dynamic Range Display

CATHODE RAY OSCILLOSCOPE. Basic block diagrams Principle of operation Measurement of voltage, current and frequency

The Color Reproduction Problem

Composite Video vs. Component Video

The XYZ Colour Space. 26 January 2011 WHITE PAPER. IMAGE PROCESSING TECHNIQUES

3/2/2016. Medical Display Performance and Evaluation. Objectives. Outline

A Novel Approach towards Video Compression for Mobile Internet using Transform Domain Technique

Visual Imaging and the Electronic Age Color Science

Selected Problems of Display and Projection Color Measurement

Department of Electrical & Electronic Engineering Imperial College of Science, Technology and Medicine. Project: Real-Time Speech Enhancement

Transcription:

Improving Color Text Sharpness in Images with Reduced Chromatic Bandwidth Scott Daly, Jack Van Oosterhout, and William Kress Digital Imaging Department, Digital Video Department Sharp aboratories of America Camas, WA USA Abstract Current digital imaging systems often reduce the chromatic bandwidth to speed transmission or save storage. While this has become routine for most video applications, it is also becoming true of digital camera systems. Chromatic bandwidth is usually reduced by converting the RGB image into a single luminance and two color difference images, and then subsampling the color difference images. The motivation lies in the reduced bandwidths of the visual system for the color signals, as evidenced by contrast sensitivity function (CSF) data. Systems using chromatic subsampling can exhibit visual equivalence to otherwise identical non-subsampled systems, despite the physical color blur introduced by subsampling. While this equivalence holds true for images of real scenes and optically captured scenes, problems arise with synthetically generated imagery, such as color geometrics and type fonts. The possible sources of these problems include: 1) deviation from natural image chromatic power spectra, 2) nonlinear masking effects between the luminance and chrominance visual mechanisms, and 3) filtering processes in the luminance-chrominance domain create signals outside of the RGB gamut. By taking advantage of nonlinear perceptual effects and the gamut clipping issues, we have developed a simple point-processing algorithm that accomplishes a reduction in the visibility of chromatic blur. Introduction Since the early days of color television it has been known that the spatial bandwidths of human color mechanisms are substantially less than that for the luminance [1]. This is most dramatic when the color signals are formed as isoluminant color differences, generally having a red-green (R/G) and a blue-yellow (B/Y) modulation. By recombining these color difference signals with the luminance signal at the receiver, the original RGB image can be reconstructed. Modern data is shown below in Figure 1 for the contrast sensitivity functions (CSF) of the visual system 1 to luminance and chrominance signals [2]. The maximum perceived frequencies, or cut-off frequencies, of this data suggest that the bandwidths of luminance to R/G to B/Y are nearly: 1 : 0.5 : 0.5. Contrast Sensitivity 10 3 uminance and Chromatic CSFs 10 2 10 1 - um -- RG -. BY 10 0 10-1 10 0 10 1 10 2 Spatial Frequency (cy/deg) Figure 1: uminance and Chromatic CSFs of the visual system Different color spaces for the chromatic modulations have been used in both vision science and imaging system design. Examples include YIQ, YUV, ATD, and Y/R-Y/B-Y. One of the most predominant is the *a*b* representation, where a* is the red-green modulation and b* is the blue-yellow. Although these are not the exact color directions currently believed to be used by the visual system, it is still a useful space within which to work. Although the *a*b* representation was intended for color reproduction issues and has zero frequency components, an extension referred to as Spatial-CIEAB [3], incorporates spatial frequency. In this extension, the three CSFs shown in Figure 1 are modeled as Gaussian lowpass filters and applied to the *, a*, and b* layers of an image. This is considered a perceptually uniform image space from which spatial E values can be computed. However, there are still unresolved issues with using the a* and b* layers of an image for the main color difference signals in human vision. These relate to the order of 1 The CSF is roughly analogous to a spatial frequency response. 387 386

nonlinearity between the MS cone action spectra ( XYZ) and the cones static nonlinearity ( the cube-root in *) prior to the formation of the opponent color signals ( a* and b*) used to represent color in Spatial CIEAB. Engineering uses of the bandwidth effects of the luminance and two chrominance CSFs existed before the effects were well understood. Although early use was made by an analog low-pass filter for video, the predominant approach now is by subsampling the color difference layers of a digital image. The bandwidth reductions shown in Figure 1 suggest a scheme that subsamples the chromatic layers by 2 in each direction (with appropriate low-pass filtering). This commonly used technique works well on natural images or optically captured contone images. In such images, the distortions due to subsampling the chrominance 2 are not visible to the eye. In addition, when high resolution text (i.e., greater bandwidth than the contone image) is inserted into the continuous tone images, the sharpness of the black, grey and white text is not affected. Chromatic Bandwidth Reduction Problems Unfortunately, there are two main problems caused by chromatic subsampling (or other chromatic spatial bandwidth reduction). These occur on images containing color text or graphical objects with binary edge transitions. One problem manifests itself as a visible chromatic blur (also referred to as color bleeding) on text 3 of saturated colors. This visibility is dependent on the grey level, where it is most visible on white, and not visible at all with chromatic text on black backgrounds. The second problem is that of overshoots or undershoots around the edges (often referred to as haloes). These are similar to ringing artifacts but consist of only one ring. One could ask why the visual theory derived from Figure 1 fails in this case of colored text or graphical objects. This problem occurs even if the Spatial-CIEAB domain is used, in which case it fails both along the luminance axis 4 as well as the spatial frequency axis 5. 2 We used a* and b* of CIEAB, but any color difference signal would give similar results. 3 The term text will refer to both text and graphics objects. 4 Because the chromatic blur is not equally visible throughout the * grey scale. 5 Because the chromatic blur is visible with a spatial subsampling of only a factor of 2.0. Secondary Visual System Attributes The failure of the visual theory derived from the three CSFs is that it is over-simplified. The under-prediction of color bleeding is primarily due to viewing distance issues, the center of chromatic modulation, and cross luminance-chrominance facilitation. These three aspects will be discussed below. The other distortion, the overshoot adjacent to sharp edges, is primarily a gamut-clipping effect and results from the low-pass filtering step which creates values outside the range of the original RGB (or XYZ). Unless special precautions are taken into account when converting from a color transform such as *a*b* back to a directly displayable color format, orthogonal projections of the out-of-gamut color to the display space results in local luminance and hue shifts. These appear as halo artifacts. Due to the limited space, we will only expound on the perceptual issues. Viewing Distance and Nyquist Mapping The common engineering misunderstanding relating to the CSFs of Figure 1 is to believe that since the cut-off spatial frequencies are in the ratio of approximately 1 : 0.5 : 0.5 for luminance to R/G to B/Y, a system designed with those relative bandwidths will not exhibit chromatic blur. This is only true if the system s luminance Nyquist is high enough that it exceeds the luminance cut-off visual frequency. The mapping of the system s Nyquist to visual frequency in cy/deg depends on both the maximum physical frequency as well as the viewing distance. If the Nyquist maps to less than the visual cut-off frequency (in such a chromatically subsampled system), the individual Nyquists of the luminance and two chrominance planes all intersect their respective CSFs. The consequence is that the undersampling with respect to luminance will result in luminance blur, the undersampling with respect to the R/G CSF will cause a R/G chromatic blur, and similarly for the B/Y plane. However, in such systems which have perceivable luminance blur, the color blur is usually not noticed for natural or optically captured images. This is probably due the lack of spatial power spectra at the higher frequencies for the color difference images as well as the secondary visual effects of luminance masking chrominance at high luminance contrasts. However, with images containing digitally generated text, the chromatic blur can be quite visible. Center of Chromatic Modulation Most of the physiologically-motivated studies [2,4] of the chromatic CSFs modulate the R/G signals about a yellow mean level; assumed to be the best isolation of the visual luminance signal. However, since the space-time 388 387

average of most images is grey, engineering-motivated studies [5,6] modulate about grey. Substantial differences result between these two approaches. Modulation about grey results in R/G and B/Y CSFs which are bandpass rather than lowpass. Also, the bandwidths of the R/G and B/Y signals are much less relative to luminance 6. While these effects can be partly explained by luminance contamination of the R/G and B/Y signals due to chromatic adaptation and gamma effects, there is still debate over which is the most appropriate for practical imaging. The bandpass shapes may represent an increased R/G and B/Y sensitivity over the physiological studies CSFs obtained using a yellow mean. Cross uminance-chrominance ing & Facilitation As mentioned earlier in this section there are masking effects that act across luminance and chrominance channels of the visual system [7]. While high contrast luminance signals can mask the chrominance signals, lowto mid-luminance contrasts actually facilitate (reduces the threshold) chrominance signals of similar frequencies, making them easier to see. This is shown in Figure 2 below, where the R/G threshold is plotted as a function of a luminance masking contrast. For luminance mask contrasts in the range of 1% to 20%, the R/G threshold is actually reduced. Threshold Contrast (%) 10 0 0 deg phase chromatic contrast. The areas around a color edge have low luminance contrast for the frequencies higher than the chromatic CSF cut-off frequency. The difference between the original and the subsampled chromatic signal (i.e, the color blur) also consists of frequencies higher than this cutoff. Without any interaction from the luminance signal, this chromatic difference signal would not be detected. However, the facilitation of the chrominance signal by the lower contrast luminance signals of similar frequency, as shown in Figure 2, reduces the threshold of the chrominance signal. The chrominance signal due to the color blur now lies above this reduced threshold, and is visible. Chromatic Resolution Enhancement Algorithm Our goal is to correct for the subset of imagery where chromatic subsampling based on the simpler CSF visual model fails due to the reasons described in the previous section. Consequently, we concentrated on text and graphical objects. This acted to differentiate our work from many television applications toward increasing color resolution. The basic idea of our algorithm is to use the luminance signal as a control signal for the processing of the chrominance signals. In particular, we limit the range of the magnitude of the chrominance signals according to the luminance level, as shown in Figure 3. This is motivated by our observation that the chromatic blur was most visible on a white background. Also, the color gamut in the luminance-chrominance space narrows to a point at the minimum and maximum luminance values. That is, the values should be zero for both chrominance signals at the luminance minimum and maximum. C1 Create Modify C1 by W 1. C1 N E W W 2 10-1 10-1 10 0 10 1 10 2 Contrast (%) C2 Modify C2 by C2 N E W Figure 2: R/G Threshold vs. uminance, data from [7] We believe this is the primary cause of the chromatic blur problem. Since isoluminant text are rarely used due to their illegibility, the chromatic text contains both luminance and 6 One study [6] has luminance to RG to BY bandwidths of 1:0.3:0.1 Figure 3: Algorithm with luminance-controlled mask function Since we are mostly concerned with chromatic text on a white surrounding area, which is the most visible as well as most common case, we can remove the color bleeding from the chromatic characters with the following algorithm. Since the luminance signal is not distorted by the chromatic subsampling and we know that both chrominance signals should be zero in the white area surrounding the 389 388

text, our algorithm forces them to zero if the control signal (mask) from the pixel s luminance indicates that it is white or near white. The consequence is that we can achieve a spatial resolution enhancement in the chromatic text without resorting to spatial operators, such as filtering and other local spatial operators. The available high resolution luminance signal is used to invoke a higher resolution into the chrominance signal. The scheme is essentially a pointprocessing scheme across the *, a*, and b* images, generically indicated in the block diagrams as, C 1, C 2. Consequently, it is computationally inexpensive, especially as compared with other spatial approaches. back to *a*b* for this analysis). This is why the figure shows a blurring of even the non-subsampled luminance signal if the algorithm is not used. The algorithm strongly reduces any distortions introduced into the luminance signal by the *a*b* to RGB mapping. The algorithm is generally applied to the *, a* and b* images after a* and b* are interpolated to the * size from a compressed image buffer. It is also possible, however, to use the algorithm as a postprocessing technique to clean up color text from any source, such as digital video or digital still cameras that may output in the RGB format. 250 Amplitude of *, a*+128, & b*+128 200 150 100 50 original subsampled algorithm 0.0 X C o u t C in 0.0 c Figure 5: Modification of chrominance signal by the mask signal o d e 0 105 106 107 108 109 110 111 112 113 114 115 Pixel Position Figure 4: Algorithm effects on *, a*, and b* fontstroke profiles The resulting resolution enhancement is exemplified in Figure 4. In this figure, the curves are the cross-sections of a chromatic text fontstroke. The uppermost three curves are * signals, the middle three are a* signals, and the lowest are b* signal traces. The original *, a*, and b* chromatic text signals are shown as solid lines, while the subsampled (and upsampled back to original resolution) *, a*, and b* signals are shown as the long-dash short-dash curves. The result of the algorithm for the *, a*, and b* signals are shown as the dashed curves 7. One can see that the algorithm s a* and b* signals have steeper slopes than the subsampled signals, resulting in a sharper appearance. They also return to 0 (i.e., 128 in this plot) nearer to the original stroke edge than the subsampled signals, corresponding to less visible color spread (or color bleeding). Note that these signals are taken after the *a*b* image has been converted to RGB (then converted 7 This is hard to see for *, since it is nearly superimposed on the original * curve. The details of the modification of the chrominance images are shown in Figure 5. The simpler version use a linear transfer of C in to C out (dashed line) and multiplies this result by the mask signal. This merely scales the chrominance signal about zero (where the image is achromatic) by the mask. In this approach, the mask essentially controls the resulting slope as governed by the equation, C ou t = C in C mod (1) where C indicates the chromatic images from either a* or b*, C out is the chromatic image s pixel value, C in is the pixel s input value having the chromatic bandwidth reduction, C mod is the dashed line in Figure 4, and is a pixel dependent value to be described shortly. Independent control over the way the mask signal affects the C1 and C2 chrominance images signals is accomplished with the weights shown in Figure 3. A more detailed approach is to use the mask to control the degree of limiting of the C mod function, shown as the solid line in Figure 5. This is accomplished by using the solid line transfer function on the chrominance signal which is subsequently multiplied by the mask. The net effect is that the output chrominance value at each pixel is given by the following equation: C ou t = C in C mod () (2) 390 389

where C mod is the nonlinear transfer curve shown solid in Figure 5, whose shape is determined from the mask value, which directly controls its limiting values. Figure 6 shows the details of the generation of the masking control signal generated from the luminance image,. Again a simple version as well as a more sophisticated version are shown. The dashed line creates a simple threshold signal with binary values of 0.0 and 1.0. The threshold value is chosen so that the chrominance signal will only be affected in the brightest parts of the image, where the chromatic blur is most visible. It is chosen to be high enough that it does not affect the largearea tinted white regions, since the visual system is insensitive to the actual color of small specular highlights due to the chromatic CSFs. ower threshold values will act to increase the magnitude of sharpening 8. The result is that the C1 and C2 signal values are set to zero since the mask value is zero and the chrominance is scaled by the mask. References 1. O. Schade, On the quality of color-television images and the perception of color detail, SMPTE Journal V. 67, pp 801-819 (1958). 2. K. T. Mullen, J. The contrast sensitivity of human colour vision to red-green and blue-yellow chromatic gratings, Physiol V. 359, pp. 381-400 (1985). 3. X. M. Zhang and B. A. Wandell, A spatial extension to CIEAB for digital color image reproduction, Proceedings of the SID Symposiums (1996) 4. D. Williams, N. Sekiguchi, and D. Brainard, Color, contrast sensitivity, and the cone mosaic, Proc. Natl. Acad. Sci. USA V. 90, pp. 9770-9777 (1993) 5. A. Watanabe, H. Sakata, and H. Isono, Chromatic spatial sine-wave responses of the human visual system, NHK aboratories Note, Serial # 198, March (1976) 6. W. E. Glenn, K. G. Glenn, and C. J. Bastian, Imaging system design based on psychophysical data, Proceedings of the SID, Vol 26, pp. 71-78 (1985). 7. E. Switkes, A. Bradley, and K. De Valois, Contrast dependence and mechanisms of masking interactions among chromatic and luminance gratings, JOSA A. V. 5, pp1149-1162 1.0 c o d e c o d e th re s h Figure 6: Generation of mask signal from luminance image The more sophisticated version of the mask signal attempts to track the maximum chromatic extent as a function of luminance, in accordance with the overall color gamut possible at the printer. This uses a threshold as a function of grey level, shown as the solid line in Figure 6. Conclusions Images displayed with this algorithm demonstrate a significant decrease in the visibility of color blur. It has some difficulty with yellow text, due to its high luminance confounding the mask threshold setting. Fortunately, the use of yellow text on white backgrounds is generally avoided by the experienced user due to its illegibility. 8 In Figure 4, a threshold value of 250 was used. 391 390