PERCEPTUAL QUALITY ASSESSMENT FOR VIDEO WATERMARKING. Stefan Winkler, Elisa Drelie Gelasca, Touradj Ebrahimi

Similar documents
Video Quality Evaluation with Multiple Coding Artifacts

Understanding PQR, DMOS, and PSNR Measurements

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video

SUBJECTIVE QUALITY EVALUATION OF HIGH DYNAMIC RANGE VIDEO AND DISPLAY FOR FUTURE TV

Video Quality Evaluation for Mobile Applications

Evaluation of video quality metrics on transmission distortions in H.264 coded video

A Novel Approach towards Video Compression for Mobile Internet using Transform Domain Technique

Colour Reproduction Performance of JPEG and JPEG2000 Codecs

A SUBJECTIVE STUDY OF THE INFLUENCE OF COLOR INFORMATION ON VISUAL QUALITY ASSESSMENT OF HIGH RESOLUTION PICTURES

CERIAS Tech Report Preprocessing and Postprocessing Techniques for Encoding Predictive Error Frames in Rate Scalable Video Codecs by E

Channel models for high-capacity information hiding in images

SERIES J: CABLE NETWORKS AND TRANSMISSION OF TELEVISION, SOUND PROGRAMME AND OTHER MULTIMEDIA SIGNALS Measurement of the quality of service

Comparative Study of JPEG2000 and H.264/AVC FRExt I Frame Coding on High-Definition Video Sequences

QUALITY ASSESSMENT OF VIDEO STREAMING IN THE BROADBAND ERA. Jan Janssen, Toon Coppens and Danny De Vleeschauwer

Contents. xv xxi xxiii xxiv. 1 Introduction 1 References 4

An Efficient Low Bit-Rate Video-Coding Algorithm Focusing on Moving Regions

Project No. LLIV-343 Use of multimedia and interactive television to improve effectiveness of education and training (Interactive TV)

PERCEPTUAL QUALITY OF H.264/AVC DEBLOCKING FILTER

ABSTRACT 1. INTRODUCTION

Objective video quality measurement techniques for broadcasting applications using HDTV in the presence of a reduced reference signal

AUDIOVISUAL COMMUNICATION

PERCEPTUAL QUALITY COMPARISON BETWEEN SINGLE-LAYER AND SCALABLE VIDEOS AT THE SAME SPATIAL, TEMPORAL AND AMPLITUDE RESOLUTIONS. Yuanyi Xue, Yao Wang

Lecture 2 Video Formation and Representation

OBJECT-BASED IMAGE COMPRESSION WITH SIMULTANEOUS SPATIAL AND SNR SCALABILITY SUPPORT FOR MULTICASTING OVER HETEROGENEOUS NETWORKS

Perceptual Analysis of Video Impairments that Combine Blocky, Blurry, Noisy, and Ringing Synthetic Artifacts

Intra-frame JPEG-2000 vs. Inter-frame Compression Comparison: The benefits and trade-offs for very high quality, high resolution sequences

Research Topic. Error Concealment Techniques in H.264/AVC for Wireless Video Transmission in Mobile Networks

Project Proposal: Sub pixel motion estimation for side information generation in Wyner- Ziv decoder.

DELTA MODULATION AND DPCM CODING OF COLOR SIGNALS

WYNER-ZIV VIDEO CODING WITH LOW ENCODER COMPLEXITY

Spatial Error Concealment Technique for Losslessly Compressed Images Using Data Hiding in Error-Prone Channels

Estimating the impact of single and multiple freezes on video quality

An Evaluation of Video Quality Assessment Metrics for Passive Gaming Video Streaming

Error Resilience for Compressed Sensing with Multiple-Channel Transmission

Free Viewpoint Switching in Multi-view Video Streaming Using. Wyner-Ziv Video Coding

Robust Joint Source-Channel Coding for Image Transmission Over Wireless Channels

The Development of a Synthetic Colour Test Image for Subjective and Objective Quality Assessment of Digital Codecs

Research Article. ISSN (Print) *Corresponding author Shireen Fathima

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur

Using enhancement data to deinterlace 1080i HDTV

CHAPTER 8 CONCLUSION AND FUTURE SCOPE

Region Adaptive Unsharp Masking based DCT Interpolation for Efficient Video Intra Frame Up-sampling

ENCODING OF PREDICTIVE ERROR FRAMES IN RATE SCALABLE VIDEO CODECS USING WAVELET SHRINKAGE. Eduardo Asbun, Paul Salama, and Edward J.

EMBEDDED ZEROTREE WAVELET CODING WITH JOINT HUFFMAN AND ARITHMETIC CODING

Methodology for Objective Evaluation of Video Broadcasting Quality using a Video Camera at the User s Home

Analysis of Packet Loss for Compressed Video: Does Burst-Length Matter?

1 Evolution of measurement techniques from analogue to compressed digital

DATA hiding technologies have been widely studied in

ERROR CONCEALMENT TECHNIQUES IN H.264 VIDEO TRANSMISSION OVER WIRELESS NETWORKS

1. INTRODUCTION. Index Terms Video Transcoding, Video Streaming, Frame skipping, Interpolation frame, Decoder, Encoder.

Joint Optimization of Source-Channel Video Coding Using the H.264/AVC encoder and FEC Codes. Digital Signal and Image Processing Lab

ARTEFACTS. Dr Amal Punchihewa Distinguished Lecturer of IEEE Broadcast Technology Society

ELEC 691X/498X Broadcast Signal Transmission Fall 2015

OBJECTIVE VIDEO QUALITY METRICS: A PERFORMANCE ANALYSIS

Lund, Sweden, 5 Mid Sweden University, Sundsvall, Sweden

SUBJECTIVE AND OBJECTIVE EVALUATION OF HDR VIDEO COMPRESSION

Reduced complexity MPEG2 video post-processing for HD display

Measuring and Interpreting Picture Quality in MPEG Compressed Video Content

CHROMA CODING IN DISTRIBUTED VIDEO CODING

INTRA-FRAME WAVELET VIDEO CODING

University of Bristol - Explore Bristol Research. Peer reviewed version. Link to published version (if available): /ISCAS.2005.

Steganographic Technique for Hiding Secret Audio in an Image

Adaptive Key Frame Selection for Efficient Video Coding

An Overview of Video Coding Algorithms

Study of White Gaussian Noise with Varying Signal to Noise Ratio in Speech Signal using Wavelet

Comparative Analysis of Wavelet Transform and Wavelet Packet Transform for Image Compression at Decomposition Level 2

PAPER Parameter Embedding in Motion-JPEG2000 through ROI for Variable-Coefficient Invertible Deinterlacing

DICOM medical image watermarking of ECG signals using EZW algorithm. A. Kannammal* and S. Subha Rani

White Paper. Uniform Luminance Technology. What s inside? What is non-uniformity and noise in LCDs? Why is it a problem? How is it solved?

Video Over Mobile Networks

DWT Based-Video Compression Using (4SS) Matching Algorithm

Video Transmission. Thomas Wiegand: Digital Image Communication Video Transmission 1. Transmission of Hybrid Coded Video. Channel Encoder.

MULTI-STATE VIDEO CODING WITH SIDE INFORMATION. Sila Ekmekci Flierl, Thomas Sikora

INTERNATIONAL JOURNAL OF ELECTRONICS AND COMMUNICATION ENGINEERING & TECHNOLOGY (IJECET)

Reduced-reference image quality assessment using energy change in reorganized DCT domain

ROBUST ADAPTIVE INTRA REFRESH FOR MULTIVIEW VIDEO

Digital Video Telemetry System

INTERNATIONAL TELECOMMUNICATION UNION. SERIES H: AUDIOVISUAL AND MULTIMEDIA SYSTEMS Coding of moving video

Motion Video Compression

RECOMMENDATION ITU-R BT Methodology for the subjective assessment of video quality in multimedia applications

Video coding standards

Unequal Error Protection Codes for Wavelet Image Transmission over W-CDMA, AWGN and Rayleigh Fading Channels

A look at the MPEG video coding standard for variable bit rate video transmission 1

Real Time PQoS Enhancement of IP Multimedia Services Over Fading and Noisy DVB-T Channel

2-Dimensional Image Compression using DCT and DWT Techniques

COMPRESSION OF DICOM IMAGES BASED ON WAVELETS AND SPIHT FOR TELEMEDICINE APPLICATIONS

Objective Video Quality Assessment of Direct Recording and Datavideo HDR-40 Recording System

A Perceptual Distortion Metric for Digital Color Video

MULTIVIEW DISTRIBUTED VIDEO CODING WITH ENCODER DRIVEN FUSION

52 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 7, NO. 1, FEBRUARY 2005

Keep your broadcast clear.

Watermarking Robustness Evaluation Using Enhanced Performance Metrics.

Perceptual Coding: Hype or Hope?

1 Introduction to PSQM

SUBJECTIVE ASSESSMENT OF H.264/AVC VIDEO SEQUENCES TRANSMITTED OVER A NOISY CHANNEL

Fast MBAFF/PAFF Motion Estimation and Mode Decision Scheme for H.264

COMP 249 Advanced Distributed Systems Multimedia Networking. Video Compression Standards

FAST SPATIAL AND TEMPORAL CORRELATION-BASED REFERENCE PICTURE SELECTION

ON THE USE OF REFERENCE MONITORS IN SUBJECTIVE TESTING FOR HDTV. Christian Keimel and Klaus Diepold

Image Resolution and Contrast Enhancement of Satellite Geographical Images with Removal of Noise using Wavelet Transforms

Transcription:

PERCEPTUAL QUALITY ASSESSMENT FOR VIDEO WATERMARKING Stefan Winkler, Elisa Drelie Gelasca, Touradj Ebrahimi Genista Corporation EPFL PSE Genimedia 15 Lausanne, Switzerland http://www.genista.com/ swinkler@genimedia.com ABSTRACT The reliable evaluation of the performance of watermarking algorithms is difficult. An important aspect in this process is the assessment of the visibility of the watermark. In this paper, we address this issue and propose a methodology for evaluating the visual quality of watermarked video. Using a software tool that measures different types of perceptual video artifacts, we determine the most relevant impairments and design the corresponding objective metrics. We demonstrate their performance through subjective experiments on several different watermarking algorithms and video sequences. 1. INTRODUCTION The rapid spread of digital media (audio, images and video) and the ease of their reproduction and distribution has created a need for copyright enforcement schemes in order to protect content creators and owners. In recent years, digital watermarking has emerged as an effective way to prevent users from violating copyrights. This concept is based on the insertion of information into the data in such a way that the added information is not visible yet resistant to (intentional or unintentional) alterations of the watermarked data. In the watermarking algorithms, three factors must be considered: Capacity, i.e. the amount of information that can be put into the watermark and recovered without errors; Robustness, i.e. the resistance of the watermark to alterations of the original content such as compression, filtering or cropping; Visibility, i.e. how easily the watermark can be discerned by the user. These factors are inter-dependent; for example, increasing the capacity will decrease the robustness or increase the visibility. Therefore, it is essential to consider all three factors for a fair evaluation or comparison of watermarking algorithms. Organizations such as Certimark or the Content E. Drelie Gelasca and T. Ebrahimi are with the Signal Processing Lab at EPFL, Switzerland. http://www.certimark.org ID forum are working on the definition of procedures for such evaluations. While benchmark tests have already been proposed for the robustness of watermarking algorithms, such as CheckMark [9] or StirMark [], much less attention has been directed at the evaluation of the visual effects of the watermarking process. In this paper, we propose methods for the objective evaluation of watermarked video quality.. PERCEPTUAL QUALITY ASSESSMENT.1. Background Unfortunately, the accurate measurement of quality as perceived by the typical user is a big challenge in image or video processing in general. The reason for this is that the amount and visibility of distortions such as those introduced by watermarking strongly depend on the actual image/video content. The benchmark for any kind of visual quality assessment are subjective experiments, where a number of people are asked to watch test clips and to rate their quality. Procedures for such experiments have been formalized in ITU-R Recommendation BT.5 [] or ITU-T Recommendation P.9 [], which suggest standard viewing conditions, criteria for the selection of observers and test material, assessment procedures, and data analysis methods. However, subjective experiments require careful setup and are timeconsuming, hence expensive and often impractical. Furthermore, for many applications such as online quality monitoring and control subjective experiments cannot be used at all. Given these limitations, engineers have turned to simple error measures such as mean squared error (MSE) or peak signal-to-noise ratio (PSNR), suggesting that they would be as valid as subjective experiments. However, these simple measures operate solely on the basis of pixel-wise differences and neglect the important influence of video content and viewing conditions on the actual visibility of artifacts. Therefore, they cannot be expected to be reliable predictors of perceived quality. The shortcomings of these methods have led to the intensive study of more advanced perceptual quality metrics in recent years. An up-to-date review of such metrics can http://www.cidf.org

be found in []. Essentially two different approaches can be distinguished: Approaches based on models of the human visual system. These are the most general and potentially most accurate ones [1]. Examples of such metrics are described in [, 1, 17] among others. However, the human visual system is extremely complex, and many of its properties are not well understood even today. Besides, implementing these models is very expensive from a computational point of view due to their complexity. Approaches based on a priori knowledge about the compression methods as well as the pertinent types of artifacts. Examples of such specialized metrics include [11, 15]. While such metrics are not as versatile, they normally perform well in a given application area. Their main advantage lies in the fact that they often permit a computationally more efficient implementation. In this paper, we take the latter approach... Metrics We have chosen the VideoQoS software for the evaluation of the visual impact of the watermarks. VideoQoS is an application that provides for the measurement of artifacts affecting the perceptual quality of digital video. It does this through full reference quality metrics, i.e. it compares a reference video with a processed one to measure the quality of the degraded video in relation to the reference. Video- QoS provides separate metrics for different types of visual artifacts, which are divided into three categories: ANSI Metrics, which rely on algorithms defined by ANSI [1]. This document represents an attempt by a standards body to define objective measures that serve as a basis for the measurement of video quality. Perceptual Metrics, which measure specific visual artifacts introduced into the video in a way that is correlated with human perception. These artifacts are intuitive and well known, and are easily recognized even by inexperienced viewers. Fidelity Metrics, which are rather standard and represent mathematical fidelity measures of the video, e.g. MSE/PSNR. They do not take into account human perception. From our experience with the numerous video watermarking algorithms that we tested, we have seen mainly two kinds of impairments: Flicker, which result from visible changes of the watermark pattern between consecutive frames; High-frequency (HF) noise, which is the fundamental footprint of most watermarks. Based on these observations, we have designed objective metrics that measure the perceptual impact of these two impairments, which we refer to as Flicker metric and Noise metric in the following. VideoQoS version 1..1, July 1, Genista Corp. 3. EXPERIMENTS 3.1. Watermarking Algorithms Our interest here is mainly video watermarking. Most video watermarking techniques today are derived from algorithms for still images. Therefore, we adopt a number of watermarking schemes for still images and apply them to each frame of a video sequence. We chose four algorithms from the literature (Cox, Dugad, Xia and Wang). To enhance the test set, we also included a genuine video watermarking algorithm for videos developed by AlpVision. A brief description of each of these algorithms is given in the following. The scheme of Cox et al. [] is based on the discrete cosine transform (DCT). In practice, in order to place a length n watermark into an image, first the DCT of the entire image is computed, and a sequence of n real numbers is generated from a uniform distribution of zero mean and unit variance, which is then placed into the n highest magnitude coefficients of the transform matrix. Additionally, a scaling parameter α is specified to determine the extent to which the watermark alters the image. Dugad et al. [3] use a three level discrete wavelet transform (DWT) with an eight-tap Daubechies filter. The coefficients above a given threshold in all sub-bands except the low-pass band are picked. As in Cox s scheme, the watermark is generated by a sequence of n real numbers and is added to these coefficients. Thus, the scaling parameter α has the same meaning. Similarly, Xia et al. [19] decompose an image into several bands. The watermark is added to the largest coefficients in the high- and middle-frequency bands of the DWT. A parameter α is tuned to control the level of the watermark. The output of the inverse DWT is modified such that the resulting image has the same dynamic range as the original. Wang et al. [1] adopt a successive subband quantization scheme in the multi-threshold wavelet codec to choose perceptually significant coefficients for watermark embedding. The coefficients above the threshold in the current subband are the significant coefficients. The watermark is cast in the significant coefficients taking into account the scaling factors α and β. The value of α is adjustable by users to increase the watermarked image fidelity and decrease the security of watermark protection at the same time. The video watermarking scheme developed by AlpVision is based on technique initially proposed for still images by Kutter [7]. It uses spread-spectrum modulation to insert a watermark with variable amplitude and density in the spatial domain. In contrast to the other four algorithms, it considers the temporal content changes in the video. The default settings of each algorithm were used for all of its parameters. The source code for these algorithms can be downloaded from http://www.cosy.sbg.ac.at/ pmeerw/watermarking/source/ http://www.alpvision.com

(a) Train (b) Harp (c) Graphics (d) Race Figure 1: Frames from test clips. 3.. Test Clips We have watermarked four different test clips for our analysis. These clips were selected from the set of scenes in the VQEG test [13] to include spatial detail, saturated colors, motion, and synthetic content. They are seconds long with a frame rate of 5 Hz. They were de-interlaced and subsampled from the interlaced ITU-R Rec. BT.1 format [5] to a resolution of 3 pixels per frame for progressive display. The implementations of four watermarking algorithms are limited to frame sizes of powers of, therefore we have cropped a 5 5 pixel region from each frame in the video for watermarking and subsequent quality evaluation. A sample frame from each of the four scenes is shown in Figure 1. 3.3. Subjective Experiments For the evaluation of our metrics, subjective experiments were performed. Non-expert observers were asked to rank a total of watermarked test clips from best to worst according to perceived flicker and noise in two separate trials. The viewing order of the clips was not fixed; oberservs could freely choose between clips and play them as often as they liked. They could also watch the original clips for comparison. Six observers participated in the flicker trial, and five in the noise trial. The data obtained from the subjective ratings were combined to an average rank for comparison with the objective metrics. According to the subjective experiments, the most annoying artifacts in video are produced by watermarking algorithms that add noise patterns with relatively low spatial frequencies, which change from frame to frame and thus create clearly visible flicker. Other algorithms that add mainly high-frequency noise or temporally unchanging patterns to the video exhibit hardly any flicker at all. 3.. Results A statistical analysis of the data was carried out to evaluate the two proposed metrics with respect to the subjective ratings. Two correlation coefficients are used here to quantify and compare the metrics performance, namely the (linear) Pearson correlation coefficient as well as the (nonparametric) Spearman rank-order correlation coefficient. The scatter plot of perceived versus measured flicker for the above-mentioned watermarking algorithms and test clips is shown in Figure (a). For comparison, the scatter plot of perceived flicker versus PSNR is shown in Figure (b). The respective correlation coefficients are reported in the table in Figure (c). Figure 3 shows the same data for perceived noise and the HF-Noise metric. The proposed metrics clearly outperform PSNR in both cases. The plots show that adding a temporal component such as flicker to the measurements is essential for the evaluation of video watermarks, because PSNR is unable by design to take this into account. More surprisingly perhaps, PSNR is not well correlated with perceived noise either. This shows the importance of more discriminatory metrics for the perceptual quality evaluation of watermarks. For further improvement and testing of our metrics, more genuine video watermarking algorithms should be used. Also, an extension of the set of test clips could give better indications of how well these metrics generalize to different types of content.. CONCLUSIONS We have discussed the importance of perceptual quality assessment in watermarking. While this remains a difficult problem, we have presented a possible solution. We have introduced a measurement tool that analyzes video impairments by looking for certain types of artifacts. Using this tool, we have demonstrated that typical video watermarks suffer mostly from added high-frequency noise and/or flicker. The watermarking artifacts, which may be hardly noticeable in still images, become emphasized through the motion effects in video. We have proposed a Flicker metric and an HF-Noise metric to measure the perceptual impact of these specific distortions. Through subjective experiments we have demonstrated that the proposed metrics are reliable predictors of perceived flicker and perceived noise and clearly outperform PSNR in terms of prediction accuracy.

Perceived flicker 1 1 1.... 1 Flicker metric (a) Subjective flicker ratings vs. Flicker metric. Perceived noise 1 1 1.... 1 Noise metric (a) Subjective noise ratings vs. HF-Noise metric. Perceived flicker 1 1 1 5 3 35 5 5 55 PSNR [db] (b) Subjective flicker ratings vs. PSNR. Perceived noise 1 1 1 5 3 35 5 5 55 PSNR [db] (b) Subjective noise ratings vs. PSNR. Flicker Metric PSNR Pearson.95.5 Spearman.9.5 (c) Correlations. Figure : Perceived flicker vs. Flicker metric and PSNR (subjective data are shown with 9%-confidence intervals). Noise Metric PSNR Pearson.1. Spearman.1.1 (c) Correlations. Figure 3: Perceived noise vs. HF-Noise metric and PSNR (subjective data are shown with 9%-confidence intervals).

5. ACKNOWLEDGMENTS The authors wish to thank Frederic Jordan and Martin Kutter of AlpVision, Switzerland, for watermarking the test clips with their algorithm.. REFERENCES [1] ANSI T1.1.3: Digital transport of one-way video signals parameters for objective performance assessment. American National Standards Institute, New York, NY, 199. [] I. J. Cox et al.: Secure spread spectrum watermarking for multimedia. in Proc. ICIP, vol., pp. 173 7, Santa Barbara, CA, USA, 1997. [3] R. Dugad, K. Ratakonda, N. Ahuja: A new waveletbased scheme for watermarking for multimedia. in Proc. ICIP, Chicago, IL, USA, 199. [] ITU-R Recommendation BT.5-: Methodology for the subjective assessment of the quality of television pictures. International Telecommuncation Union, Geneva, Switzerland,. [5] ITU-R Recommendation BT.1-5: Studio encoding parameters of digital television for standard :3 and wide-screen 1:9 aspect ratios. International Telecommuncation Union, Geneva, Switzerland, 1995. [] ITU-T Recommendation P.9: Subjective video quality assessment methods for multimedia applications. International Telecommuncation Union, Geneva, Switzerland, 199. [7] M. Kutter: Digital Image Watermarking: Hiding Information in Images. Ph.D. thesis, École Polytechnique Fédérale de Lausanne, Switzerland, 1999. [] J. Lubin, D. Fibush: Sarnoff JND vision model. T1A1.5 Working Group Document #97-1, ANSI T1 Standards Committee, 1997. [9] S. Pereira et al.: Second generation benchmarking and application oriented evaluation. in Proc. Information Hiding Workshop, Pittsburgh, PA, 1. [] F. A. P. Petitcolas, R. J. Anderson: Evaluation of copyright marking systems. in Proc. IEEE Multimedia Systems (ICMCS), Florence, Italy, 1999. [11] K. T. Tan, M. Ghanbari, D. E. Pearson: An objective measurement tool for MPEG video quality. Signal Processing 7(3):79 9, 199. [1] C. J. van den Branden Lambrecht: Perceptual Models and Architectures for Video Coding Applications. Ph.D. thesis, École Polytechnique Fédérale de Lausanne, Switzerland, 199. [13] VQEG: Final report from the Video Quality Experts Group on the validation of objective models of video quality assessment., available at http://www.vqeg.org/. [1] H.-J. Wang, P.-C. Su, C.-C. J. Kuo: Wavelet-based digital image watermarking. Opt. Express 3(1):91 9, 199. [15] A. B. Watson et al.: Design and performance of a digital video quality metric. in Proc. SPIE, vol. 3, pp. 17, San Jose, CA, 1999. [1] S. Winkler: Issues in vision modeling for perceptual video quality assessment. Signal Processing 7():31 5, 1999. [17] S. Winkler: A perceptual distortion metric for digital color video. in Proc. SPIE, vol. 3, pp. 175, San Jose, CA, 1999. [] S. Winkler: Vision Models and Quality Metrics for Image Processing Applications. Ph.D. thesis, École Polytechnique Fédérale de Lausanne, Switzerland,. [19] X.-G. Xia, C. G. Boncelet, G. R. Arce: Wavelet transform based watermark for digital images. Opt. Express 3(1):97 51, 199.