Distributed Video Coding Using LDPC Codes for Wireless Video

Similar documents
WYNER-ZIV VIDEO CODING WITH LOW ENCODER COMPLEXITY

CHROMA CODING IN DISTRIBUTED VIDEO CODING

A Study of Encoding and Decoding Techniques for Syndrome-Based Video Coding

Wyner-Ziv Coding of Motion Video

Energy Efficient Video Compression for Wireless Sensor Networks *

Free Viewpoint Switching in Multi-view Video Streaming Using. Wyner-Ziv Video Coding

Systematic Lossy Forward Error Protection for Error-Resilient Digital Video Broadcasting

COMP 249 Advanced Distributed Systems Multimedia Networking. Video Compression Standards

Robust wireless video multicast based on a distributed source coding approach $

Video coding standards

LAYERED WYNER-ZIV VIDEO CODING FOR NOISY CHANNELS. A Thesis QIAN XU

MULTI-STATE VIDEO CODING WITH SIDE INFORMATION. Sila Ekmekci Flierl, Thomas Sikora

Modeling and Optimization of a Systematic Lossy Error Protection System based on H.264/AVC Redundant Slices

Systematic Lossy Error Protection of Video based on H.264/AVC Redundant Slices

MULTIVIEW DISTRIBUTED VIDEO CODING WITH ENCODER DRIVEN FUSION

UC San Diego UC San Diego Previously Published Works

Distributed Video Coding

Decoder-driven mode decision in a block-based distributed video codec

An Efficient Low Bit-Rate Video-Coding Algorithm Focusing on Moving Regions

A Novel Approach towards Video Compression for Mobile Internet using Transform Domain Technique

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur

Video compression principles. Color Space Conversion. Sub-sampling of Chrominance Information. Video: moving pictures and the terms frame and

Robust Transmission of H.264/AVC Video using 64-QAM and unequal error protection

Research Topic. Error Concealment Techniques in H.264/AVC for Wireless Video Transmission in Mobile Networks

Systematic Lossy Error Protection of Video Signals Shantanu Rane, Member, IEEE, Pierpaolo Baccichet, Member, IEEE, and Bernd Girod, Fellow, IEEE

Perspectives in distributed source coding

Systematic Lossy Error Protection based on H.264/AVC Redundant Slices and Flexible Macroblock Ordering

Joint Optimization of Source-Channel Video Coding Using the H.264/AVC encoder and FEC Codes. Digital Signal and Image Processing Lab

Robust Transmission of H.264/AVC Video Using 64-QAM and Unequal Error Protection

Marie Ramon, François-XavierCoudoux, andmarcgazalet. 1. Introduction

INFORMATION THEORY INSPIRED VIDEO CODING METHODS : TRUTH IS SOMETIMES BETTER THAN FICTION

Project Proposal: Sub pixel motion estimation for side information generation in Wyner- Ziv decoder.

The H.26L Video Coding Project

CERIAS Tech Report Preprocessing and Postprocessing Techniques for Encoding Predictive Error Frames in Rate Scalable Video Codecs by E

Analysis of Video Transmission over Lossy Channels

Exploring the Distributed Video Coding in a Quality Assessment Context

Video Quality Monitoring for Mobile Multicast Peers Using Distributed Source Coding

ELEC 691X/498X Broadcast Signal Transmission Fall 2015

Chapter 10 Basic Video Compression Techniques

Real-Time Distributed Video Coding for 1K-pixel Visual Sensor Networks

The H.263+ Video Coding Standard: Complexity and Performance

Overview: Video Coding Standards

Wyner-Ziv video coding for wireless lightweight multimedia applications

INTRA-FRAME WAVELET VIDEO CODING

Chapter 2 Introduction to

Multimedia Communications. Video compression

ROBUST ADAPTIVE INTRA REFRESH FOR MULTIVIEW VIDEO

Comparative Study of JPEG2000 and H.264/AVC FRExt I Frame Coding on High-Definition Video Sequences

Error Concealment for SNR Scalable Video Coding

Fast MBAFF/PAFF Motion Estimation and Mode Decision Scheme for H.264

AUDIOVISUAL COMMUNICATION

Example: compressing black and white images 2 Say we are trying to compress an image of black and white pixels: CSC310 Information Theory.

Multimedia Communications. Image and Video compression

PACKET-SWITCHED networks have become ubiquitous

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 58, NO. 5, MAY Note that the term distributed coding in this paper is always employed

Dual frame motion compensation for a rate switching network

COMP 9519: Tutorial 1

Rate-Adaptive Codes for Distributed Source Coding

Error Resilience for Compressed Sensing with Multiple-Channel Transmission

University of Bristol - Explore Bristol Research. Peer reviewed version. Link to published version (if available): /ISCAS.2005.

SYSTEMATIC LOSSY ERROR PROTECTION OF VIDEO SIGNALS

Video Transmission. Thomas Wiegand: Digital Image Communication Video Transmission 1. Transmission of Hybrid Coded Video. Channel Encoder.

Introduction to Video Compression Techniques. Slides courtesy of Tay Vaughan Making Multimedia Work

176 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 13, NO. 2, FEBRUARY 2003

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video

MPEG-2. ISO/IEC (or ITU-T H.262)

Intra-frame JPEG-2000 vs. Inter-frame Compression Comparison: The benefits and trade-offs for very high quality, high resolution sequences

Part1 박찬솔. Audio overview Video overview Video encoding 2/47

1934 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 4, APRIL 2012

Performance Evaluation of Error Resilience Techniques in H.264/AVC Standard

High performance and low complexity decoding light-weight video coding with motion estimation and mode decision at decoder

Contents. xv xxi xxiii xxiv. 1 Introduction 1 References 4

Investigation of the Effectiveness of Turbo Code in Wireless System over Rician Channel

Understanding IP Video for

H.261: A Standard for VideoConferencing Applications. Nimrod Peleg Update: Nov. 2003

Research Article. ISSN (Print) *Corresponding author Shireen Fathima

Digital Video Telemetry System

Analysis of Packet Loss for Compressed Video: Does Burst-Length Matter?

Video 1 Video October 16, 2001

Compressed-Sensing-Enabled Video Streaming for Wireless Multimedia Sensor Networks Abstract:

Video Over Mobile Networks

Interleaved Source Coding (ISC) for Predictive Video Coded Frames over the Internet

WITH the demand of higher video quality, lower bit

ROBUST IMAGE AND VIDEO CODING WITH ADAPTIVE RATE CONTROL

Reduced Decoder Complexity and Latency in Pixel-Domain Wyner-Ziv Video Coders

Minimax Disappointment Video Broadcasting

An Overview of Video Coding Algorithms

DELTA MODULATION AND DPCM CODING OF COLOR SIGNALS

Region Adaptive Unsharp Masking based DCT Interpolation for Efficient Video Intra Frame Up-sampling

yintroduction to video compression ytypes of frames ysome video compression standards yinvolves sending:

DWT Based-Video Compression Using (4SS) Matching Algorithm

Advanced Video Processing for Future Multimedia Communication Systems

Lecture 1: Introduction & Image and Video Coding Techniques (I)

UNBALANCED QUANTIZED MULTI-STATE VIDEO CODING

A video signal consists of a time sequence of images. Typical frame rates are 24, 25, 30, 50 and 60 images per seconds.

Principles of Video Compression

Lecture 2 Video Formation and Representation

Agrowing percentage of the world population now uses image and

THE CAPABILITY of real-time transmission of video over

International Journal for Research in Applied Science & Engineering Technology (IJRASET) Motion Compensation Techniques Adopted In HEVC

Transcription:

Wireless Sensor Network, 2009, 1, 334-339 doi:10.4236/wsn.2009.14041 Published Online November 2009 (http://www.scirp.org/journal/wsn). Distributed Video Coding Using LDPC Codes for Wireless Video Abstract P. APARNA, Sivaprakash REDD, Sumam DAVID * Department of Electronics and Communication Engineering, National Institute of Technology Karnataka, Surathkal, India Email: sumam@ieee.org Received May 19, 2009; revised June 16, 2009; accepted June 23, 2009 Popular video coding standards like H.264 and MPEG working on the principle of motion-compensated predictive coding demand much of the computational resources at the encoder increasing its complexity. Such bulky encoders are not suitable for applications like wireless low power surveillance, multimedia sensor networks, wireless PC cameras, mobile camera phones etc. New video coding scheme based on the principle of distributed source coding is looked upon in this paper. This scheme supports a low complexity encoder, at the same time trying to achieve the rate distortion performance of conventional video codecs. Current implementation uses LDPC codes for syndrome coding. Keywords: Syndrome Coding, Cosets, Distributed Source Coding, Distributed Video Coding (DVC). 1. Introduction With the proliferation of various complex video applications it is necessary to have advanced video and image compression techniques. Popular video standards like ISO MPEG and ITU-H.26x have been successful in accomplishing the requirements in terms of compression efficiency and quality. However these standards are pertinent to downlink friendly applications like video telephony, video streaming, broadcasting etc. These conventional video codecs work on the principle of motion compensated prediction which increases the encoder complexity due to the coexistence of the decoder with the encoder. Also motion-search algorithm makes the encoder computationally intensive. The downlink friendly architectures belong to the class of Broadcast model, where in high encoder complexity is not an issue. The encoder of a Broadcast model resides at the base-station where power consumption and computational resources are not an issue. However this Broadcast model of video is not suitable for uplink friendly applications like mobile video cameras, wireless video sensor networks, wireless surveillance etc which demands a low power, low complexity encoder. These uplink friendly applications which belong to wireless-video model demands a simple encoder since the power and the computational resources are of primary concern in the wireless scenario. Based on the information theoretic bounds established in 1970 s by Slepian-Wolf [1] for distributed lossless *SMIEEE. coding and by Wyner-Ziv [2] for lossy coding with decoder side information, it is seen that efficient compression can also be achieved by exploiting source statistics partially or wholly at the decoder. Video compression schemes that build upon these theorems are referred as distributed video coding which befits uplink friendly video applications. Distributed video coding shifts the encoder complexity to the decoder making it suitable for wireless video model. Unlike conventional video codecs distributed coding exploits the source statistics at the decoder alone, thus interchanging the traditional balance of complex encoder and simple decoder. Hence the encoder of such a video codec is very simple, at the expense of a more complex decoder. Such algorithms hold great promise for new generation mobile video cameras and wireless sensor networks. In the design of a new video coding paradigm, issues like compression efficiency, robustness to packet losses, encoder complexity are of prime importance in comparison with conventional coding system. In this paper we present the simulation results of distributed video coding with syndrome coding as in PRISM [3], using LDPC codes for coset channel coding [4]. 2. Background 2.1. Slepian-Wolf Theorem for Lossless Distributed Coding [1] Consider two correlated information sequences and.

P. APARNA ET AL. 335 of each source is constrained to operate without the knowledge of the other source while the decoder has access to both encoded binary message streams as shown in Figure 1. The problem that Slepian-Wolf theorem addresses is to determine the minimum number of bits per source character required for encoding the message stream in order to ensure accurate reconstruction at the decoder. Considering separate encoder and the decoder for and, the rate required is R H() and R H() where H() and H() represents the entropy of and respectively. Slepian-Wolf [1] showed that good compression can be achieved with joint decoding but separate encoding. For doing this an admissible rate region is defined [6] as shown in Figure 2 given by: R + R H(,) (1) R H(/), R H() (2) R H(), R H(/) (3) Thus Slepian-Wolf [1] showed that Equation (1) is the necessary condition and Equation (2) or Equation (3) are the sufficient conditions required to encode the data in case of joint decoding. With the above result as the base, we can consider the distributed coding with side information at the decoder as shown in the Figure 3. Let be the source data that is statistically dependent to the side information. Side information is separately encoded at a rate R H() and is available only at the decoder. Thus as seen from Figure 2 can be encoded at a rate R H(/). Correlated Sources Joint Decoder Figure 1. Compression of correlated sources by separate encoder but decoded jointly. R y bits Rx Ry No Errors ˆ Ŷ Source / R x H(/) Lossless Lossless Decoder Side Information R H() Figure 3. Lossless decoder with side information. 2.1. Wyner-Ziv Rate Distortion Theory[2,6] Aaron Wyner and Jacob Ziv [2,6] extended Slepian- Wolf theorem and showed that conditional Rate-MSE distortion function for is same whether the side information is available only at the decoder or both at encoder and decoder; where and are statistically dependent Gaussian random processes. Let and be the samples of two random sequences representing the source data and side information respectively. encodes without access to side information as shown in Figure 4. Decoder reconstructs using as side information. Let D = E [d (, )] is the acceptable distortion. Let R / (D) be the rate required for the case where side information is available at the encoder also and R WZ / represent the Wyner-Ziv rate required when encoder doesn t have access to side information. Wyner-Ziv proved that Wyner-Ziv rate WZ distortion function R / (D) is the achievable lower bound for the bitrate for a distortion D WZ R / R / 0 (4) They also showed that for Gaussian memoryless sources WZ R / R / 0 (5) As a result source sequence can be considered as the sum of arbitrarily distributed side information and independent Gaussian Noise. Distributed video coding is based on these two fundamental theories, specifically works on the Wyner-Ziv coding considering a distortion measure. In such a coding system the encoder encodes each video frame separately H(y) H(y/x) Vanishing Errors for long sequences R wz (D) H(/) x Source / Lossy Lossy Decoder ' H(x,y)=R x +R y H(x/y) H(x) R x bits Side Information R H() Figure 2. Admissible rate region [5]. Figure 4. Lossy decoder with side information.

336 P. APARNA ET AL. with respect to the correlation statistics between itself and the side information. The decoder decodes the frames jointly using the side information available only at the decoder. This video paradigm is as opposed to the conventional coding system where the side information is available both at the encoder and decoder as shown in Figure 5. 2.2. Syndrome Coding [5] Let be a source that is to be transmitted using least average number of bits. Statistically dependent side information, such that = + N is available only at the decoder. The encoder must therefore encode in the absence of, whereas the decoder jointly decodes using. Distributed source encoder compresses in to syndromes S with respect to a Channel code C [7]. Decoder on receiving the syndrome can identify the coset to which belongs and using side information can reconstruct back. 2.3. Correlation Channel and the Channel Codes [4] The performance of the channel codes is the key factor of the distributed video coding system in both error correcting and data compression. Turbo and LDPC codes are two advanced channel codes which have astonishing performance near the Shannon Capacity limit. The use of LDPC codes for syndrome coding was first suggested by Liveris in [4], where the message passing algorithm was modified to take syndrome information in to account. Side Information Decoder Figure 5. Lossless decoder with side information. ˆ The correlation between binary sources = [ 1, 2..., n ] and = [ 1, 2,..., n ] is modeled using a binary symmetric channel. We consider i and i to be correlated according to Pr [ i i ] = p < 0.5. The rate used for is its entropy R = H(), therefore the theoretical limit for lossless compression of is given by nr x nh( i / i ) = nh(p) =n( plog 2 p (1 p)log 2 (1 p)) (6) The compressed version of is the syndrome S which is the input to the channel. The source is assumed to be available at the decoder as side information. Using a linear (n,k) binary block code, it is possible to have 2 n k distinct syndromes, each indexing a set of 2 k binary words of length n. This compression results in mapping a sequence of n input symbols into (n k) syndrome symbols. 3. Implementation 3.1. The encoder block diagram is shown in the Figure 6. The video frames are divided into blocks of 8x8 and each block is processed one by one. Block DCT (Discrete Cosine Transform) is applied to each 8x8 block (or 16x16) and the DCT coefficients are zig-zag scanned so that they are arranged as an array of coefficients in order of their importance. Then the transformed coefficients are uniform quantized with reference to target distortion measure and desired reconstruction quality. After quantization a bitplane is formed for each block as shown in Figure 7 [3]. Main idea behind distributed video coding is to code source assuming that the side information is available at the decoder such that = + N, where N is Gaussian random noise. This is done in the classification step where bitplane for each coefficient is divided into different levels of importance. Classification step strongly rely on the correlation noise Input Block DCT & Zigzag Scan Uniform Quantization One Frame Delay Bit Plane Formation Correlation Noise Estimation Classification LSB MSB Syndrome Coding using LDPC codes Entropy Coding using Adaptive Huffman Coding Encode d Bit Stream Class Info Figure 6. Video encoder.

P. APARNA ET AL. 337 Bitplanes b 0 b 1 b 2 b m Coefficients x 0 x 1 x 2 x 3 x 63 structure N between the source block and the side information block. Less is the correlation noise between and, more is the similarity and hence less number of bits of can be transmitted to the decoder. In order to classify the bitplanes offline training is done for different types of video files without any motion search. On the basis of offline process 16 types of classes are formed, where each class considers different number of bitplanes for entropy coding and syndrome coding for each coefficient in the block. In the classification process, MSE (mean square error) for each block is computed with respect to the zero motion blocks in the previous frame. Based on the MSE and the offline process appropriate class for that particular block is chosen. As a result some of the least significant bit planes are syndrome coded and some of the bitplanes that can be reconstructed from side information are totally ignored. The syndrome coding bitplanes shown in black and gray in Figure 7 and skip planes shown in white in Figure 7. Skip planes can be reconstructed back using side information at the decoder and hence need not be sent to the decoder. The important bits of each coefficient that cannot be determined by side information has to be syndrome coded [3]. In our implementation we code two bitplanes using coset channel coding and the remaining syndrome bitplanes using Adaptive Huffman coding. Among the syndrome coding bitplanes we code the most significant bit planes using Adaptive Huffman coding. The number of bitplanes to be syndrome coded is directly used from class information that is hard coded. Hence we need not send four-tuple data (run, depth, path, last) as in PRISM [3]. Rest of the least significant bitplanes is coded using coset channel coding. This is done by using a parity check matrix H of a (n,k) linear channel code. Compression is achieved by generating syndrome bits of length (n-k) for each n bits of data. These syndrome bits are obtained by multiplying the source bits with the parity check matrix H such that Figure 7. Bit planes for each coefficient blocks. S = Hb where S represents the syndrome bits. H represents the parity check matrix of linear channel code. b represents the source bits. These syndromes identify the coset to which the source data belongs to. In this implementation we have considered two biplanes for coset coding marked gray in the Figure 7. We have implemented this using irregular 3/4 rate LDPC coder [4]. 3.2. Decoder The Decoder block diagram is shown in the Figure 8. The entropy coded bits are decoded by an entropy decoder and the coset coded bits are passed to the LDPC decoder. In this implementation, previous frame is considered as the side information required for syndrome decoding. Once the syndrome coded bits are recovered they identify the coset to which i belongs and hence using the side information i we can correctly decode the entire bits of i. The quantized codeword sequence is then dequantized and inverse transformed to get the original coefficients. 4. Simulation Results Video Codec is designed for a single camera scenario which is an application to wireless network of video camera equipped with cell phones. The video codec is simulated and tested with a object oriented approach Side Information Generation Syndrome Decoding using LDPC decoder LSB Encoded Bit Stream Entropy Decoding Bit Plane Formation MSB Uniform DeQuantization IDCT & Zigzag Scan Reconstructed Block Figure 8. Video decoder.

338 P. APARNA ET AL. Table 1. Filename: foreman. QCIF, frame rate=30fps. Luma PSNR (db) for different Methods BitRate (Mbps) DVC Implementation H.263+Predictive Coder IntraCoder (Motion JPEG) 2.57 31.357 34.72 30.092 2.67 33.554 35.03 32.863 3.55 35.534 35.86 34.92 Table 2. Filename: football. QCIF, frame rate=30fps. Luma PSNR (db) for different Methods BitRate (Mbps) DVC Implementation H.263+ Predictive Coder IntraCoder (Motion JPEG) 3.52 30.724 25.62 30.07 3.67 31.834 25.76 30.92 4.87 34.005 26.59 33.80 Figure 9. a) Error resilience characteristics of DVC, 4th, 10th, 20th frames are lost for football; b) Error resilience characteristics of DVC, 4th, 10th, 20 th frames are lost for foreman. using C++ in gcc. The program processes frames one by one and within each frame, block wise processing is done. The input to the encoder is a QCIF video file (Quarter Common Intermediate Format). allows the storage of one previous frame. Objective performance evaluation of the system is done by measuring the Compression Ratio (CR), MSE and the Peak Signal to Noise Ratio (PSNR) between the original and the reconstructed video. The PSNR and CR for various video sequences is computed. These are compared with that of H.263+ Intra and H.263+ Predictive video codec [8]. The encoder and decoder block as shown in Figure 6 and Figure 8 respectively are implemented and some preliminary simulation results are presented in this paper for two video files Football and Foreman in QCIF resolution with a frame rate of 30 fps. The rate distortion performance and the error resilience characteristics of the distributed video coder is presented in this paper. As seen from the Table 1, for the same bitrate distributed video coder has better PSNR than DCT based intraframe coder and but is slightly inferior to H.263+ predictive coder [8] for Foreman file. As seen from Table 2 distributed video coder has better PSNR than DCT based intraframe coder and H.263+ predictive coder for Football file. With some enhancements to the current coding scheme such as accurate modeling of correlation statistics between the source data and the side information, proper motion search module for side information generation etc, better rate-distortion performance can be achieved with a low complexity encoder model. Error Resilience characteristics of Distributed video scheme is as shown in Figure 9a for Football and Figure 9b for Foreman. Effect on the quality of the reconstructed video sequence is seen by dropping 4th, 10th, 20th frames at the decoder in our implementation. It is seen that distributed video coder recovers quickly. In Distributed video scheme, decoding is dependent on the side information that is universal for all source data as long as correlation structure is satisfied. 5. Conclusion In this paper we have tried PRISM [3] like implementation using LDPC coset channel coding. By proper modeling of correlation structure of source and the side information for video we can achieve better compression performance with better quality of reconstructed video sequence. However the main aim of distributed video coding scheme is to reduce encoder complexity to conform with wireless-video model, which seems to be satisfied. Distributed codec is more robust to packet /frame

P. APARNA ET AL. 339 loss due to the absence of prediction loop in the encoder. In a Predictive coder accuracy of decoding is strongly dependent on a single predictor from the encoder, loss of which results in erroneous decoding and error propagation. Hence Predictive coder can recover from packet or frame loss by only some extent. The quality of the reconstructed signal for the same CR can be improved by performing more complex motion search. However it is seen that the current implementation operates well in high quality (PSNR of order of 30dB) regime. The extension to lower bit rates without any compromise in the quality so that it is comparable with the conventional codecs will be the next part of the work. 6. References [1] J. D. Slepian and J. K. Wolf, Noiseless coding of correlated information sources, IEEE Transactions on Information Theory, Vol. IT-19, pp. 471 480, July 1973. [2] A. D. Wyner and J. Ziv, The rate-distortion function for source coding with side information at the decoder, IEEE Transactions on Information Theory, Vol. IT-22, No. 1, pp. 1 10, January 1976. [3] R. Puri, A. Majumdar, and K. Ramachandran, PRISM: A video coding paradigm with motion estimation at the decoder, IEEE Transactions on Image Processing, Vol. 16, No. 10, October 2007. [4] A. D. Liveris, Compression of binary sources with side information at the decoder using LDPC codes, IEEE Communication Letters, Vol. 6, No. 10, October 2002. [5] S. S. Pradhan and K. Ramchandran, Distributed source coding using syndromes (DISCUS): Design and construction, Proc. IEEE Data Compression Conference, Snowbird, UT, pp. 158 167, March 1999. [6] A. D. Wyner, Recent results in the Shannon theory, IEEE Transactions on Information Theory, Vol. 20, No. 1, pp. 2 10, January 1974. [7] R. Puri and K. Ramchandran, PRISM: A new robust video coding architecture based on distributed compression principles, Proc. Allerton Conference on Communication, Control and Computing, Allerton, IL, October 2002. [8] G. Cote, B. Erol, M. Gallant, and F. Kosssentini, H.263+: Video coding at low bitrates, IEEE Transactions. Circuits Sys. Video Technology, Vol. 8, No. 7, pp. 849 866, November 1998. [9] B. Girod, A. M. Aaron, S. R. and D. Rebollo-Monedero, Distributed video coding, Proceedings of the IEEE, Vol. 93, No. 1, pp. 71 83, January 2005.