Analysis of a Two Step MPEG Video System

Similar documents
Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur

AUDIOVISUAL COMMUNICATION

Chapter 10 Basic Video Compression Techniques

An Efficient Low Bit-Rate Video-Coding Algorithm Focusing on Moving Regions

Contents. xv xxi xxiii xxiv. 1 Introduction 1 References 4

Video compression principles. Color Space Conversion. Sub-sampling of Chrominance Information. Video: moving pictures and the terms frame and

Video coding standards

CHROMA CODING IN DISTRIBUTED VIDEO CODING

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video

WYNER-ZIV VIDEO CODING WITH LOW ENCODER COMPLEXITY

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur

MPEG has been established as an international standard

Color Quantization of Compressed Video Sequences. Wan-Fung Cheung, and Yuk-Hee Chan, Member, IEEE 1 CSVT

Multimedia Communications. Video compression

SCENE CHANGE ADAPTATION FOR SCALABLE VIDEO CODING

Chapter 2 Introduction to

Selective Intra Prediction Mode Decision for H.264/AVC Encoders

An Overview of Video Coding Algorithms

SHOT DETECTION METHOD FOR LOW BIT-RATE VIDEO CODING

PAL uncompressed. 768x576 pixels per frame. 31 MB per second 1.85 GB per minute. x 3 bytes per pixel (24 bit colour) x 25 frames per second

FLEXIBLE SWITCHING AND EDITING OF MPEG-2 VIDEO BITSTREAMS

Motion Video Compression

Reduced complexity MPEG2 video post-processing for HD display

Multimedia Communications. Image and Video compression

Fast MBAFF/PAFF Motion Estimation and Mode Decision Scheme for H.264

Motion Re-estimation for MPEG-2 to MPEG-4 Simple Profile Transcoding. Abstract. I. Introduction

Impact of scan conversion methods on the performance of scalable. video coding. E. Dubois, N. Baaziz and M. Matta. INRS-Telecommunications

MPEG-2. ISO/IEC (or ITU-T H.262)

Implementation of an MPEG Codec on the Tilera TM 64 Processor

Bit Rate Control for Video Transmission Over Wireless Networks

Research Topic. Error Concealment Techniques in H.264/AVC for Wireless Video Transmission in Mobile Networks

University of Bristol - Explore Bristol Research. Peer reviewed version. Link to published version (if available): /ISCAS.2005.

ENCODING OF PREDICTIVE ERROR FRAMES IN RATE SCALABLE VIDEO CODECS USING WAVELET SHRINKAGE. Eduardo Asbun, Paul Salama, and Edward J.

A Statistical Framework to Enlarge the Potential of Digital TV Broadcasting

The H.263+ Video Coding Standard: Complexity and Performance

AN IMPROVED ERROR CONCEALMENT STRATEGY DRIVEN BY SCENE MOTION PROPERTIES FOR H.264/AVC DECODERS

MULTI-STATE VIDEO CODING WITH SIDE INFORMATION. Sila Ekmekci Flierl, Thomas Sikora

Visual Communication at Limited Colour Display Capability

CERIAS Tech Report Preprocessing and Postprocessing Techniques for Encoding Predictive Error Frames in Rate Scalable Video Codecs by E

Joint Optimization of Source-Channel Video Coding Using the H.264/AVC encoder and FEC Codes. Digital Signal and Image Processing Lab

Constant Bit Rate for Video Streaming Over Packet Switching Networks

Drift Compensation for Reduced Spatial Resolution Transcoding

MPEGTool: An X Window Based MPEG Encoder and Statistics Tool 1

Project Proposal: Sub pixel motion estimation for side information generation in Wyner- Ziv decoder.

Adaptive Key Frame Selection for Efficient Video Coding

Performance Evaluation of Error Resilience Techniques in H.264/AVC Standard

Express Letters. A Novel Four-Step Search Algorithm for Fast Block Motion Estimation

Overview: Video Coding Standards

Introduction to Video Compression Techniques. Slides courtesy of Tay Vaughan Making Multimedia Work

PERCEPTUAL QUALITY OF H.264/AVC DEBLOCKING FILTER

A look at the MPEG video coding standard for variable bit rate video transmission 1

RECOMMENDATION ITU-R BT.1203 *

Research Article. ISSN (Print) *Corresponding author Shireen Fathima

Minimax Disappointment Video Broadcasting

INTRA-FRAME WAVELET VIDEO CODING

Improvement of MPEG-2 Compression by Position-Dependent Encoding

complex than coding of interlaced data. This is a significant component of the reduced complexity of AVS coding.

COMP 249 Advanced Distributed Systems Multimedia Networking. Video Compression Standards

Video Over Mobile Networks

ROBUST ADAPTIVE INTRA REFRESH FOR MULTIVIEW VIDEO

Region Adaptive Unsharp Masking based DCT Interpolation for Efficient Video Intra Frame Up-sampling

UC San Diego UC San Diego Previously Published Works

06 Video. Multimedia Systems. Video Standards, Compression, Post Production

Implementation of MPEG-2 Trick Modes

Key Techniques of Bit Rate Reduction for H.264 Streams

MPEG-2. Lecture Special Topics in Signal Processing. Multimedia Communications: Coding, Systems, and Networking

yintroduction to video compression ytypes of frames ysome video compression standards yinvolves sending:

WITH the rapid development of high-fidelity video services

Improved Error Concealment Using Scene Information

Distributed Video Coding Using LDPC Codes for Wireless Video

Analysis of Packet Loss for Compressed Video: Does Burst-Length Matter?

Video Transmission. Thomas Wiegand: Digital Image Communication Video Transmission 1. Transmission of Hybrid Coded Video. Channel Encoder.

Robust 3-D Video System Based on Modified Prediction Coding and Adaptive Selection Mode Error Concealment Algorithm

The Development of a Synthetic Colour Test Image for Subjective and Objective Quality Assessment of Digital Codecs

The H.26L Video Coding Project

SCALABLE video coding (SVC) is currently being developed

ERROR CONCEALMENT TECHNIQUES IN H.264 VIDEO TRANSMISSION OVER WIRELESS NETWORKS

Comparative Study of JPEG2000 and H.264/AVC FRExt I Frame Coding on High-Definition Video Sequences

A Unified Approach to Restoration, Deinterlacing and Resolution Enhancement in Decoding MPEG-2 Video

Adaptive reference frame selection for generalized video signal coding. Carnegie Mellon University, Pittsburgh, PA 15213

Shot Transition Detection Scheme: Based on Correlation Tracking Check for MB-Based Video Sequences

Colour Reproduction Performance of JPEG and JPEG2000 Codecs

A Novel Approach towards Video Compression for Mobile Internet using Transform Domain Technique

Free Viewpoint Switching in Multi-view Video Streaming Using. Wyner-Ziv Video Coding

Digital Video Telemetry System

Wipe Scene Change Detection in Video Sequences

THE CAPABILITY of real-time transmission of video over

Workload Prediction and Dynamic Voltage Scaling for MPEG Decoding

Adaptive Distributed Compressed Video Sensing

VIDEO ANALYSIS IN MPEG COMPRESSED DOMAIN

1. Introduction. 2. System requirements. Abstract

Principles of Video Compression

ELEC 691X/498X Broadcast Signal Transmission Fall 2015

Bridging the Gap Between CBR and VBR for H264 Standard

Error Resilience for Compressed Sensing with Multiple-Channel Transmission

Error prevention and concealment for scalable video coding with dual-priority transmission q

Video 1 Video October 16, 2001

CM3106 Solutions. Do not turn this page over until instructed to do so by the Senior Invigilator.

Error concealment techniques in H.264 video transmission over wireless networks

Content storage architectures

Transcription:

Analysis of a Two Step MPEG Video System Lufs Telxeira (*) (+) (*) INESC- Largo Mompilhet 22, 4000 Porto Portugal (+) Universidade Cat61ica Portnguesa, Rua Dingo Botelho 1327, 4150 Porto, Portugal Abstract: The efficiency of a rate control mechanism depends on the knowledge about the video complexity. The high number of different occurrences that can exist in a given video sequence makes it very difficult to encode at a constant bit rate while maintaining constant the picture quality. This paper presents a 'feed forward bit rate control' to improve the perceptual quality of a video sequence,, The algorithm also produces indexing information that allows detecting and classifying scene cuts. Keywords: MPEG, scene change, scene complexity, bit rate control 1. Introduction The success of the MPEG family of standards is based on the fact that they were designed to be generic, Therefore, they can be used in a wider range of applications when compared with their predecessors created to meet the requirements of one single application. MPEG1 was designed for multimedia storage applications. MPEG2 standard is expected to cover a larger field of applications such as Digital Terrestrial Television Broadcasting (dtrb), Video-on Demand, pay TV, SNG, ENG. The MPEG rate control algorithm plays an important role for improving and stabilising the quality of the compressed video sequence. As MPEG does not specify how to control the bit rate several solutions have been presented in the literature [3,4,5,6]. There are two different approaches: 'feed forward bit rate control' and "feed backward bit rate controro In the last approach we have a limited knowledge of the sequence complexity and the encoder performance. Bits are allocated on a picture basis and spatially uniform distributed throughout the image. Thus too many bits may be overspent in the beginning of the picture while the end of the picture may present a higher degree of complexity. In the first approach, a pre-analysis is performed in order to determine the optimum setting, which will increase the computational complexity. The feed backward bit rate control is suitable for real time applications and feed forward bit rate control for applications where the quality is the main goal and time isn't a constraint. This paper describes a "feed forward bit rate control' that allows not only to improve the overall video sequence quality but also to extract additional information that will be useful for operations like indexing and browsing. 2. Bit Rate Control The bit rate control may be expressed as follows: given a particular video input and a desired bit rate what should be the encoder setting to maintain the picture quality as high and constant as possible. In MPEG encoding, the trade-off between picture quality and the bit rate is controlled by a quantization scale, denoted as q. This parameter is used to compute the step size of the uniform quantizers used for the different AC DCT coefficients [2]. Each macroblock can be quantised with a different q. The scheme for adjusting the value of q between macroblocks within an image frame is called 'adaptive quantization". There are several schemes for doing the

63 adaptive quantization. For example, in MPEG Test Model 5 (TM5) [3], a non-linear mapping based on the block variance is used to adapt the q's. Besides the quantization scale, q, quantization coarseness depends also on the quantization matrix. In MPEG1, the quantization matrix can be altered in each sequence while in MPEG2 on a picture basis. The actual step size is obtained by multiplying the parameter q by a matrix of weights, one for each DCT coefficient. There are lots of different characteristics in video sequences and the MPEG-2 encoder reacts differently to each of these characteristics. We have implemented a two step coding strategy. First, the video sequence is analysed (pre-coded) using fixed parameters. Statistics are obtained and the video quality is assessed, on a frame basis, in a high quality monitor (in case of assisted encoded). One of the goals is to obtain the Bit Usage Profile. The pre-analyser (first step) generates a histogram of the number of bits, the q signal, the PSNR for each macrohlock (both for a particular position and regardless of the spatial position). The Bit Usage Profile allows us to determine how, when and where the complexity varies. Thus when the second step coding is performed, the final result presents an improvement of the perceptual picture quality. 201 Analysis of Scene Complexity The first stage analyses the video sequence and extracts characteristics, such as local activity, location of frames that are particularly hard to code, etc., that will enable greater coding efficiency in the second stage. The quality with which a video sequence can be coded at a given data rate is largely dependent on its complexity. As the complexity of a video scene can vary widely we define measures of picture and scene complexity. All interframe processing, such as motion analysis, is also performed in the first stage. Picture Complexity The picture complexity is measured by computation of the average value of spatial local activity. Spatial local activity, for the macroblock j, is measured from the four luminance frame-organised sub-blocks and the four luminance field-organised subblocks using the original pixel values: actj = 1 + min ~,=~,, (vat_ sblk) (1) where v sbl - (2) F-iV" p - 64 X.~k=l (3) and Pk are the pixel values in the original 8 8 block.

64 Hard to code pictures We defined a hard to code picture as a picture for which the prediction falls, resulting in large prediction-error energy and a highly non-smooth vector field. A hard to code picture is detected when a mismatch takes place between the coding model (which assumes purely translation motion) and the content of the images (which may contain newly revealed or occluded regions, non-transitional motion or a scene change). To solve this problem, one possible solution would be to encode the hard to code picture with intra.coded blocks. As the overhead would still remain high, the hard to code picture is associated with the beginning of a new GOP, becoming an Intra frame. Thus there is an alignment between the Intra frames and hard to code pictures Scene background In most scenes part of the image does not change within a GOP time frame. The macroblocks belonging to flxed parts of the image, the background, need only to be encoded in the first image of the GOP (as intra). Usually, due to a coarse coding the same spatial macroblock is encoded again in P and B frames (as error information). The presence of noise may affect this decision. Whenever the difference between the sum of pixel values of maeroblocks in two consecutive frames, at the same spatial position, is inferior to a fixed threshold, the macroblock is classified as fixed background. Video Sequences Characteristics The characteristics studied where: scene cuts, noise characteristics, level of detail and movement. The encoder automatically detects scene cuts and divides the sequence into scenes. Each scene is classified according to the classes identified ha MPEG4: Class A B C Characteristics Low detail and low movement Low detail and medium movement or vice-versa medium detail and high movement or vice-versa Table 1) MPEG video classes Example Akiyo, Hall Monitor Coast Guard, Silent Bus, Calendar & Mobile After characteristic extraction, each scene goes through a first step encoding process where the input parameters are set according to its class. In images where scene cuts occur all the macroblocks are coded as Intra blocks to insure that the block effect won't be visible in that frame 2.2 Improved rate-control (second step) In an one-pass scheme, since no information about the video sequence is known in advance, the target number of bits per frame is often estimate in a linear fashion, i.e. allocating to each frame of the same type the same number of bits. This scheme doesn't take advantage of information about local variations of activity and can conduct to inefficient performance. For example, in a GOP where a scene change

65 occurs within the first frames, a linear distribution of bits as described above is suboptimal and the perceptual quality of the reconstructed video may be compromised. As information regarding picture complexity, scene complexity and bit allocation is available at the second encoding stage a different strategy should be used to control rate and buffer. In the first stage, not only characteristics of the video sequence are extracted but also a representation of the data with constant quality is produced: the Bit Usage Profile. In the second stage, a rate-control scheme is implemented allocating bits to each frame proportional to the number of bits used for its constant quality representation and complexity. The number of bits necessary to encode each frame in the first stage is a measure of the exigency associated with that frame, and, for constant quality, the target number of bits allocated to each frame in the second stage should be proportional to this exigency. The target number of bits to encode frame i, for the second stage, is determined by T~ =Kx~ax S, xr (4) " s~o~-s~, where gl,p. B are weighting factors (introduced to compensated the effect of the distinct quantization matrices applied to distinct picture types), Sc, oe, the number of bits spent on the first stage encoding the GOP which frame i belongs to, S i the number of bits spent encoding frame i and S(i, the number of bits spent on the first stage encoding up to frame i and R is initially equal to (5) (at the beginning of the GOP) and updated after encoding a picture (6): bit rate R = x N _of_framec, oe (5) frame rate R = R- St.p.~ (6) where Si,p, B is the number of bits generated to encode the last image. To determine the weighting factors, Kt, p.b, several simulations were performed to measure the effect of distinct quantization parameters in the number of allocated bits in variable bit rate, using several SIF625 sequences (Table 2, Figure 1). Q 2 4 6 8 10 12 14 Frame I 1.6 2.3 2.9 3.7 4.5 5.5 6.6 Frame P 1.5 1.8 2.1 2.5 2.8 3.2 3.6 Frame B 1 1 1 1 1 1 1 Tabel 2.a) Bits Frame Ratio (q)

66 Sequence Bit usage SNR Flower Garden 3227.40 mq_p" J.6892 47.437 mq_.p- '2 90 Coast Guard 1798.49 mq_p - ~.7351 45.030 mq_p -o.185o Bus 2095.52 mq..p -0.7o42 42.463 mq._p -0.23o3 Akiyo 812.24 mq_p -o,5s~7 42.581 mq._p -o.n22 Tabel 2.b) Modeling bit and SNR in function of quantization step t800.~ 'Z ~i '<..k ' il +1 ~o 1 "~'~ ~.,-~*'~,---- - i- oi Figure l.a) MB Bits (q) il t$ " " ; ; i" "''~" 2<::. to 2o ~o 40 5o egnqjen; Figure 1.b) MB SNR (q) The parameter q is modulated by the local and global activity of, respectively, the macroblock and the image. This bit allocation strategy is susceptible of providing a non-linear bit allocation, as determined by the video sequence characteristics. If, for example, the complexity of the first frame is such that it requires 30% of the total number of the GOP bits in the first stage, the target set for the second stage will also be 30% of the total number of bits available for the GOP. If a linear strategy were applied, this frame would receive roughly 8% of the GOP bits (considering a GOP of 12 frames length). Nevertheless, if a linear allocation is required by the characteristics of the video sequence, it can also be easily achieved. 2.3 Simulation Results Simulations were performed using SIF625 sequences. On the first stage, we have obtained the complexity of the video sequence and a measure of the exigency associated with each frame: the number of allocated of bits in VBR mode (figure 2). L,+!++!! -+,+~,,,l,,illlj~j~lb~l+j~j Jl.lllllt I -III~IIHIIIIII[IIHIIIII~! BI~tI~]t -kl~~~gl~iflllr 7'~ klilll "~I I II I!I I II I I~ Figure 2.1) Bits per frame (q),+~" r- +,if1... ~r - ;,... F""'- l Figure 2.2) Differential Local Activity '

67 t I Figure 3.1) Quantization step size Figure 3.2) PSNR From the local activity, we have obtained the differential local activity by subtracting adjacent local activity values. By correctly detecting shot scenes (pictures 100, 300 and 400), an improved bit rate algorithm control was developed. Using the information obtained in the first step the sequence is encoded. In general we obtain a small gain in PSNR (0.3 db) but the improvement in subjective quality is larger than what might be expected from such gain (mainly in shot scenes and hard to code pictures). The reason is the increased smoothness in the temporal variations of the coding quality and the reduction of the negative peaks associated with scenes changes Work with other video sequences is being performed to validate initial results. The way these techniques will interact is also being studied. 3. Discussion This paper describes a two-stage encoding process. Quality improvement is obtained by using a more global model of the source characteristics Improvement thus results from two passes with look-ahead bit allocation, special processing of scenes changes and an intelligent rate-control scheme. Future work will also address the problem of detection and classification of scene breaks (including cuts, fades, dissolves and wipes) To achieve this goal we are using techniques to determine real motion such as phase correlation 4o Acknowledgements This work was supported by the Junta Nacional Cientifica e Tecnol6gica (grant PRAXIS -XXI/4/4.1 ~D/2256). 5. References [1] ISO/IEC 11172-2, "Information Technology - Coding of Moving Pictures and Associated Audio for Digital Storage Media at up to about 1,5 Mbit/s-video", Geneva, 1993. [2] ISO/IEC IS 13818-2: Generic coding of moving pictures and associated audio, November 1994 [3] ISOflEC-JTCl/SC29/WGlt MPEG93/457 MPEG Video Test Mode/ 5 (TM-5), April 1993.

68 [4] K. Ramchandran, Ao Ortega and M. Vetterli, "Bit allocation for dependent quantization with applications to MPEG video codec", Proc. Internat. Conf. Acoust. Speech Signal Processing. 1993, Minneapolis, March 1993 [5] Gerjan Keesman, Imran Shah, Rene Klein-Gunnewiek, "Bit-rate control for MPEG encoders", Image Comm., vol. 6, pp. 545-560, June 1995 [6] Liang-Jin Lin, A. Ortega, C.-C. Jay Kuo, "oradient-based buffer control technique for MPEG", VCIP 95, Taipei, Taiwan, May 1995