MULTIVIEW DISTRIBUTED VIDEO CODING WITH ENCODER DRIVEN FUSION

Similar documents
Free Viewpoint Switching in Multi-view Video Streaming Using. Wyner-Ziv Video Coding

WYNER-ZIV VIDEO CODING WITH LOW ENCODER COMPLEXITY

CHROMA CODING IN DISTRIBUTED VIDEO CODING

Distributed Video Coding Using LDPC Codes for Wireless Video

A Study of Encoding and Decoding Techniques for Syndrome-Based Video Coding

Project Proposal: Sub pixel motion estimation for side information generation in Wyner- Ziv decoder.

University of Bristol - Explore Bristol Research. Peer reviewed version. Link to published version (if available): /ISCAS.2005.

Modeling and Optimization of a Systematic Lossy Error Protection System based on H.264/AVC Redundant Slices

Systematic Lossy Error Protection of Video based on H.264/AVC Redundant Slices

INFORMATION THEORY INSPIRED VIDEO CODING METHODS : TRUTH IS SOMETIMES BETTER THAN FICTION

MULTI-STATE VIDEO CODING WITH SIDE INFORMATION. Sila Ekmekci Flierl, Thomas Sikora

Robust 3-D Video System Based on Modified Prediction Coding and Adaptive Selection Mode Error Concealment Algorithm

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video

Wyner-Ziv Coding of Motion Video

UC San Diego UC San Diego Previously Published Works

Chapter 10 Basic Video Compression Techniques

Energy Efficient Video Compression for Wireless Sensor Networks *

Adaptive mode decision with residual motion compensation for distributed video coding

Fast MBAFF/PAFF Motion Estimation and Mode Decision Scheme for H.264

Systematic Lossy Forward Error Protection for Error-Resilient Digital Video Broadcasting

Interactive multiview video system with non-complex navigation at the decoder

Video compression principles. Color Space Conversion. Sub-sampling of Chrominance Information. Video: moving pictures and the terms frame and

Multi-view Video Streaming with Mobile Cameras

Popularity-Aware Rate Allocation in Multi-View Video

ROBUST ADAPTIVE INTRA REFRESH FOR MULTIVIEW VIDEO

An Efficient Low Bit-Rate Video-Coding Algorithm Focusing on Moving Regions

Decoder-driven mode decision in a block-based distributed video codec

Reduced complexity MPEG2 video post-processing for HD display

Multiview Video Coding

Joint Optimization of Source-Channel Video Coding Using the H.264/AVC encoder and FEC Codes. Digital Signal and Image Processing Lab

Research Topic. Error Concealment Techniques in H.264/AVC for Wireless Video Transmission in Mobile Networks

Multiple Description H.264 Video Coding with Redundant Pictures

Camera Motion-constraint Video Codec Selection

Video Quality Monitoring for Mobile Multicast Peers Using Distributed Source Coding

Advanced Video Processing for Future Multimedia Communication Systems

Principles of Video Compression

Dual frame motion compensation for a rate switching network

A Novel Approach towards Video Compression for Mobile Internet using Transform Domain Technique

Systematic Lossy Error Protection of Video Signals Shantanu Rane, Member, IEEE, Pierpaolo Baccichet, Member, IEEE, and Bernd Girod, Fellow, IEEE

Systematic Lossy Error Protection based on H.264/AVC Redundant Slices and Flexible Macroblock Ordering

Dual Frame Video Encoding with Feedback

Digital Video Telemetry System

Error Resilience for Compressed Sensing with Multiple-Channel Transmission

Agrowing percentage of the world population now uses image and

Wyner-Ziv to H.264 Video Transcoder for Low Cost Video Encoding

Real-Time Distributed Video Coding for 1K-pixel Visual Sensor Networks

UNIVERSAL SPATIAL UP-SCALER WITH NONLINEAR EDGE ENHANCEMENT

Adaptive Key Frame Selection for Efficient Video Coding

Exploring the Distributed Video Coding in a Quality Assessment Context

1022 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 19, NO. 4, APRIL 2010

Video coding standards

SCALABLE video coding (SVC) is currently being developed

Visual Communication at Limited Colour Display Capability

EMBEDDED ZEROTREE WAVELET CODING WITH JOINT HUFFMAN AND ARITHMETIC CODING

COMP 249 Advanced Distributed Systems Multimedia Networking. Video Compression Standards

AN IMPROVED ERROR CONCEALMENT STRATEGY DRIVEN BY SCENE MOTION PROPERTIES FOR H.264/AVC DECODERS

Motion Re-estimation for MPEG-2 to MPEG-4 Simple Profile Transcoding. Abstract. I. Introduction

UNBALANCED QUANTIZED MULTI-STATE VIDEO CODING

DELTA MODULATION AND DPCM CODING OF COLOR SIGNALS

High performance and low complexity decoding light-weight video coding with motion estimation and mode decision at decoder

ENCODING OF PREDICTIVE ERROR FRAMES IN RATE SCALABLE VIDEO CODECS USING WAVELET SHRINKAGE. Eduardo Asbun, Paul Salama, and Edward J.

Performance Evaluation of Error Resilience Techniques in H.264/AVC Standard

Feasibility Study of Stochastic Streaming with 4K UHD Video Traces

Wireless Multi-view Video Streaming with Subcarrier Allocation by Frame Significance

WITH the rapid development of high-fidelity video services

The H.263+ Video Coding Standard: Complexity and Performance

A robust video encoding scheme to enhance error concealment of intra frames

Selective Intra Prediction Mode Decision for H.264/AVC Encoders

Constant Bit Rate for Video Streaming Over Packet Switching Networks

Comparative Study of JPEG2000 and H.264/AVC FRExt I Frame Coding on High-Definition Video Sequences

Adaptive Distributed Compressed Video Sensing

1. INTRODUCTION. Index Terms Video Transcoding, Video Streaming, Frame skipping, Interpolation frame, Decoder, Encoder.

Distributed Video Coding: Selecting the Most Promising Application Scenarios

AUDIOVISUAL COMMUNICATION

PACKET-SWITCHED networks have become ubiquitous

Encoder-driven rate control and mode decision for distributed video coding

ARTICLE IN PRESS. Signal Processing: Image Communication

Error Resilient Video Coding Using Unequally Protected Key Pictures

Error concealment techniques in H.264 video transmission over wireless networks

Analysis of Video Transmission over Lossy Channels

GLOBAL DISPARITY COMPENSATION FOR MULTI-VIEW VIDEO CODING. Kwan-Jung Oh and Yo-Sung Ho

Dual frame motion compensation for a rate switching network

Hierarchical SNR Scalable Video Coding with Adaptive Quantization for Reduced Drift Error

Region Adaptive Unsharp Masking based DCT Interpolation for Efficient Video Intra Frame Up-sampling

Reduced Decoder Complexity and Latency in Pixel-Domain Wyner-Ziv Video Coders

Analysis of MPEG-2 Video Streams

HEVC Subjective Video Quality Test Results

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur

Intra-frame JPEG-2000 vs. Inter-frame Compression Comparison: The benefits and trade-offs for very high quality, high resolution sequences

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 58, NO. 5, MAY Note that the term distributed coding in this paper is always employed

Marie Ramon, François-XavierCoudoux, andmarcgazalet. 1. Introduction

ROBUST IMAGE AND VIDEO CODING WITH ADAPTIVE RATE CONTROL

17 October About H.265/HEVC. Things you should know about the new encoding.

Robust Transmission of H.264/AVC Video using 64-QAM and unequal error protection

PAPER Wireless Multi-view Video Streaming with Subcarrier Allocation

PERCEPTUAL QUALITY ASSESSMENT FOR VIDEO WATERMARKING. Stefan Winkler, Elisa Drelie Gelasca, Touradj Ebrahimi

Scalable multiple description coding of video sequences

Chapter 2 Introduction to

Understanding Compression Technologies for HD and Megapixel Surveillance

1934 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 4, APRIL 2012

Transcription:

MULTIVIEW DISTRIBUTED VIDEO CODING WITH ENCODER DRIVEN FUSION Mourad Ouaret, Frederic Dufaux and Touradj Ebrahimi Institut de Traitement des Signaux Ecole Polytechnique Fédérale de Lausanne (EPFL), CH-1015 Lausanne, Switzerland {mourad.ouaret, frederic.dufaux, touradj.ebrahimi}@epfl.ch ABSTRACT In this paper, a scheme based on multiview and Distributed Video Coding (DVC) is introduced. More specifically, a new fusion technique between temporal, homography and Disparity Compensation View Prediction (DCVP) side informations is introduced. For this purpose, a binary mask is computed at the encoder based on knowledge of the original video. This mask is compressed and then transmitted to the decoder. The latter merges the different side informations with the help of the reconstructed binary mask. The simulation results show that the fusion improves the rate-distortion performance over monoview DVC by a maximum gap of around 1.0 db. 1. INTRODUCTION Due to the reducing cost of cameras and improved display technology, multi-camera systems are being used in many fields such as video surveillance and monitoring, 3D reconstruction and Free Viewpoint Television (FTV). However, the amount of captured video in such systems is often huge. This makes video compression important and a key issue in multiview video systems. Multiview Video Coding (MVC) [1] is the consequence of work conducted by MPEG in 3D Audio-Video (3DAV). It is based on H.264/AVC [2] used for single camera encoding. It performs block-based predictive coding across the different views as well as the time axis of each camera. Predictive coding gives the best compression efficiency. On the other hand, the encoder requires high computational power in addition to communication between the cameras in a practical scenario. However, this is not feasible as it requires complex inter-camera communicating systems, which is time and power consuming and entails complex networking issues. For the sake of having less complex encoders, work is conducted in the field of Distributed Video Coding (DVC) [4]. Theoretically, it states that the rate achieved when performing joint encoding and decoding of two sources can be reached by doing separate encoding and joint decoding. In a practical scenario, this implies low power / low complexity cameras as well as no communication between the cameras. On the other hand, DVC shift the complexity towards the decoder. In monoview DVC schemes, side information is often generated temporally using frames from the same camera, usually the previous and the forward ones. In multiview DVC [5, 6, 7], frames from the side cameras are also involved in generating the side information. The latter is combined, or fused, with the one generated temporally in order to improve the compression efficiency. In [5], View Synthesis Prediction (VSP) is used to generate side information from the side cameras. However, the rate-distortion performance of the approach is not investigated. In addition, VSP requires depth map estimation, which is a hard problem for real world scenes. A fusion technique is used with multiview DVC based on pixel-difference and motion vector thresholding in [6]. Finally in [7], a fusion technique is introduced between temporal and homography-based side informations to improve the overall rate-distortion performance. The homography is estimated using a robust gradient descent algorithm. The fusion consists of two merging algorithms, one is used at low bitrates and the other one at high bitrates. The simulations in [7] report that the fusion outperforms monoview DVC by around 0.2~0.5 db. In this paper, a fusion technique is introduced based on some prior knowledge of the original video. For this purpose, a binary mask is computed at the encoder. Then, it is compressed using JBIG [8] and transmitted to the decoder. The latter uses the recovered mask along with temporal, homography and DCVP side informations to construct a better side information. The paper is structured as follows. First, the monoview and multiview DVC architectures are introduced in section 2 and 3 respectively. Then, the different side information generation techniques are presented in section 4. Then, the simulation results are presented in section 5. Finally, some concluding remarks are drawn in section 6. 2. MONOVIEW DISTRIBUTED VIDEO CODING DVC is the consequence of information-theoretic bounds established by Slepian and Wolf [9] for distributed lossless coding, and by Wyner and Ziv [10] for lossy coding with decoder side information. In a practical scenario, lossy coding is used. In this paper, the DVC scheme from [11] is used. The latter is illustrated in Fig. 1. 2007 EURASIP 2100

Figure 1. Conventional DVC scheme (GOP =2). The Wyner-Ziv encoder operates in the DCT domain. In other words, an interleaved turbo encoder is used to generate parity bits for the quantized DCT coefficients. Moreover, the quantized DCT coefficients are organized in bands, which are separated into bitplanes. The latter are organized from the most to the least significant one. The parity bits are generated for a certain number of bitplanes. As the number of bitplanes increases, the quality of the refined frame improves. For simplicity, the case where the Group Of Pictures (GOP) is equal to two is considered. The conventionally decoded previous and forward key frames are used to generate side information by motion compensated interpolation. To exploit the side information, the decoder assumes a statistical model, which is a Laplacian distribution of the difference between the individual DCT coefficients of the original Wyner-Ziv frame and the side information. The decoder combines the side information and the received parity bits to recover the original frame. For more details on the used DVC scheme, see [4], [10] and [11]. 3. MULTIVIEW DISTRIBUTED VIDEO CODING Multiview DVC differs from the monoview one in the frames involved in the side information generation process. In monoview DVC, frames within the same camera are used for that purpose. In multiview DVC, frames from other cameras can additionally be used to generate side information. In this work, DCVP, homography and temporal side informations are merged at the decoder based on some prior knowledge of the original video. For this purpose, the encoder transmits a binary mask to the decoder in order to define the reference used for each pixel. The mask is compressed using JBIG prior to transmission. The resulting multiview DVC scheme is shown in Figure 2. Figure 2. Multiview DVC scheme (GOP =2). 4. SIDE INFORMATION GENERATION 4.1 Temporal Motion Compensated side information Temporal side information is used in monoview DVC. It is generated by temporal motion estimation using the previous and the forward key frames. Block-based motion vectors from the previous frame towards the forward frame are computed. Then, the motion vectors are interpolated at mid point to generate the side information. This is done by considering the intersection point of each motion vector with a virtual frame at mid-distance from both key frames as shown in Figure 3. The virtual frame is filled at the block position, which is closest to the intersection point. The latter is filled by a weighted sum of both key frame blocks. Figure 3. Temporal side information (GOP =2). 4.2 Homography side information The homography is a 3x3 matrix that relates one view to another one in the homogenous coordinates system. The matrix has 8 parameters a, b, c, d, e, f, g and h, such that each point from the first view (x 1,y 1 ) is mapped to a point (x 2,y 2 ) in the second view up to a scale λ such that : x 2 a b c x 1 λ y d e f y 2 = 1 1 g h 1 1 x 2 = ax 1 + by 1 + c gx 1 + hy 1 + 1 y 2 = dx 1 + ey 1 + f gx + hy + 1 1 When a=e=1 and b=d=g=h=0 the model is a pure translation. When g=h=0 the model is called an affine transformation. Otherwise, it is called a perspective transformation. These models are suitable when the scene can be approximated by a 1 2007 EURASIP 2101

planar surface, or when the scene is static and the camera motion is a pure rotation around its optical center [12]. In our case, the first assumption applies. Depending on the model used, the parameters are computed such that the sum of squared differences between the current frame and the warped frame is minimized. To compute the model parameters, a gradient descent method [12] is used. The latter minimizes a truncated quadratic error function to remove the influence of outliers. Different side informations are computed using the homogaphies from the side cameras. More precisely, the side information is computed from the left, right or both cameras as shown in Figure 4. At the encoder: Each pixel of the Wyner-Ziv frame is compared to the ones from the previous and the forward frames. If the one from the previous pixel has a closer value, the binary mask at the pixel position is set to one. On the other hand, if the forward pixel has a closer value, it is set to zero. This process is illustrated in Figure 5. Figure 4. Homographies relating the central view to the side ones. 4.3 Disparity Compensation View Prediction with variable motion vector weighting DCVP uses the same algorithm as the one used to generate temporal side information. But instead of using the previous and forward frames of the same camera, DCVP uses the left and right frames from the side cameras. Interpolating the motion vectors at mid-point means that middle camera is located exactly at equal distance from the two side cameras. This is not true in all cases. To calculate the optimal motion vectors weight, the first frame of each camera is conventionally encoded and then decoded. The decoder performs block-based motion estimation between the left and the right camera frames. Then, the motion vectors are weighted with the weights 0.1, 0.2 until 0.9. For each weight the Peak Signal to Noise Ratio (PSNR) with respect to the central camera frame is computed. The weight with maximum PSNR is kept and used for the rest of the video. 4.4 Fusion-based side information In general, the side information generated from the side cameras (either by homography or DCVP) has a poorer quality than the one generated temporally. This is due to the larger disparity between the side camera frames when compared to the one between the previous and forward frames. The fusion merges the different side informations (temporal, homography and DCVP) in order to improve the quality of the final one. The idea is to determine a very good estimate of the Wyner-Ziv frame, which is called the reference. The decision for each pixel (i.e. either DCVP, homography or temporal side information) is taken with respect to this reference, as detailed here after. Figure 5. The binary mask generation at the encoder. The binary mask is encoded using JBIG from the Joint Bi-level Image experts Group [8]. The encoding is preceded by a morphological closing and opening [13]. This eliminates isolated points in the binary mask that would compromise the compression efficiency of the JBIG software. At the decoder: The fusion-based side information generation is illustrated in Figure 6. The binary mask is recovered and the different side informations, temporal, homography and DCVP are computed. The binary mask defines for each pixel which reference to use. One in the binary mask means that the pixel values from both side informations should be compared with pixel from the previously decoded frame. On the other hand, a zero in the binary mask means that the comparison is made with respect to the forward one. 5. RATE-DISTORTION SIMULATIONS 5.1 Test material and conditions The sequences Ballet and Breakdancers [14] are used to compute the rate-distortion curves for both monoview and multiview DVC schemes. The spatio-temporal resolution used is 256x192@15 frames per second. The GOP size is set to two. H.264/AVC Intra is used for encoding the keyframes. The camera setup shown in Figure 7 is used. The side cameras are conventionally encoded. Thus, Only the middle camera contains Wyner-Ziv frames. 2007 EURASIP 2102

greater in the Breakdancers case. This makes temporal interpolation less efficient. Side infromation quality (ballet) Temporal Fusion H avg H Left H Right DCVP 34.5 33.5 32.5.5.5 29.5 28.5 27.5 26.5 25.5 24.5 23.5 22.5 21.5 1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 33 35 37 39 41 43 45 47 49 Frame index Figure 9. Side information quality for Ballet. Figure 6. Fusion process at the decoder. The R-D plots in Figure 10 show that fusion outperforms temporal side information by around 1.0 db at low bit rates for the sequence Breakdancers. As the bit rate increases, the performance gap decreases. The plots show a 0.5 db performance gap at average bit rates and a similar performance at very high video quality. Further, the homography has a similar performance at low bit rates as temporal side information. The gap increases with bit rate to reach a maximum of around 1.0 db at very high bit rates. Finally, DCVP has the worst performance with a gap around 1.0~2.0 db in favour of temporal side information. Figure 7. Multiview DVC camera setup. I stands for Intra frame and WZ for Wyner-Ziv frame. 29 28 27 26 25 24 23 22 Side infromation quality (Breakdancers) Temporal Fusion H avg DCVP 1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 33 35 37 39 41 43 45 47 49 Frame index Figure 8. Side information quality for Breakdancers. 5.2 Rate-Distortion performance In Figures 8 and 9, the PSNR of the different side informations is plotted. It is obvious that the temporal side information quality for the sequence Ballet is by far superior to homography and DCVP by around 7.0 and 8.0 db respectively. This gap is much smaller for the sequence Breakdancers. The latter contains higher motion than the sequence Ballet. Moreover, the area occupied by the moving object is 41 40 39 38 37 36 35 34 33 32 Breakdancers Temporal Fusion DCVP Homography Ideal fusion 100 200 0 400 500 600 700 800 Bit rate Figure 10. R-D plots for the sequence Breakdancers. H.264 Intra + Wyner Ziv rate Mask rate Total rate Mask rate (%) 128.44.49 11.92 140.36 8.492447991 234.38 33.95 12.9 247.28 5.2167583 358.08 35.71 13.38 371.46 3.602002907 780.05 40.44 14.7 794.75 1.849638251 Table 1. The four R-D points for the Breakdancers sequence. For the sequence Ballet, The fusion behaves in a similar way with respect to temporal side information as for Breakdancers. It is illustrated in Figure 11. On the other hand, the 2007 EURASIP 2103

homography and DCVP have a poorer performance with respect to temporal side information when compared to Breakdancers. When comparing these results to the ones in [7] (i.e. when the mask is entirely computed at the decoder), the gain is around 0.5 db at low bit rates and 0.3 db at average bit rates. For both sequences, an ideal fusion is computed by using the original Wyner-Ziv frame as a reference. It has a better performance than temporal side information by around 3.0 and 4.0 db for the sequences Ballet and Breakdancers respectively. 43 42 41 40 39 38 37 36 35 34 33 32 29 Ballet Temporal Fusion DCVP Homography Ideal fusion 100 200 0 400 500 600 700 Bit rate Figure 11. R-D plots for the sequence Ballet. H.264 Intra + Wyner Ziv rate Mask rate Total rate Mask rate (%) 125.79 34.06 13.8 139.59 9.886094992 221.27 36.57 15 236.27 6.348668896 321.44 38.45 15.4 336.84 4.571903574 629.53 42.19 16.15 645.68 2.501239004 Table 2. The four R-D points for the sequence Ballet. The tables 1 and 2 show the four R-D points for both sequences. It is noticed that the mask s rate decreases as the video quality increases. Thus, the mask s rate becomes less significant with increased quality as the total rate increases as well. 6. CONCLUSION In this paper, a novel multiview DVC scheme is introduced. In this scheme, a binary mask is calculated at the encoder based on knowledge of the original video. Then, the mask is compressed and transmitted to the decoder. The final side information is constructed from the temporal, homography and DCVP side informations using the binary mask. It is shown that the fusion improves the side information and the overall rate-distortion performance. The system introduced is interesting for applications requiring low complexity encoders such as distributed sensor networks. This work can be extended by investigating other fusion techniques. The focus should be on the ones where the fusion mask is entirely calculated at the decoder. It is obvious that the fusion has to improve the overall rate-distortion performance and close the gap on the ideal fusion as well. 7. ACKNOWLEDGEMENT This work was developed within DISCOVER, a European Project (www.discoverdvc.org), funded under the European Commission IST FP6 programme. We would like to thank Microsoft research for providing the video sequences Ballet and Breakdnacers. 8. REFERENCES [1] http://www.chiariglione.org/mpeg. [2] ThomasWiegand, Gary J. Sullivan, Gisle Bjøntegaard, and Ajay Luthra, "Overview of the H.264/AVC Video Coding Standard,'' IEEE Trans. on Circuits and Systems for Video Technology, vol. 13, no. 7, July 2003. [3] Martinian, E., Behrens, A., Xin, J., Vetro, A., "View Synthesis for Multiview Video Compression,'' Picture Coding Symposium (PCS), April 2006. [4] Bernd Girod, Anne Aaron, Shantanu Rane and David Rebollo-Monedero, "Distributed Video Coding,'' Proceedings of the IEEE, vol. 93, no. 1, pp. 71-83, January 2005. [5] Xavi Artigas, Egon Angeli, and Luis Torres, "Side Information Generation for Multiview Distributed Video Coding Using a Fusion Approach,'' 7th Nordic Signal Processing Symposium (NORSIG), Reykjavik, Iceland, June 7-9, 2006. [6] Xun Guo, Yan Lu, Feng Wu, Wen Gao, Shipeng Li, "Distributed Multi-view Video Coding,'' Visual Communications and Image Processing 2006, San Jose, California, USA, January 17-19 2006. [7] M.Ouaret, F. Dufaux, T.Ebrahimi, "Fusion-based Multiview Distributed Video Coding,'' fourth ACM international workshop on Video surveillance and sensor networks 2006, Santa Barbara, California, October 27-27, 2006. [8] http://www.jpeg.org/jbig/index.html. [9] J. Slepian and J. Wolf, "Noiseless Coding of Correlated Information Sources,'' IEEE Trans. on Information Theory, vol. 19, no. 4, July 1973. [10] A. Wyner and J. Ziv, "The Rate-Distortion Function for Source Coding with Side Information at the Decoder,'' IEEE Trans. on Information Theory, vol. 22, no. 1, January 1976. [11] C. Brites, J. Ascenso, F. Pereira, "Improving Transform Domain Wyner-Ziv Video Coding Performance,'' International Conference on Acoustics, Speech and Signal Processing, Toulouse, France, May 2006. [12] Frederic Dufaux and Janusz Konrad, Efficient, "Robust, and Fast Global Motion Estimation for Video Coding,'' IEEE transactions on image processing, vol. 9, no.3, March 2000. [13] R. M. Haralick, S. R. Stemberg, and X. Zhuang, "Image analysis using mathematical morphology,'' IEEE Trans. Pattern Anal Machine Intell., vol. 9, pp. 523-550, July 1987. [14] http://research.microsoft.com/ivm/3dvideodownload/. 2007 EURASIP 2104