Efficient depth video coding based on view synthesis distortion estimation

Tae Young Chung, Won Dong Jang, Chang-Su Kim

Research output: Chapter in Book/Report/Conference proceedingConference contribution

10 Citations (Scopus)

Abstract

An efficient coding algorithm for depth map images and videos, based on view synthesis distortion estimation, is proposed in this work. We first analyze how a depth error is related to a disparity error and how the disparity vector error affects the energy spectral density of a synthesized color video in the frequency domain. Based on the analysis, we propose an estimation technique to predict the view synthesis distortion without requiring the actual synthesis of intermediate view frames. To encode the depth information efficiently, we employ a Lagrangian cost function to minimize the view synthesis distortion subject to the constraint on a transmission bit rate. In addition, we develop a quantization scheme for residual depth data, which adaptively assigns bits according to block complexities. Simulation results demonstrate that the proposed depth video coding algorithm provides significantly better R-D performance than conventional algorithms.

Original languageEnglish
Title of host publication2012 IEEE Visual Communications and Image Processing, VCIP 2012
DOIs
Publication statusPublished - 2012
Event2012 IEEE Visual Communications and Image Processing, VCIP 2012 - San Diego, CA, United States
Duration: 2012 Nov 272012 Nov 30

Publication series

Name2012 IEEE Visual Communications and Image Processing, VCIP 2012

Other

Other2012 IEEE Visual Communications and Image Processing, VCIP 2012
Country/TerritoryUnited States
CitySan Diego, CA
Period12/11/2712/11/30

Keywords

  • Multi-view plus depth
  • depth video coding
  • energy spectral density
  • view synthesis distortion

ASJC Scopus subject areas

  • Computer Vision and Pattern Recognition

Fingerprint

Dive into the research topics of 'Efficient depth video coding based on view synthesis distortion estimation'. Together they form a unique fingerprint.

Cite this