Efficient depth video coding based on view synthesis distortion estimation

Tae Young Chung, Won Dong Jang, Chang-Su Kim

    Research output: Chapter in Book/Report/Conference proceedingConference contribution

    10 Citations (Scopus)

    Abstract

    An efficient coding algorithm for depth map images and videos, based on view synthesis distortion estimation, is proposed in this work. We first analyze how a depth error is related to a disparity error and how the disparity vector error affects the energy spectral density of a synthesized color video in the frequency domain. Based on the analysis, we propose an estimation technique to predict the view synthesis distortion without requiring the actual synthesis of intermediate view frames. To encode the depth information efficiently, we employ a Lagrangian cost function to minimize the view synthesis distortion subject to the constraint on a transmission bit rate. In addition, we develop a quantization scheme for residual depth data, which adaptively assigns bits according to block complexities. Simulation results demonstrate that the proposed depth video coding algorithm provides significantly better R-D performance than conventional algorithms.

    Original languageEnglish
    Title of host publication2012 IEEE Visual Communications and Image Processing, VCIP 2012
    DOIs
    Publication statusPublished - 2012
    Event2012 IEEE Visual Communications and Image Processing, VCIP 2012 - San Diego, CA, United States
    Duration: 2012 Nov 272012 Nov 30

    Publication series

    Name2012 IEEE Visual Communications and Image Processing, VCIP 2012

    Other

    Other2012 IEEE Visual Communications and Image Processing, VCIP 2012
    Country/TerritoryUnited States
    CitySan Diego, CA
    Period12/11/2712/11/30

    Keywords

    • Multi-view plus depth
    • depth video coding
    • energy spectral density
    • view synthesis distortion

    ASJC Scopus subject areas

    • Computer Vision and Pattern Recognition

    Fingerprint

    Dive into the research topics of 'Efficient depth video coding based on view synthesis distortion estimation'. Together they form a unique fingerprint.

    Cite this