Bit allocation algorithm with novel view synthesis distortion model for multiview video plus depth coding

Tae Young Chung, Jae Young Sim, Chang-Su Kim

Research output: Contribution to journalArticlepeer-review

11 Citations (Scopus)

Abstract

An efficient bit allocation algorithm based on a novel view synthesis distortion model is proposed for the rate-distortion optimized coding of multiview video plus depth sequences in this paper. We decompose an input frame into nonedge blocks and edge blocks. For each nonedge block, we linearly approximate its texture and disparity values, and derive a view synthesis distortion model, which quantifies the impacts of the texture and depth distortions on the qualities of synthesized virtual views. On the other hand, for each edge block, we use its texture and disparity gradients for the distortion model. In addition, we formulate a bit-rate allocation problem in terms of the quantization parameters for texture and depth data. By solving the problem, we can optimally divide a limited bit budget between the texture and depth data, in order to maximize the qualities of synthesized virtual views, as well as those of encoded real views. Experimental results demonstrate that the proposed algorithm yields the average PSNR gains of 1.98 and 2.04 dB in two-view and three-view scenarios, respectively, as compared with a benchmark conventional algorithm.

Original languageEnglish
Article number6824213
Pages (from-to)3254-3267
Number of pages14
JournalIEEE Transactions on Image Processing
Volume23
Issue number8
DOIs
Publication statusPublished - 2014 Aug

Keywords

  • Multi-view video plus depth
  • and rate-distortion optimization
  • bit allocation
  • view synthesis distortion
  • virtual view synthesis

ASJC Scopus subject areas

  • Software
  • Computer Graphics and Computer-Aided Design

Fingerprint

Dive into the research topics of 'Bit allocation algorithm with novel view synthesis distortion model for multiview video plus depth coding'. Together they form a unique fingerprint.

Cite this