Estimation of Virtual View Synthesis Distortion Toward Virtual View Position

Lu Fang, Yijian Xiang, Ngai Man Cheung, Feng Wu

Research output: Contribution to journalJournal Articlepeer-review

23 Citations (Scopus)

Abstract

We propose an analytical model to estimate the depth-error-induced virtual view synthesis distortion (VVSD) in 3D video, taking the distance between reference and virtual views (virtual view position) into account. In particular, we start with a comprehensive preanalysis and discussion over several possible VVSD scenarios. Taking intrinsic characteristic of each scenario into consideration, we specifically classify them into four clusters: 1) overlapping region; 2) disocclusion and boundary region; 3) edge region; and 4) infrequent region. We propose to model VVSD as the linear combination of the distortion under different scenarios (DDSs) weighted by the probability under different scenarios (PDSs). We show analytically that DDS and PDS can be related to the virtual view position using quadratic/biquadratic models and linear models, respectively. Experimental results verify that the proposed model is capable of estimating the relationship between VVSD and the distance between reference and virtual views. Therefore, our model can be used to inform a reference view setup for capturing, or distortion at certain virtual view positions, when depth information is compressed.

Original languageEnglish
Article number7430323
Pages (from-to)1961-1966
Number of pages6
JournalIEEE Transactions on Image Processing
Volume25
Issue number5
DOIs
Publication statusPublished - May 2016
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 2016 IEEE.

Keywords

  • 3d video
  • camera setup
  • depth error
  • virtual view synthesis

Fingerprint

Dive into the research topics of 'Estimation of Virtual View Synthesis Distortion Toward Virtual View Position'. Together they form a unique fingerprint.

Cite this