Spatial-Angular Attention Network for Light Field Reconstruction

Gaochang Wu, Yingqian Wang, Yebin Liu, Lu Fang, Tianyou Chai*

*Corresponding author for this work

Research output: Contribution to journalJournal Articlepeer-review

57 Citations (Scopus)

Abstract

Typical learning-based light field reconstruction methods demand in constructing a large receptive field by deepening their networks to capture correspondences between input views. In this paper, we propose a spatial-angular attention network to perceive non-local correspondences in the light field, and reconstruct high angular resolution light field in an end-to-end manner. Motivated by the non-local attention mechanism (Wang et al., 2018; Zhang et al., 2019), a spatial-angular attention module specifically for the high-dimensional light field data is introduced to compute the response of each query pixel from all the positions on the epipolar plane, and generate an attention map that captures correspondences along the angular dimension. Then a multi-scale reconstruction structure is proposed to efficiently implement the non-local attention in the low resolution feature space, while also preserving the high frequency components in the high-resolution feature space. Extensive experiments demonstrate the superior performance of the proposed spatial-angular attention network for reconstructing sparsely-sampled light fields with Non-Lambertian effects.

Original languageEnglish
Pages (from-to)8999-9013
Number of pages15
JournalIEEE Transactions on Image Processing
Volume30
DOIs
Publication statusPublished - 2021
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 1992-2012 IEEE.

Keywords

  • Light field reconstruction
  • attention mechanism
  • deep learning

Fingerprint

Dive into the research topics of 'Spatial-Angular Attention Network for Light Field Reconstruction'. Together they form a unique fingerprint.

Cite this