Synthesizing light field from a single image with variable MPI and two network fusion Academic Article uri icon

abstract

  • We propose a learning-based approach to synthesize a light field with a small baseline from a single image. We synthesize the novel view images by first using a convolutional neural network (CNN) to promote the input image into a layered representation of the scene. We extend the multiplane image (MPI) representation by allowing the disparity of the layers to be inferred from the input image. We show that, compared to the original MPI representation, our representation models the scenes more accurately. Moreover, we propose to handle the visible and occluded regions separately through two parallel networks. The synthesized images using these two networks are then combined through a soft visibility mask to generate the final results. To effectively train the networks, we introduce a large-scale light field dataset of over 2,000 unique scenes containing a wide range of objects. We demonstrate that our approach synthesizes high-quality light fields on a variety of scenes, better than the state-of-the-art methods.

published proceedings

  • ACM Transactions on Graphics

altmetric score

  • 15

author list (cited authors)

  • Li, Q., & Kalantari, N. K.

citation count

  • 5

complete list of authors

  • Li, Qinbo||Kalantari, Nima Khademi

publication date

  • November 2020