• Home
  • About
    • Hoàng-Ân Lê photo

      Hoàng-Ân Lê

      I live, yet no longer I, but Christ lives in me (Gal 2:20)

    • Learn More
    • Twitter
    • Github
  • Posts
    • All Notes
    • All Tags
  • Projects

Novel View Synthesis from Single Images via Point Cloud Transformation

Hoang-An Le, Thomas Mensink, Partha Das, Theo Gevers
Computer Vision Group, Informatics Institute, University of Amsterdam

Reading time ~3 minutes

Figure 1 Overview of the proposed model for training and inference. From a single input image, the pixel-wise depth map is predicted. The depth map is subsequently used to compute a coarse novel view (forward warping), and trained by making use of backward warping (from the target view back to the source view). The model is trained end-to-end.

Abstract

In this paper the argument is made that for true novel view synthesis of objects, where the object can be synthesized from any viewpoint, an explicit 3D shape representation is desired. Our method estimates point clouds to capture the geometry of the object, which can be freely rotated into the desired view and then projected into a new image. This image, however, is sparse by nature and hence this coarse view is used as the input of an image completion network to obtain the dense target view. The point cloud is obtained using the predicted pixel-wise depth map, estimated from a single RGB input image, combined with the camera intrinsics. By using forward warping and backward warping between the input view and the target view, the network can be trained end-to-end without supervision on depth. The benefit of using point clouds as an explicit 3D shape for novel view synthesis is experimentally validated on the 3D ShapeNet benchmark.

Figure 2 Illustration of the forward and backward warping operation of point clouds. The forward warping is used to generate a coarse target view, while the backward warping is used to reconstruct the source view from a target view for self-supervised depth estimation.

Paper

BMVC | arxiv

Presentation

Code

Coming soon at github

Citation

If you find the material useful please consider citing our work

@inproceedings{le20bmvc,
 author = {Le, Hoang{-}An and Mensink, Thomas and Das, Partha and Gevers, Theo},
 title = {{Novel View Synthesis from Single Images via Point Cloud Transformation}},
 booktitle = {Proceedings of the British Machine Vision Conference (BMVC)},
 year = {2020},
}


researchcomputer visionCGInovel view synthesisself-supervisedmonocular depth predictionpoint cloudcamera model Share Tweet +1