Header logo is

Image-guided Neural Object Rendering

2020

Book Chapter

ncs


We propose a learned image-guided rendering technique that combines the benefits of image-based rendering and GAN-based image synthesis. The goal of our method is to generate photo-realistic re-renderings of reconstructed objects for virtual and augmented reality applications (e.g., virtual showrooms, virtual tours and sightseeing, the digital inspection of historical artifacts). A core component of our work is the handling of view-dependent effects. Specifically, we directly train an object-specific deep neural network to synthesize the view-dependent appearance of an object. As input data we are using an RGB video of the object. This video is used to reconstruct a proxy geometry of the object via multi-view stereo. Based on this 3D proxy, the appearance of a captured view can be warped into a new target view as in classical image-based rendering. This warping assumes diffuse surfaces, in case of view-dependent effects, such as specular highlights, it leads to artifacts. To this end, we propose EffectsNet, a deep neural network that predicts view-dependent effects. Based on these estimations, we are able to convert observed images to diffuse images. These diffuse images can be projected into other views. In the target view, our pipeline reinserts the new view-dependent effects. To composite multiple reprojected images to a final output, we learn a composition network that outputs photo-realistic results. Using this image-guided approach, the network does not have to allocate capacity on ``remembering’’ object appearance, instead it learns how to combine the appearance of captured images. We demonstrate the effectiveness of our approach both qualitatively and quantitatively on synthetic as well as on real data.

Author(s): Thies, Justus and Zollhöfer, Michael and Theobalt, Christian and Stamminger, Marc and Nießner, Matthias
Book Title: International Conference on Learning Representations
Year: 2020

Department(s): Neural Capture and Synthesis
Bibtex Type: Book Chapter (incollection)

Event Place: Addis Ababa, Ethiopia
URL: https://justusthies.github.io/posts/ignor/

Links: Paper
Video
Video:

BibTex

@incollection{thies2020imageguided,
  title = {Image-guided Neural Object Rendering},
  author = {Thies, Justus and Zollh{\"o}fer, Michael and Theobalt, Christian and Stamminger, Marc and Nie{\ss}ner, Matthias},
  booktitle = {International Conference on Learning Representations},
  year = {2020},
  doi = {},
  url = {https://justusthies.github.io/posts/ignor/}
}