A General Differentiable Mesh Renderer for Image-based 3D Reasoning (bibtex)
by Liu, Shichen, Li, Tianye, Chen, Weikai and Li, Hao
Abstract:
Rendering bridges the gap between 2D vision and 3D scenes by simulating the physical process of image formation. By inverting such renderer, one can think of a learning approach to infer 3D information from 2D images. However, standard graphics renderers involve a fundamental step called rasterization, which prevents rendering to be differentiable. Unlike the state-of-the-art differentiable renderers [25], [35], which only approximate the rendering gradient in the backpropagation, we propose a natually differentiable rendering framework that is able to (1) directly render colorized mesh using differentiable functions and (2) back-propagate efficient supervisions to mesh vertices and their attributes from various forms of image representations. The key to our framework is a novel formulation that views rendering as an aggregation function that fuses the probabilistic contributions of all mesh triangles with respect to the rendered pixels. Such formulation enables our framework to flow gradients to the occluded and distant vertices, which cannot be achieved by the previous state-of-the-arts. We show that by using the proposed renderer, one can achieve significant improvement in 3D unsupervised single-view reconstruction both qualitatively and quantitatively. Experiments also demonstrate that our approach can handle the challenging tasks in image-based shape fitting, which remain nontrivial to existing differentiable renders.
Reference:
A General Differentiable Mesh Renderer for Image-based 3D Reasoning (Liu, Shichen, Li, Tianye, Chen, Weikai and Li, Hao), In IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020.
Bibtex Entry:
@article{liu_general_2020,
	title = {A {General} {Differentiable} {Mesh} {Renderer} for {Image}-based {3D} {Reasoning}},
	issn = {0162-8828, 2160-9292, 1939-3539},
	url = {https://ieeexplore.ieee.org/document/9134794/},
	doi = {10.1109/TPAMI.2020.3007759},
	abstract = {Rendering bridges the gap between 2D vision and 3D scenes by simulating the physical process of image formation. By inverting such renderer, one can think of a learning approach to infer 3D information from 2D images. However, standard graphics renderers involve a fundamental step called rasterization, which prevents rendering to be differentiable. Unlike the state-of-the-art differentiable renderers [25], [35], which only approximate the rendering gradient in the backpropagation, we propose a natually differentiable rendering framework that is able to (1) directly render colorized mesh using differentiable functions and (2) back-propagate efficient supervisions to mesh vertices and their attributes from various forms of image representations. The key to our framework is a novel formulation that views rendering as an aggregation function that fuses the probabilistic contributions of all mesh triangles with respect to the rendered pixels. Such formulation enables our framework to flow gradients to the occluded and distant vertices, which cannot be achieved by the previous state-of-the-arts. We show that by using the proposed renderer, one can achieve significant improvement in 3D unsupervised single-view reconstruction both qualitatively and quantitatively. Experiments also demonstrate that our approach can handle the challenging tasks in image-based shape fitting, which remain nontrivial to existing differentiable renders.},
	journal = {IEEE Transactions on Pattern Analysis and Machine Intelligence},
	author = {Liu, Shichen and Li, Tianye and Chen, Weikai and Li, Hao},
	month = jul,
	year = {2020},
	keywords = {Graphics, ARO-Coop},
	pages = {1--1}
}
Powered by bibtexbrowser