论文标题

反图形gan:学习从非结构化的2D数据生成3D形状

Inverse Graphics GAN: Learning to Generate 3D Shapes from Unstructured 2D Data

论文作者

Lunz, Sebastian, Li, Yingzhen, Fitzgibbon, Andrew, Kushman, Nate

论文摘要

最近的工作表明,仅从非结构化的2D图像中学习3D形状的生成模型的能力。但是,培训此类模型需要通过渲染过程的栅格化步骤进行区分,因此过去的工作重点是开发定制渲染模型,这些模型以各种方式平滑地胜过这个非差异过程。因此,这样的模型无法利用由游戏和图形行业建造的照片真实,完整的工业渲染器。在本文中,我们从2D数据中介绍了第一个用于3D生成模型的可扩展培训技术,该技术利用了现成的非差异渲染器。为了说明非差异性,我们引入了一个代理神经渲染器,以匹配非差异性渲染器的输出。我们进一步提出歧视器输出匹配,以确保神经渲染器学会适当地对栅格化进行平滑。我们对从生成的3D形状呈现的图像进行评估,并表明我们的模型可以始终如一地学习使用独家非结构化2D图像训练的现有模型比现有模型更好的形状。

Recent work has shown the ability to learn generative models for 3D shapes from only unstructured 2D images. However, training such models requires differentiating through the rasterization step of the rendering process, therefore past work has focused on developing bespoke rendering models which smooth over this non-differentiable process in various ways. Such models are thus unable to take advantage of the photo-realistic, fully featured, industrial renderers built by the gaming and graphics industry. In this paper we introduce the first scalable training technique for 3D generative models from 2D data which utilizes an off-the-shelf non-differentiable renderer. To account for the non-differentiability, we introduce a proxy neural renderer to match the output of the non-differentiable renderer. We further propose discriminator output matching to ensure that the neural renderer learns to smooth over the rasterization appropriately. We evaluate our model on images rendered from our generated 3D shapes, and show that our model can consistently learn to generate better shapes than existing models when trained with exclusively unstructured 2D images.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源