论文标题

差异:渲染引导的3D辐射场扩散

DiffRF: Rendering-Guided 3D Radiance Field Diffusion

论文作者

Müller, Norman, Siddiqui, Yawar, Porzi, Lorenzo, Bulò, Samuel Rota, Kontschieder, Peter, Nießner, Matthias

论文摘要

我们介绍了DIFFRF,这是一种基于deno的扩散概率模型的3D辐射场合成的新方法。尽管现有的基于扩散的方法在图像,潜在代码或点云数据上运行,但我们是第一个直接生成体积辐射字段的人。为此,我们提出了一个直接在显式体素电网表示上运行的3D denoising模型。但是,由于从一组摆姿势的图像产生的辐射场可能是模棱两可的,并且包含伪影,因此获得地面真理辐射场样品是非平凡的。我们通过将剥离配方与渲染损失配对来应对这一挑战,从而使我们的模型能够学习偏离的先验,从而有利于良好的图像质量,而不是试图复制诸如浮动工件之类的拟合错误。与2D扩散模型相反,我们的模型学习了多视图一致的先验,从而实现了自由视图的合成和准确的形状生成。与3D GAN相比,我们基于扩散的方法自然可以实现有条件的生成,例如在推理时间时掩盖的完成或单视3D合成。

We introduce DiffRF, a novel approach for 3D radiance field synthesis based on denoising diffusion probabilistic models. While existing diffusion-based methods operate on images, latent codes, or point cloud data, we are the first to directly generate volumetric radiance fields. To this end, we propose a 3D denoising model which directly operates on an explicit voxel grid representation. However, as radiance fields generated from a set of posed images can be ambiguous and contain artifacts, obtaining ground truth radiance field samples is non-trivial. We address this challenge by pairing the denoising formulation with a rendering loss, enabling our model to learn a deviated prior that favours good image quality instead of trying to replicate fitting errors like floating artifacts. In contrast to 2D-diffusion models, our model learns multi-view consistent priors, enabling free-view synthesis and accurate shape generation. Compared to 3D GANs, our diffusion-based approach naturally enables conditional generation such as masked completion or single-view 3D synthesis at inference time.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源