论文标题
剪贴壁:纹理3D形态模型的文本指导编辑
ClipFace: Text-guided Editing of Textured 3D Morphable Models
论文作者
论文摘要
我们提出了夹层,这是一种新颖的自我监督方法,用于文本引导的纹理3D面孔模型的编辑。具体来说,我们采用用户友好的语言提示来控制表达式以及3D面的外观。我们利用3D形态模型的几何表现力,其本质上具有有限的可控性和纹理表达性,并开发了自我监督的生成模型,以共同合成3D表达式,纹理和铰接的面孔。我们通过对抗性自我监督的训练为3D面提供高质量的纹理生成,并以可区分的渲染为指导,以抵抗真正的RGB图像的收集。可控的编辑和操纵是由语言提示给出的,以调整3D形态模型的纹理和表达。为此,我们提出了一个神经网络,该神经网络可以预测可变形模型的纹理和表达潜在代码。我们的模型通过基于预先训练的剪辑模型利用可区分的渲染和损失来以自我监督的方式进行培训。一旦受过训练,我们的模型将共同预测紫外空间中的面部纹理,以及表达参数,以捕获单个正向传球中面部表情的几何和纹理变化。我们进一步展示了我们方法在给定动画序列生成时间变化的纹理的适用性。
We propose ClipFace, a novel self-supervised approach for text-guided editing of textured 3D morphable model of faces. Specifically, we employ user-friendly language prompts to enable control of the expressions as well as appearance of 3D faces. We leverage the geometric expressiveness of 3D morphable models, which inherently possess limited controllability and texture expressivity, and develop a self-supervised generative model to jointly synthesize expressive, textured, and articulated faces in 3D. We enable high-quality texture generation for 3D faces by adversarial self-supervised training, guided by differentiable rendering against collections of real RGB images. Controllable editing and manipulation are given by language prompts to adapt texture and expression of the 3D morphable model. To this end, we propose a neural network that predicts both texture and expression latent codes of the morphable model. Our model is trained in a self-supervised fashion by exploiting differentiable rendering and losses based on a pre-trained CLIP model. Once trained, our model jointly predicts face textures in UV-space, along with expression parameters to capture both geometry and texture changes in facial expressions in a single forward pass. We further show the applicability of our method to generate temporally changing textures for a given animation sequence.