论文标题
Humannerf:从单眼视频中移动人们的免费观看点渲染
HumanNeRF: Free-viewpoint Rendering of Moving People from Monocular Video
论文作者
论文摘要
我们介绍了一种免费观看点渲染方法 - Humannerf-在给定的人类表演复杂身体运动的给定视频中工作,例如来自YouTube的视频。我们的方法可以在任何框架上暂停视频,并从任意的新相机观点甚至为特定框架和身体姿势的完整360度相机路径渲染主题。这项任务尤其具有挑战性,因为它需要综合身体的光真逼真的细节,从输入视频中可能不存在的各种摄像头角度看,以及综合细节(例如布褶皱和面部外观)。我们的方法优化了在规范t置孔中对人的体积表示,并与运动场一起通过向后扭曲将估计的规范表示形式映射到视频的每个帧。运动场分解为由深网产生的骨骼刚性和非刚性运动。我们在先前的工作中显示出显着的性能改善,并从单眼视频中引起了自由视频渲染的示例,这些视频在挑战不受控制的捕获场景中移动人类。
We introduce a free-viewpoint rendering method -- HumanNeRF -- that works on a given monocular video of a human performing complex body motions, e.g. a video from YouTube. Our method enables pausing the video at any frame and rendering the subject from arbitrary new camera viewpoints or even a full 360-degree camera path for that particular frame and body pose. This task is particularly challenging, as it requires synthesizing photorealistic details of the body, as seen from various camera angles that may not exist in the input video, as well as synthesizing fine details such as cloth folds and facial appearance. Our method optimizes for a volumetric representation of the person in a canonical T-pose, in concert with a motion field that maps the estimated canonical representation to every frame of the video via backward warps. The motion field is decomposed into skeletal rigid and non-rigid motions, produced by deep networks. We show significant performance improvements over prior work, and compelling examples of free-viewpoint renderings from monocular video of moving humans in challenging uncontrolled capture scenarios.