论文标题
与合成数据的点云中人类的3D分割
3D Segmentation of Humans in Point Clouds with Synthetic Data
论文作者
论文摘要
随着以人为本的机器人技术和AR/VR应用的兴起,在3D室内场景中分割人类变得越来越重要。为此,我们提出了联合3D人类语义分割,实例分割和多人体身体部位分割的任务。很少有作品试图直接在混乱的3D场景中分割人类,这主要是由于缺乏人类与3D场景相互作用的注释培训数据。我们应对这一挑战,并提出一个框架,以生成合成人类与真实3D场景相互作用的培训数据。此外,我们提出了一种基于变压器的新型模型Human3D,该模型是第一个以统一的方式分割多个人类实例及其身体零件的端到端模型。我们的合成数据生成框架的主要优点是它具有高度准确的地面真理的能力。我们的实验表明,对合成数据的预训练可改善各种3D人类分割任务的性能。最后,我们证明了人类3D甚至超过特定于任务的最先进的3D分割方法。
Segmenting humans in 3D indoor scenes has become increasingly important with the rise of human-centered robotics and AR/VR applications. To this end, we propose the task of joint 3D human semantic segmentation, instance segmentation and multi-human body-part segmentation. Few works have attempted to directly segment humans in cluttered 3D scenes, which is largely due to the lack of annotated training data of humans interacting with 3D scenes. We address this challenge and propose a framework for generating training data of synthetic humans interacting with real 3D scenes. Furthermore, we propose a novel transformer-based model, Human3D, which is the first end-to-end model for segmenting multiple human instances and their body-parts in a unified manner. The key advantage of our synthetic data generation framework is its ability to generate diverse and realistic human-scene interactions, with highly accurate ground truth. Our experiments show that pre-training on synthetic data improves performance on a wide variety of 3D human segmentation tasks. Finally, we demonstrate that Human3D outperforms even task-specific state-of-the-art 3D segmentation methods.