论文标题
在野外恢复和模拟行人
Recovering and Simulating Pedestrians in the Wild
论文作者
论文摘要
传感器仿真是测试自动驾驶汽车性能和数据增强以更好训练感知系统的关键组成部分。典型的方法依靠艺术家来创建3D资产及其动画来生成新的场景。但是,这不是扩展。相比之下,我们建议通过驾驶的自动驾驶汽车从野外捕获的传感器读数中恢复行人的形状和运动。为了实现这一目标,我们将问题最小化为深度结构化模型,该模型利用了人类的形状先验,与从图像中提取的2D姿势进行了重新注射的一致性,以及鼓励重建网格的射线掌握式射击者,以同意激光雷达的读数。重要的是,我们不需要任何地面3D扫描或3D姿势注释。然后,我们通过执行运动重新定位,将重建的行人资产库纳入现实的激光雷达模拟系统中,并证明模拟的LIDAR数据可用于显着减少视觉感知任务所需的带注释的现实世界数据的量。
Sensor simulation is a key component for testing the performance of self-driving vehicles and for data augmentation to better train perception systems. Typical approaches rely on artists to create both 3D assets and their animations to generate a new scenario. This, however, does not scale. In contrast, we propose to recover the shape and motion of pedestrians from sensor readings captured in the wild by a self-driving car driving around. Towards this goal, we formulate the problem as energy minimization in a deep structured model that exploits human shape priors, reprojection consistency with 2D poses extracted from images, and a ray-caster that encourages the reconstructed mesh to agree with the LiDAR readings. Importantly, we do not require any ground-truth 3D scans or 3D pose annotations. We then incorporate the reconstructed pedestrian assets bank in a realistic LiDAR simulation system by performing motion retargeting, and show that the simulated LiDAR data can be used to significantly reduce the amount of annotated real-world data required for visual perception tasks.