论文标题

pai3d:绘画3D对象检测的自适应实例prior

PAI3D: Painting Adaptive Instance-Prior for 3D Object Detection

论文作者

Liu, Hao, Xu, Zhuoran, Wang, Dan, Zhang, Baofeng, Wang, Guan, Dong, Bo, Wen, Xin, Xu, Xinyu

论文摘要

3D对象检测是自主驾驶中的关键任务。最近,基于多模式融合的3D对象检测方法结合了LIDAR和相机的互补优势,对单模式方法的性能进行了巨大的改进。但是,到目前为止,尚未尝试利用实例级上下文图像语义来指导3D对象检测。在本文中,我们为3D对象检测(PAI3D)提出了一个简单有效的绘画自适应实例 - 绘画,以灵活地将实例级图像语义与点云特征融合在一起。 PAI3D是一个多模式的顺序实例级融合框架。它首先从图像中提取实例级别的语义信息,然后使用提取的信息,包括对象的分类标签,点对点成员资格和对象位置,用于在随后的3D检测网络中增加每个激光雷达点,以指导和提高检测性能。 PAI3D在Nuscenes数据集上胜过最先进的最先进,在MAP中达到71.4,在测试拆分中在NDS中获得74.2。我们的全面实验表明,实例级图像语义对性能增长贡献最大,PAI3D与任何良好质量实例分割模型和任何现代点云3D编码器都很好地配合在一起,使其成为在自动驾驶汽车上部署的有力候选者。

3D object detection is a critical task in autonomous driving. Recently multi-modal fusion-based 3D object detection methods, which combine the complementary advantages of LiDAR and camera, have shown great performance improvements over mono-modal methods. However, so far, no methods have attempted to utilize the instance-level contextual image semantics to guide the 3D object detection. In this paper, we propose a simple and effective Painting Adaptive Instance-prior for 3D object detection (PAI3D) to fuse instance-level image semantics flexibly with point cloud features. PAI3D is a multi-modal sequential instance-level fusion framework. It first extracts instance-level semantic information from images, the extracted information, including objects categorical label, point-to-object membership and object position, are then used to augment each LiDAR point in the subsequent 3D detection network to guide and improve detection performance. PAI3D outperforms the state-of-the-art with a large margin on the nuScenes dataset, achieving 71.4 in mAP and 74.2 in NDS on the test split. Our comprehensive experiments show that instance-level image semantics contribute the most to the performance gain, and PAI3D works well with any good-quality instance segmentation models and any modern point cloud 3D encoders, making it a strong candidate for deployment on autonomous vehicles.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源