论文标题

ST-P3:通过时空特征学习基于端到端的基于端视力的自主驾驶

ST-P3: End-to-end Vision-based Autonomous Driving via Spatial-Temporal Feature Learning

论文作者

Hu, Shengchao, Chen, Li, Wu, Penghao, Li, Hongyang, Yan, Junchi, Tao, Dacheng

论文摘要

许多现有的自主驾驶范式涉及多个任务的多个阶段离散管道。为了更好地预测控制信号并增强用户安全性,希望从联合时空特征学习中受益的端到端方法是可取的。尽管基于激光雷达的输入或隐式设计有一些开创性的作品,但在本文中,我们在基于可解释的基于视觉的设置中提出了问题。特别是,我们提出了一种空间特征学习方案,以同时同时进行感知,预测和计划任务的一组代表性的特征,称为ST-P3。具体而言,提出了一种以自我为中心的累积技术来保留3D空间中的几何信息。设计了双重途径建模,以考虑将来的前进预测。引入了一个基于时间的改进单元,以弥补识别基于视觉的计划的元素。据我们所知,我们是第一个系统地研究基于端到视觉的自主驾驶系统的每个部分的人。我们在开环Nuscenes数据集和闭环CARLA模拟上对以前的最先进的方法进行了基准测试。结果显示了我们方法的有效性。源代码,模型和协议详细信息可在https://github.com/openperceptionx/st-p3上公开获得。

Many existing autonomous driving paradigms involve a multi-stage discrete pipeline of tasks. To better predict the control signals and enhance user safety, an end-to-end approach that benefits from joint spatial-temporal feature learning is desirable. While there are some pioneering works on LiDAR-based input or implicit design, in this paper we formulate the problem in an interpretable vision-based setting. In particular, we propose a spatial-temporal feature learning scheme towards a set of more representative features for perception, prediction and planning tasks simultaneously, which is called ST-P3. Specifically, an egocentric-aligned accumulation technique is proposed to preserve geometry information in 3D space before the bird's eye view transformation for perception; a dual pathway modeling is devised to take past motion variations into account for future prediction; a temporal-based refinement unit is introduced to compensate for recognizing vision-based elements for planning. To the best of our knowledge, we are the first to systematically investigate each part of an interpretable end-to-end vision-based autonomous driving system. We benchmark our approach against previous state-of-the-arts on both open-loop nuScenes dataset as well as closed-loop CARLA simulation. The results show the effectiveness of our method. Source code, model and protocol details are made publicly available at https://github.com/OpenPerceptionX/ST-P3.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源