论文标题
Depthnet:自动驾驶汽车的实时LIDAR点云深度完成
DepthNet: Real-Time LiDAR Point Cloud Depth Completion for Autonomous Vehicles
论文作者
论文摘要
自动驾驶汽车在很大程度上依赖于诸如相机和激光镜等传感器,这些传感器提供了有关其周围环境的实时信息,以完成感知,计划和控制的任务。通常,由于扫描线数量有限,激光雷达只能提供稀疏点云。通过采用深度完成,可以通过为每个相机像素分配一个相应的深度值来生成密集的深度图。但是,现有的深度完成卷积神经网络非常复杂,需要高端GPU进行处理,因此它们不适用于实时自主驾驶。在本文中,提出了一个轻巧的网络,以完成LIDAR点云深度完成任务。由于参数数量的惊人减少了96.2%,因此它仍然可以达到可比的性能(MAE增长了9.3%,但RMSE的3.9%差3.9%)比最先进的网络。对于实时嵌入式平台,将深度可分离技术应用于卷积和反卷积操作,参数数量进一步降低了7.3倍,只有RMSE和MAE的性能仅增加了百分比。此外,在基于PYNQ的FPGA平台上开发了用于深度完成的系统结构,该平台以每秒的速度11.1帧来实现HDL-64E激光雷达的实时处理。
Autonomous vehicles rely heavily on sensors such as camera and LiDAR, which provide real-time information about their surroundings for the tasks of perception, planning and control. Typically a LiDAR can only provide sparse point cloud owing to a limited number of scanning lines. By employing depth completion, a dense depth map can be generated by assigning each camera pixel a corresponding depth value. However, the existing depth completion convolutional neural networks are very complex that requires high-end GPUs for processing, and thus they are not applicable to real-time autonomous driving. In this paper, a light-weight network is proposed for the task of LiDAR point cloud depth completion. With an astonishing 96.2% reduction in the number of parameters, it still achieves comparable performance (9.3% better in MAE but 3.9% worse in RMSE) to the state-of-the-art network. For real-time embedded platforms, depthwise separable technique is applied to both convolution and deconvolution operations and the number of parameters decreases further by a factor of 7.3, with only a small percentage increase in RMSE and MAE performance. Moreover, a system-on-chip architecture for depth completion is developed on a PYNQ-based FPGA platform that achieves real-time processing for HDL-64E LiDAR at the speed 11.1 frame per second.