论文标题

端到端的自主驾驶感知和连续的潜在表示学习

End-to-end Autonomous Driving Perception with Sequential Latent Representation Learning

论文作者

Chen, Jianyu, Xu, Zhuo, Tomizuka, Masayoshi

论文摘要

当前的自主驾驶系统由感知系统和决策系统组成。他们俩都分为多个由许多人类启发式方法构建的子系统。端到端的方法可能会清理系统并避免人工工程的巨大努力,并通过增加数据和计算资源获得更好的性能。与决策系统相比,感知系统更适合在端到端框架中设计,因为它不需要在线驾驶探索。在本文中,我们提出了一种新颖的端到端方法,用于自主驾驶感知。引入了潜在空间,以捕获所有相关功能,可用于感知,这是通过连续的潜在表示学习来学习的。博学的端到端感知模型能够通过最低限度的人类工程工作来解决检测,跟踪,本地化和映射问题,而无需在线存储任何地图。在现实的城市驾驶模拟器中评估了所提出的方法,相机图像和LiDar Point Cloud作为传感器输入。这项工作的代码和视频可在我们的GitHub Repo和Project网站上找到。

Current autonomous driving systems are composed of a perception system and a decision system. Both of them are divided into multiple subsystems built up with lots of human heuristics. An end-to-end approach might clean up the system and avoid huge efforts of human engineering, as well as obtain better performance with increasing data and computation resources. Compared to the decision system, the perception system is more suitable to be designed in an end-to-end framework, since it does not require online driving exploration. In this paper, we propose a novel end-to-end approach for autonomous driving perception. A latent space is introduced to capture all relevant features useful for perception, which is learned through sequential latent representation learning. The learned end-to-end perception model is able to solve the detection, tracking, localization and mapping problems altogether with only minimum human engineering efforts and without storing any maps online. The proposed method is evaluated in a realistic urban driving simulator, with both camera image and lidar point cloud as sensor inputs. The codes and videos of this work are available at our github repo and project website.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源