论文标题
SelfVoxelo:基于体素的深神经网络的自我监督的底部射仪
SelfVoxeLO: Self-supervised LiDAR Odometry with Voxel-based Deep Neural Networks
论文作者
论文摘要
最近基于学习的LiDAR进程方法证明了它们的竞争力。但是,大多数方法仍然面临两个重大挑战:1)LIDAR数据的2D投影表示无法从点云有效地编码3D结构; 2)大量标记数据的训练需求限制了这些方法的应用范围。在本文中,我们提出了一种称为“ selfvoxelo”的自我监督的激光射击方法,以解决这两个困难。具体而言,我们提出了一个3D卷积网络,以直接处理RAW LIDAR数据,该数据提取了更好地编码3D几何模式的功能。为了适合我们的网络,我们设计了几种利用LiDar Point Clouds固有特性的新型损失功能。此外,不确定性感知的机制已纳入损失功能,以减轻移动对象/噪声的干扰。我们在两个大规模数据集(即Kitti和Apollo-Southbay)上评估了方法的性能。就KITTI数据集上的翻译/旋转错误而言,我们的方法优于最先进的无监督方法,在Apollo-Southbay数据集上的性能很好。通过包括更多未标记的培训数据,我们的方法可以进一步提高与监督方法相当的性能。
Recent learning-based LiDAR odometry methods have demonstrated their competitiveness. However, most methods still face two substantial challenges: 1) the 2D projection representation of LiDAR data cannot effectively encode 3D structures from the point clouds; 2) the needs for a large amount of labeled data for training limit the application scope of these methods. In this paper, we propose a self-supervised LiDAR odometry method, dubbed SelfVoxeLO, to tackle these two difficulties. Specifically, we propose a 3D convolution network to process the raw LiDAR data directly, which extracts features that better encode the 3D geometric patterns. To suit our network to self-supervised learning, we design several novel loss functions that utilize the inherent properties of LiDAR point clouds. Moreover, an uncertainty-aware mechanism is incorporated in the loss functions to alleviate the interference of moving objects/noises. We evaluate our method's performances on two large-scale datasets, i.e., KITTI and Apollo-SouthBay. Our method outperforms state-of-the-art unsupervised methods by 27%/32% in terms of translational/rotational errors on the KITTI dataset and also performs well on the Apollo-SouthBay dataset. By including more unlabelled training data, our method can further improve performance comparable to the supervised methods.