论文标题
为自我监督点云学习提高采样自动编码器
Upsampling Autoencoder for Self-Supervised Point Cloud Learning
论文作者
论文摘要
在计算机辅助设计(CAD)社区中,点云数据普遍应用于反向工程,其中点云分析起着重要作用。尽管已经提出了大量的监督学习方法来处理无序的点云并证明了它们的显着成功,但其性能和适用性仅限于昂贵的数据注释。在这项工作中,我们提出了一个新颖的自我监督预审计模型,用于没有人类注释的点云学习,该模型仅依赖于提高采样操作,以有效地对点云进行特征学习。我们方法的关键前提是,UPSMPLING操作鼓励网络捕获点云的高级语义信息和低级几何信息,因此,分类和细分等下游任务将从预训练的模型中受益。具体而言,我们的方法首先以低比例为12.5%的输入点云进行随机子采样。然后,我们将它们送入编码器架构中,在该体系结构中设计了一个编码器仅在子采样点上操作,并采用了UPSMPLINGINPING解码器来基于学习的功能来重建原始点云。最后,我们设计了一种新型的关节损耗函数,该功能强制实施与原始点云相似的上采样点,并均匀分布在下面的形状表面上。通过采用预先训练的编码权重作为下游任务的模型的初始化,我们发现我们的阿联酋在形状分类,部分分割和点云上提高任务中都优于先前最先进的方法。接受后将公开提供代码。
In computer-aided design (CAD) community, the point cloud data is pervasively applied in reverse engineering, where the point cloud analysis plays an important role. While a large number of supervised learning methods have been proposed to handle the unordered point clouds and demonstrated their remarkable success, their performance and applicability are limited to the costly data annotation. In this work, we propose a novel self-supervised pretraining model for point cloud learning without human annotations, which relies solely on upsampling operation to perform feature learning of point cloud in an effective manner. The key premise of our approach is that upsampling operation encourages the network to capture both high-level semantic information and low-level geometric information of the point cloud, thus the downstream tasks such as classification and segmentation will benefit from the pre-trained model. Specifically, our method first conducts the random subsampling from the input point cloud at a low proportion e.g., 12.5%. Then, we feed them into an encoder-decoder architecture, where an encoder is devised to operate only on the subsampled points, along with a upsampling decoder is adopted to reconstruct the original point cloud based on the learned features. Finally, we design a novel joint loss function which enforces the upsampled points to be similar with the original point cloud and uniformly distributed on the underlying shape surface. By adopting the pre-trained encoder weights as initialisation of models for downstream tasks, we find that our UAE outperforms previous state-of-the-art methods in shape classification, part segmentation and point cloud upsampling tasks. Code will be made publicly available upon acceptance.