论文标题
VNT网络:旋转不变矢量神经元变压器
VNT-Net: Rotational Invariant Vector Neuron Transformers
论文作者
论文摘要
在机器学习中学习3D点集是机器学习中的一个重要且具有挑战性的问题。通过旋转不变体系结构,3D点云神经网络免除了需要典型的全球姿势和所有可能旋转的详尽数据增强。在这项工作中,我们通过将最近引入的向量神经元与自发层相结合以构建点云矢量神经元变压器网络(VNT-NET)来引入旋转不变的神经网络。向量神经元以其在代表SO(3)作用中的简单性和多功能性而闻名,从而将其纳入了常见的神经操作中。同样,变压器体系结构已获得流行,最近通过直接应用图像贴片的序列并实现了出色的性能和收敛性,以表明图像成功。为了使两者受益,我们通过主要展示如何适应多头注意层以符合向量神经元操作来结合这两种结构。通过这种适应性注意层变得如此(3),整个网络变为旋转不变。实验表明,我们的网络有效地以任意姿势处理3D点云对象。我们还表明,与相关的最新方法相比,我们的网络可以达到更高的准确性,并且由于常见分类和分割任务的较少数量的超参数,需要更少的培训。
Learning 3D point sets with rotational invariance is an important and challenging problem in machine learning. Through rotational invariant architectures, 3D point cloud neural networks are relieved from requiring a canonical global pose and from exhaustive data augmentation with all possible rotations. In this work, we introduce a rotational invariant neural network by combining recently introduced vector neurons with self-attention layers to build a point cloud vector neuron transformer network (VNT-Net). Vector neurons are known for their simplicity and versatility in representing SO(3) actions and are thereby incorporated in common neural operations. Similarly, Transformer architectures have gained popularity and recently were shown successful for images by applying directly on sequences of image patches and achieving superior performance and convergence. In order to benefit from both worlds, we combine the two structures by mainly showing how to adapt the multi-headed attention layers to comply with vector neurons operations. Through this adaptation attention layers become SO(3) and the overall network becomes rotational invariant. Experiments demonstrate that our network efficiently handles 3D point cloud objects in arbitrary poses. We also show that our network achieves higher accuracy when compared to related state-of-the-art methods and requires less training due to a smaller number of hyperparameters in common classification and segmentation tasks.