论文标题
随机神经领域的深度学习:通过神经切线内核进行数值实验
Deep Learning in Random Neural Fields: Numerical Experiments via Neural Tangent Kernel
论文作者
论文摘要
皮质中的生物神经网络形成神经场。该领域的神经元具有自己的接受场,并且两个神经元之间的连接权重是随机的,但是当它们在接收领域处于近距离时,它们高度相关。在本文中,我们研究了多层体系结构中的此类神经领域,以调查对领域的监督学习。我们从经验上将现场模型的性能与随机连接的深网的性能进行比较。基于神经切线核制度的关键思想进行了随机连接网络的行为,这是过度参数化网络的机器学习理论的最新发展;对于大多数随机连接的神经网络,这表明全球最小始终存在于其小社区中。我们从数字上表明,这一说法也适用于我们的神经领域。更详细地说,我们的模型具有两个结构:i)磁场中的每个神经元具有连续分布的接收场,ii)初始连接权重是随机的,但不是独立的,当神经元的位置在每个层的位置接近时具有相关性。我们表明,当输入模式被噪声干扰变形时,这种多层神经场比常规模型更强大。此外,其概括能力可以略优于常规模型。
A biological neural network in the cortex forms a neural field. Neurons in the field have their own receptive fields, and connection weights between two neurons are random but highly correlated when they are in close proximity in receptive fields. In this paper, we investigate such neural fields in a multilayer architecture to investigate the supervised learning of the fields. We empirically compare the performances of our field model with those of randomly connected deep networks. The behavior of a randomly connected network is investigated on the basis of the key idea of the neural tangent kernel regime, a recent development in the machine learning theory of over-parameterized networks; for most randomly connected neural networks, it is shown that global minima always exist in their small neighborhoods. We numerically show that this claim also holds for our neural fields. In more detail, our model has two structures: i) each neuron in a field has a continuously distributed receptive field, and ii) the initial connection weights are random but not independent, having correlations when the positions of neurons are close in each layer. We show that such a multilayer neural field is more robust than conventional models when input patterns are deformed by noise disturbances. Moreover, its generalization ability can be slightly superior to that of conventional models.