论文标题
基于Koopman操作员理论的训练有素的神经网络的信用分配
Credit Assignment for Trained Neural Networks Based on Koopman Operator Theory
论文作者
论文摘要
神经网络的信用分配问题是指评估每个网络组件对最终产出的信用。对于未经训练的神经网络,应对它的方法为训练阶段的参数更新和模型革命做出了巨大贡献。训练有素的神经网络上的这个问题引起了极大的关注,但是,它在神经网络贴片,规范和验证中起着越来越重要的作用。基于Koopman运营商理论,本文介绍了有关处理训练有素的神经网络的信用分配问题的线性动态的替代观点。关于神经网络作为子动力学系列的组成,我们利用踩踏嵌入来捕获每个组件的快照,并尽可能准确地表征已建立的映射。为了避免嵌入过程中遇到的维差异问题,辅助线性层的组成和分解(称为最小线性尺寸对齐)是通过严格的正式保证精心设计的。之后,每个组件都由Koopman运算符近似,我们得出了Jacobian矩阵及其相应的决定因素,类似于向后传播。然后,我们可以为每个网络组件的信用分配定义具有代数解释性的度量。此外,在典型的神经网络上进行的实验证明了该方法的有效性。
Credit assignment problem of neural networks refers to evaluating the credit of each network component to the final outputs. For an untrained neural network, approaches to tackling it have made great contributions to parameter update and model revolution during the training phase. This problem on trained neural networks receives rare attention, nevertheless, it plays an increasingly important role in neural network patch, specification and verification. Based on Koopman operator theory, this paper presents an alternative perspective of linear dynamics on dealing with the credit assignment problem for trained neural networks. Regarding a neural network as the composition of sub-dynamics series, we utilize step-delay embedding to capture snapshots of each component, characterizing the established mapping as exactly as possible. To circumvent the dimension-difference problem encountered during the embedding, a composition and decomposition of an auxiliary linear layer, termed minimal linear dimension alignment, is carefully designed with rigorous formal guarantee. Afterwards, each component is approximated by a Koopman operator and we derive the Jacobian matrix and its corresponding determinant, similar to backward propagation. Then, we can define a metric with algebraic interpretability for the credit assignment of each network component. Moreover, experiments conducted on typical neural networks demonstrate the effectiveness of the proposed method.