论文标题
Wasserstein距离和BaryCenter的拓扑持续学习
Topological Continual Learning with Wasserstein Distance and Barycenter
论文作者
论文摘要
在神经网络中的持续学习遭受了一种称为灾难性遗忘的现象,其中一个网络很快忘记了以前的任务中学到的东西。但是,人脑能够不断学习新任务并在一生中积累知识。神经科学的发现表明,人脑的持续学习成功可能与其模块化结构和记忆巩固机制有关。在本文中,我们提出了一种新型的拓扑正则化,该拓扑正则化在训练期间使用持续的同源性和最佳运输的原则理论来惩罚神经网络中的周期结构。惩罚鼓励网络在培训期间学习模块化结构。该惩罚基于Wasserstein距离的封闭形式表达式和BaryCenter的拓扑特征,用于网络的1骨骨骼表示。我们的拓扑连续学习方法将拟议的正则化和微小的情节记忆结合在一起,以减轻遗忘。我们证明我们的方法在多个图像分类数据集的浅网和深网架构中都是有效的。
Continual learning in neural networks suffers from a phenomenon called catastrophic forgetting, in which a network quickly forgets what was learned in a previous task. The human brain, however, is able to continually learn new tasks and accumulate knowledge throughout life. Neuroscience findings suggest that continual learning success in the human brain is potentially associated with its modular structure and memory consolidation mechanisms. In this paper we propose a novel topological regularization that penalizes cycle structure in a neural network during training using principled theory from persistent homology and optimal transport. The penalty encourages the network to learn modular structure during training. The penalization is based on the closed-form expressions of the Wasserstein distance and barycenter for the topological features of a 1-skeleton representation for the network. Our topological continual learning method combines the proposed regularization with a tiny episodic memory to mitigate forgetting. We demonstrate that our method is effective in both shallow and deep network architectures for multiple image classification datasets.