论文标题
基于固定的预训练特征提取器的贝叶斯模型的持续学习
Continual Learning with Bayesian Model based on a Fixed Pre-trained Feature Extractor
论文作者
论文摘要
深度学习表明了其在各种应用中的人类水平表现。但是,当前的深度学习模型的特征是在学习新课程时灾难性地忘记了旧知识。这尤其是在智能诊断系统中提出的挑战,在智能诊断系统中,最初只能培训有限数量的疾病的数据。在这种情况下,使用新疾病的数据更新智能系统将不可避免地降低其先前学习的疾病的性能。受到人类大脑中新知识的过程的启发,我们提出了一个贝叶斯生成模型,用于持续学习以固定的预训练的特征提取器为基础。在此模型中,每个旧类的知识可以由统计分布的集合(例如随着时间的流逝,使用高斯混合模型,并且自然而然地忘记了持续学习。与现有的课堂学习方法不同,所提出的方法对持续学习过程不敏感,并且可以很好地应用于数据收入学习方案。关于多个医学和自然图像分类任务的实验表明,所提出的方法的表现优于最先进的方法,这些方法甚至可以在不断学习新课程期间保留一些旧课程的图像。
Deep learning has shown its human-level performance in various applications. However, current deep learning models are characterised by catastrophic forgetting of old knowledge when learning new classes. This poses a challenge particularly in intelligent diagnosis systems where initially only training data of a limited number of diseases are available. In this case, updating the intelligent system with data of new diseases would inevitably downgrade its performance on previously learned diseases. Inspired by the process of learning new knowledge in human brains, we propose a Bayesian generative model for continual learning built on a fixed pre-trained feature extractor. In this model, knowledge of each old class can be compactly represented by a collection of statistical distributions, e.g. with Gaussian mixture models, and naturally kept from forgetting in continual learning over time. Unlike existing class-incremental learning methods, the proposed approach is not sensitive to the continual learning process and can be additionally well applied to the data-incremental learning scenario. Experiments on multiple medical and natural image classification tasks showed that the proposed approach outperforms state-of-the-art approaches which even keep some images of old classes during continual learning of new classes.