论文标题

限制了几乎没有阶级学习的学习

Constrained Few-shot Class-incremental Learning

论文作者

Hersche, Michael, Karunaratne, Geethan, Cherubini, Giovanni, Benini, Luca, Sebastian, Abu, Rahimi, Abbas

论文摘要

从新的数据中不断学习新课程,而不会忘记以前对旧课程的知识是一个非常具有挑战性的研究问题。此外,这种学习必须尊重某些记忆和计算约束,例如(i)培训样本仅限于每个类别的几个,(ii)学习新颖类的计算成本保持恒定,并且(iii)模型的记忆足迹在最多线性地增长,观察到的类数量。为了满足上述约束,我们提出了C-FSCIL,该C-FSCIL在架构上由冷冻的元学习特征提取器组成,可训练的固定尺寸完全连接的层以及一个动态增长的存储器,该内存存储了与遇到类的数量相同的向量。 C-FSCIL提供了三种更新模式,可在学习新颖课程的准确性和计算记忆成本之间进行权衡。 C-FSCIL利用高维嵌入,该嵌入允许不断地表达比矢量空间中的固定维度更大的类别,并且具有最小的干扰。通过通过新颖的损失函数将准矢量表示的质量互相对齐,从而进一步提高了它们。 CIFAR100,Miniimagenet和Omniglot数据集的实验表明,C-FSCIL的表现优于基准的精度和压缩。通过学习423个新颖的课程,在1200个基础类别的基础上,精度下降了1.6%,它还可以扩展到在此几次设置中尝试过的最大问题大小。我们的代码可在https://github.com/ibm/ccontrated-fscil上找到。

Continually learning new classes from fresh data without forgetting previous knowledge of old classes is a very challenging research problem. Moreover, it is imperative that such learning must respect certain memory and computational constraints such as (i) training samples are limited to only a few per class, (ii) the computational cost of learning a novel class remains constant, and (iii) the memory footprint of the model grows at most linearly with the number of classes observed. To meet the above constraints, we propose C-FSCIL, which is architecturally composed of a frozen meta-learned feature extractor, a trainable fixed-size fully connected layer, and a rewritable dynamically growing memory that stores as many vectors as the number of encountered classes. C-FSCIL provides three update modes that offer a trade-off between accuracy and compute-memory cost of learning novel classes. C-FSCIL exploits hyperdimensional embedding that allows to continually express many more classes than the fixed dimensions in the vector space, with minimal interference. The quality of class vector representations is further improved by aligning them quasi-orthogonally to each other by means of novel loss functions. Experiments on the CIFAR100, miniImageNet, and Omniglot datasets show that C-FSCIL outperforms the baselines with remarkable accuracy and compression. It also scales up to the largest problem size ever tried in this few-shot setting by learning 423 novel classes on top of 1200 base classes with less than 1.6% accuracy drop. Our code is available at https://github.com/IBM/constrained-FSCIL.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源