论文标题
用属性探测很少的概括
Probing Few-Shot Generalization with Attributes
论文作者
论文摘要
尽管在深度学习方面取得了令人印象深刻的进步,但越来越多的训练分配是一个重要的开放挑战。在这项工作中,我们考虑了很少的射击分类,并旨在阐明什么使某些新颖的课程比其他类型更容易学习,以及哪些类型的学术表现能够更好地推广。为此,我们根据属性(形成概念的简单构建块)定义了一个新的范式,以量化不同概念的相关性程度。我们的经验分析表明,监督学习的概括性范围很少,但自我监督预处理和监督的填充的结合会导致更强的概括。通过使用属性空间的随机分裂进行对照实验,进一步研究了自我监管预处理和监督者的好处,我们发现测试属性的可预测性提供了对模型概括能力的信息估计。
Despite impressive progress in deep learning, generalizing far beyond the training distribution is an important open challenge. In this work, we consider few-shot classification, and aim to shed light on what makes some novel classes easier to learn than others, and what types of learned representations generalize better. To this end, we define a new paradigm in terms of attributes -- simple building blocks of which concepts are formed -- as a means of quantifying the degree of relatedness of different concepts. Our empirical analysis reveals that supervised learning generalizes poorly to new attributes, but a combination of self-supervised pretraining with supervised finetuning leads to stronger generalization. The benefit of self-supervised pretraining and supervised finetuning is further investigated through controlled experiments using random splits of the attribute space, and we find that predictability of test attributes provides an informative estimate of a model's generalization ability.