论文标题

信任:使用知识蒸馏的值得信赖的积极学习

TrustAL: Trustworthy Active Learning using Knowledge Distillation

论文作者

Kwak, Beong-woo, Kim, Youngwook, Kim, Yu Jin, Hwang, Seung-won, Yeo, Jinyoung

论文摘要

主动学习可以定义为数据标记,模型培训和数据采集的迭代,直到获得足够的标签为止。数据获取的传统观点是,通过迭代,人类标签和模型的知识被隐式蒸馏,以单调地提高准确性和标签一致性。在这个假设下,最近训练的模型是当前标记数据的良好替代品,根据不确定性/多样性,请求数据采集。我们的贡献是揭穿这个神话,并提出了一个新的蒸馏目标。首先,我们发现了示例忘记,这表明跨迭代中学习的知识丧失。其次,因此,最后一个模型不再是最好的老师 - 为了减轻这种被遗忘的知识,我们通过提出的“一致性”概念选择了它的前身模型之一。我们表明,这种新颖的蒸馏在以下三个方面是独特的。首先,一致性确保避免忘记标签。其次,一致性提高了标记数据的不确定性/多样性。最后,一致性赎回了人类注释者产生的有缺陷的标签。

Active learning can be defined as iterations of data labeling, model training, and data acquisition, until sufficient labels are acquired. A traditional view of data acquisition is that, through iterations, knowledge from human labels and models is implicitly distilled to monotonically increase the accuracy and label consistency. Under this assumption, the most recently trained model is a good surrogate for the current labeled data, from which data acquisition is requested based on uncertainty/diversity. Our contribution is debunking this myth and proposing a new objective for distillation. First, we found example forgetting, which indicates the loss of knowledge learned across iterations. Second, for this reason, the last model is no longer the best teacher -- For mitigating such forgotten knowledge, we select one of its predecessor models as a teacher, by our proposed notion of "consistency". We show that this novel distillation is distinctive in the following three aspects; First, consistency ensures to avoid forgetting labels. Second, consistency improves both uncertainty/diversity of labeled data. Lastly, consistency redeems defective labels produced by human annotators.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源