论文标题
重新考虑深度学习中可转移性的两个共识
Rethinking Two Consensuses of the Transferability in Deep Learning
论文作者
论文摘要
深度转移学习(DTL)已经为使深度神经网络(DNN)像人类一样有效地重复历史经验而形成了长期的追求。此功能被命名为知识传递性。用于DTL的常用范例是首先学习通用知识(预训练),然后重复使用(微调)来执行特定目标任务。预先训练的DNN的可传递性有两种共识:(1)训练和下游数据之间的较大域间隙带来较低的可传递性; (2)可传递性逐渐从下层(接近输入)降低到较高的层(接近输出)。但是,这些共识基本上是根据基于自然图像的实验得出的,这限制了它们的应用范围。这项工作旨在通过提出一种测量预训练的DNN参数可转移性的方法来研究和补充它们。我们对十二种不同图像分类数据集的实验得出了与以前的共识相似的结论。更重要的是,提出了两个新的发现,即(1)除了域间隙,更大的数据量和下游目标任务的巨大数据集多样性也禁止可转移性; (2)尽管下层学习基本图像特征,但由于域的灵敏度,它们通常不是最可转移的层。
Deep transfer learning (DTL) has formed a long-term quest toward enabling deep neural networks (DNNs) to reuse historical experiences as efficiently as humans. This ability is named knowledge transferability. A commonly used paradigm for DTL is firstly learning general knowledge (pre-training) and then reusing (fine-tuning) them for a specific target task. There are two consensuses of transferability of pre-trained DNNs: (1) a larger domain gap between pre-training and downstream data brings lower transferability; (2) the transferability gradually decreases from lower layers (near input) to higher layers (near output). However, these consensuses were basically drawn from the experiments based on natural images, which limits their scope of application. This work aims to study and complement them from a broader perspective by proposing a method to measure the transferability of pre-trained DNN parameters. Our experiments on twelve diverse image classification datasets get similar conclusions to the previous consensuses. More importantly, two new findings are presented, i.e., (1) in addition to the domain gap, a larger data amount and huge dataset diversity of downstream target task also prohibit the transferability; (2) although the lower layers learn basic image features, they are usually not the most transferable layers due to their domain sensitivity.