论文标题
自动转移:学会路由可透式陈述
Auto-Transfer: Learning to Route Transferrable Representations
论文作者
论文摘要
近来,异质源和目标网络和任务之间的知识转移引起了很多关注,因为在许多应用中很难获得大量质量标记的数据。现有方法通常会限制目标深神经网络(DNN)特征表示与源DNNS特征表示的接近,这可能是限制的。在本文中,我们提出了一种新颖的对抗性多臂强盗方法,该方法自动学习将源代表路由到适当的目标表示形式,然后以有意义的方式组合来产生准确的目标模型。与四个基准(目标)图像数据集Cub200,Stanford Dogs,Mit67和Stanford40上的最先进的知识转移方法相比,我们看到5 \%的精度提高了5%的精度。我们通过与(最接近)竞争对手相比,我们的目标网络在不同层上着重于我们的目标网络的个别示例来定性地分析转移方案的优点。我们还观察到,对于较小的目标数据集,我们对其他方法的改进更高,使其成为可能受益于转移学习的小型数据应用程序的有效工具。
Knowledge transfer between heterogeneous source and target networks and tasks has received a lot of attention in recent times as large amounts of quality labeled data can be difficult to obtain in many applications. Existing approaches typically constrain the target deep neural network (DNN) feature representations to be close to the source DNNs feature representations, which can be limiting. We, in this paper, propose a novel adversarial multi-armed bandit approach that automatically learns to route source representations to appropriate target representations following which they are combined in meaningful ways to produce accurate target models. We see upwards of 5\% accuracy improvements compared with the state-of-the-art knowledge transfer methods on four benchmark (target) image datasets CUB200, Stanford Dogs, MIT67, and Stanford40 where the source dataset is ImageNet. We qualitatively analyze the goodness of our transfer scheme by showing individual examples of the important features focused on by our target network at different layers compared with the (closest) competitors. We also observe that our improvement over other methods is higher for smaller target datasets making it an effective tool for small data applications that may benefit from transfer learning.