论文标题
Metagater:通过联合元学习快速学习条件通道门控网络
MetaGater: Fast Learning of Conditional Channel Gated Networks via Federated Meta-Learning
论文作者
论文摘要
尽管深度学习在许多AI应用中取得了惊人的成功,但其巨大的模型大小和密集的计算要求对资源有限的节点的部署构成了巨大的挑战。最近,人们对计算效率学习方法(例如量化,修剪和通道门控)的兴趣越来越大。但是,大多数现有技术无法快速适应不同的任务。在这项工作中,我们提倡一种整体方法,用于共同训练骨干网络和通道门控,该通道门控能够动态选择过滤器的子集,以便给定数据输入,以进行更有效的本地计算。特别是,我们开发了一种联合的元学习方法,通过利用在不同节点上学习任务的模型相似性,共同学习骨干网络和门控模块的良好元定位。通过这种方式,学识渊博的元门控模块有效地捕获了良好的元靠背网络的重要过滤器,基于该网络,可以通过使用该任务的新示例在两阶段的过程中从元启动过程中快速调整特定于任务的条件通道网络网络,即通过一步梯度下降,通过一步梯度下降。在轻度条件下建立了拟议的联合元学习算法的收敛性。与相关工作相比,实验结果证实了我们方法的有效性。
While deep learning has achieved phenomenal successes in many AI applications, its enormous model size and intensive computation requirements pose a formidable challenge to the deployment in resource-limited nodes. There has recently been an increasing interest in computationally-efficient learning methods, e.g., quantization, pruning and channel gating. However, most existing techniques cannot adapt to different tasks quickly. In this work, we advocate a holistic approach to jointly train the backbone network and the channel gating which enables dynamical selection of a subset of filters for more efficient local computation given the data input. Particularly, we develop a federated meta-learning approach to jointly learn good meta-initializations for both backbone networks and gating modules, by making use of the model similarity across learning tasks on different nodes. In this way, the learnt meta-gating module effectively captures the important filters of a good meta-backbone network, based on which a task-specific conditional channel gated network can be quickly adapted, i.e., through one-step gradient descent, from the meta-initializations in a two-stage procedure using new samples of that task. The convergence of the proposed federated meta-learning algorithm is established under mild conditions. Experimental results corroborate the effectiveness of our method in comparison to related work.