论文标题
CoEdge:合作DNN推断与异质边缘设备上的自适应工作负载分区
CoEdge: Cooperative DNN Inference with Adaptive Workload Partitioning over Heterogeneous Edge Devices
论文作者
论文摘要
人工智能的最新进展推动了网络边缘的智能应用程序的增加,例如智能家居,智能工厂和智能城市。为了在资源受限的边缘设备上部署计算密集的深度神经网络(DNN),传统方法依赖于将工作负载卸载到远程云或在本地设备上优化计算。但是,云辅助方法遭受了不可靠和延迟的广阔区域网络的影响,并且本地计算方法受到约束计算能力的限制。为了进行高性能的边缘情报,合作执行机制提供了一种新的范式,最近引起了不断增长的研究兴趣。在本文中,我们提出了COEDGE,这是一个分布式DNN计算系统,该系统协调了在异质边缘设备上的合作DNN推断。 CoEdge利用了边缘的可用计算和通信资源,并动态地对DNN推理工作负载适应设备的计算功能和网络条件。基于现实原型的实验评估表明,Coedge在近距离推理潜伏期中节省能量方面的表现如何,可在四种广泛添加的CNN模型中实现高达25.5%〜66.9%的能量。
Recent advances in artificial intelligence have driven increasing intelligent applications at the network edge, such as smart home, smart factory, and smart city. To deploy computationally intensive Deep Neural Networks (DNNs) on resource-constrained edge devices, traditional approaches have relied on either offloading workload to the remote cloud or optimizing computation at the end device locally. However, the cloud-assisted approaches suffer from the unreliable and delay-significant wide-area network, and the local computing approaches are limited by the constrained computing capability. Towards high-performance edge intelligence, the cooperative execution mechanism offers a new paradigm, which has attracted growing research interest recently. In this paper, we propose CoEdge, a distributed DNN computing system that orchestrates cooperative DNN inference over heterogeneous edge devices. CoEdge utilizes available computation and communication resources at the edge and dynamically partitions the DNN inference workload adaptive to devices' computing capabilities and network conditions. Experimental evaluations based on a realistic prototype show that CoEdge outperforms status-quo approaches in saving energy with close inference latency, achieving up to 25.5%~66.9% energy reduction for four widely-adopted CNN models.