论文标题

边缘计算中自适应深神经网络的案例

A Case For Adaptive Deep Neural Networks in Edge Computing

论文作者

McNamee, Francis, Dustadar, Schahram, Kilpatrick, Peter, Shi, Weisong, Spence, Ivor, Varghese, Blesson

论文摘要

在将对隐私敏感性和关键性能的应用程序的原始数据传输到云数据中心之前,Edge Computing在靠近数据源的额外计算基础架构中更接近数据源。深神经网络(DNN)是一类应用程序,据报道,边缘和云之间的协作计算受益。将DNN的分区进行了分区,以使DNN的特定层部署到边缘和云上以满足性能和隐私目标。但是,对:(a)有限的了解:(a)是否以及如何不断发展的操作条件(在边缘和边缘和云之间的数据传输速率降低)是否会影响已经部署已部署的DNN的性能,以及(b)是否需要新的分区配置才能最大程度地提高性能。适应不断变化的操作条件的DNN称为“自适应DNN”。本文通过考虑三个问题来研究边缘计算中自适应DNN的情况:(i)DNNS对操作条件敏感? (ii)DNN对操作条件的敏感程度如何? (iii)个人或操作条件的组合是否同样影响DNN? (iv)DNN分区是否对云/边缘上的硬件体系结构敏感?该探索是在8种预训练的DNN模型的背景下进行的,所提供的结果来自分析近800万个数据点。结果表明,网络条件比CPU或与内存相关的操作条件更大。注意到重新分配在许多情况下提供性能增长,但是与基础硬件架构相关的特定趋势并未指出。尽管如此,对自适应DNN的需求得到了证实。

Edge computing offers an additional layer of compute infrastructure closer to the data source before raw data from privacy-sensitive and performance-critical applications is transferred to a cloud data center. Deep Neural Networks (DNNs) are one class of applications that are reported to benefit from collaboratively computing between the edge and the cloud. A DNN is partitioned such that specific layers of the DNN are deployed onto the edge and the cloud to meet performance and privacy objectives. However, there is limited understanding of: (a) whether and how evolving operational conditions (increased CPU and memory utilization at the edge or reduced data transfer rates between the edge and the cloud) affect the performance of already deployed DNNs, and (b) whether a new partition configuration is required to maximize performance. A DNN that adapts to changing operational conditions is referred to as an 'adaptive DNN'. This paper investigates whether there is a case for adaptive DNNs in edge computing by considering three questions: (i) Are DNNs sensitive to operational conditions? (ii) How sensitive are DNNs to operational conditions? (iii) Do individual or a combination of operational conditions equally affect DNNs? (iv) Is DNN partitioning sensitive to hardware architectures on the cloud/edge? The exploration is carried out in the context of 8 pre-trained DNN models and the results presented are from analyzing nearly 8 million data points. The results highlight that network conditions affects DNN performance more than CPU or memory related operational conditions. Repartitioning is noted to provide a performance gain in a number of cases, but a specific trend was not noted in relation to its correlation to the underlying hardware architecture. Nonetheless, the need for adaptive DNNs is confirmed.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源