论文标题
AI税:AI数据中心应用程序的隐藏成本
AI Tax: The Hidden Cost of AI Data Center Applications
论文作者
论文摘要
人工智能和机器学习正在行业和学术界广泛采用。通过日益复杂的算法和模型,AI的应用和准确性的快速进步驱动了这一点。反过来,这刺激了专业硬件AI加速器的研究。鉴于进步的速度迅速,很容易忘记它们通常是在真空中开发和评估的,而无需考虑完整的应用环境。本文强调了对AI工作负载进行整体,端到端分析的必要性,并揭示了“ AI税”。我们在边缘数据中心部署并表征了面部识别。该应用程序是使用流行的开源基础架构和ML工具构建的以AI为中心的Edge Video Analytics应用程序。尽管使用了最先进的AI和ML算法,但该应用程序在很大程度上依赖于预处理代码。随着以AI为中心的应用程序受益于加速器承诺的加速度,我们发现它们对硬件和软件基础架构施加了压力:存储和网络带宽会随着AI加速度的增加而成为主要的瓶颈。通过专门针对AI应用程序,我们表明,可以设计出专用的边缘数据中心,该中心是为TCO加速AI的应力而设计的,而不是从均质服务器和基础架构中得出的一个。
Artificial intelligence and machine learning are experiencing widespread adoption in industry and academia. This has been driven by rapid advances in the applications and accuracy of AI through increasingly complex algorithms and models; this, in turn, has spurred research into specialized hardware AI accelerators. Given the rapid pace of advances, it is easy to forget that they are often developed and evaluated in a vacuum without considering the full application environment. This paper emphasizes the need for a holistic, end-to-end analysis of AI workloads and reveals the "AI tax." We deploy and characterize Face Recognition in an edge data center. The application is an AI-centric edge video analytics application built using popular open source infrastructure and ML tools. Despite using state-of-the-art AI and ML algorithms, the application relies heavily on pre-and post-processing code. As AI-centric applications benefit from the acceleration promised by accelerators, we find they impose stresses on the hardware and software infrastructure: storage and network bandwidth become major bottlenecks with increasing AI acceleration. By specializing for AI applications, we show that a purpose-built edge data center can be designed for the stresses of accelerated AI at 15% lower TCO than one derived from homogeneous servers and infrastructure.