论文标题

学习的TCP充血控制的符号蒸馏

Symbolic Distillation for Learned TCP Congestion Control

论文作者

Sharan, S P, Zheng, Wenqing, Hsu, Kuo-Feng, Xing, Jiarong, Chen, Ang, Wang, Zhangyang

论文摘要

TCP拥塞控制(CC)的最新进展通过深度强化学习(RL)方法取得了巨大的成功,这些方法使用前馈神经网络(NN)来学习复杂的环境条件并做出更好的决策。但是,这种“黑框”策略缺乏可解释性和可靠性,并且由于使用复杂的NNS,它们通常需要在传统的TCP数据管道之外进行操作。本文提出了一种新颖的两阶段解决方案,以实现两全其美:首先训练深度RL代理,然后将其(过度)参数化的NN策略提炼成白色盒子,以符号表达式的形式进行轻重量的规则,这些规则更容易理解并在受约束环境中实现。我们提案的核心是一种新颖的符号分支算法,该算法使该规则能够在各种网络条件下意识到上下文,最终将NN策略转换为符号树。蒸馏符号规则比标准神经网络更快,更简单,在最先进的NN政策上保留并经常提高性能。我们在模拟环境和仿真环境中验证了蒸馏符号规则的性能。我们的代码可在https://github.com/vita-group/symbolicpcc上找到。

Recent advances in TCP congestion control (CC) have achieved tremendous success with deep reinforcement learning (RL) approaches, which use feedforward neural networks (NN) to learn complex environment conditions and make better decisions. However, such "black-box" policies lack interpretability and reliability, and often, they need to operate outside the traditional TCP datapath due to the use of complex NNs. This paper proposes a novel two-stage solution to achieve the best of both worlds: first to train a deep RL agent, then distill its (over-)parameterized NN policy into white-box, light-weight rules in the form of symbolic expressions that are much easier to understand and to implement in constrained environments. At the core of our proposal is a novel symbolic branching algorithm that enables the rule to be aware of the context in terms of various network conditions, eventually converting the NN policy into a symbolic tree. The distilled symbolic rules preserve and often improve performance over state-of-the-art NN policies while being faster and simpler than a standard neural network. We validate the performance of our distilled symbolic rules on both simulation and emulation environments. Our code is available at https://github.com/VITA-Group/SymbolicPCC.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源