论文标题

对抗性机器学习和防御游戏,用于Nextg信号分类,并深入学习

Adversarial Machine Learning and Defense Game for NextG Signal Classification with Deep Learning

论文作者

Sagduyu, Yalin E.

论文摘要

本文提出了一个游戏理论框架,以研究基于深度学习的NextG信号分类的攻击和防御的相互作用。 NextG系统(例如,设想的大量物联网设备都可以使用深层神经网络(DNN)来实现各种任务,例如用户设备识别,物理层认证以及现有用户的检测(例如在公民宽带无线电服务(CBRS)中)。通过培训另一个DNN作为代理模型,对手可以发起推理(探索性)攻击,以了解受害者模型的行为,预测成功的操作模式(例如,渠道访问)并堵塞了它们。防御机制可以通过在受害者模型的决定中引入受控错误(即毒害对手的训练数据),从而增加对手的不确定性。这种防御能够有效防止攻击,但在没有攻击时会降低性能。防守者和对手之间的相互作用被表述为一种非合作游戏,在该游戏中,防守者选择防御概率或防御水平本身(即伪造决策的比率)和对手选择攻击的可能性。辩护人的目标是最大化其奖励(例如,吞吐量或传播成功率),而对手的目标是最大程度地降低这一奖励及其攻击成本。 NASH均衡策略是确定的,因为操作模式使得鉴于​​对方的策略是固定的,任何玩家都无法单方面改善其实用性。为每个玩家制定了虚拟的游戏,以应对对手行动的经验频率反复玩游戏。将NASH平衡的性能与固定攻击和防御案例进行了比较,并量化了针对攻击的NextG信号分类的弹性。

This paper presents a game-theoretic framework to study the interactions of attack and defense for deep learning-based NextG signal classification. NextG systems such as the one envisioned for a massive number of IoT devices can employ deep neural networks (DNNs) for various tasks such as user equipment identification, physical layer authentication, and detection of incumbent users (such as in the Citizens Broadband Radio Service (CBRS) band). By training another DNN as the surrogate model, an adversary can launch an inference (exploratory) attack to learn the behavior of the victim model, predict successful operation modes (e.g., channel access), and jam them. A defense mechanism can increase the adversary's uncertainty by introducing controlled errors in the victim model's decisions (i.e., poisoning the adversary's training data). This defense is effective against an attack but reduces the performance when there is no attack. The interactions between the defender and the adversary are formulated as a non-cooperative game, where the defender selects the probability of defending or the defense level itself (i.e., the ratio of falsified decisions) and the adversary selects the probability of attacking. The defender's objective is to maximize its reward (e.g., throughput or transmission success ratio), whereas the adversary's objective is to minimize this reward and its attack cost. The Nash equilibrium strategies are determined as operation modes such that no player can unilaterally improve its utility given the other's strategy is fixed. A fictitious play is formulated for each player to play the game repeatedly in response to the empirical frequency of the opponent's actions. The performance in Nash equilibrium is compared to the fixed attack and defense cases, and the resilience of NextG signal classification against attacks is quantified.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源