论文标题
野外深度学习的隐私分析:反转移学习的会员推理攻击
Privacy Analysis of Deep Learning in the Wild: Membership Inference Attacks against Transfer Learning
论文作者
论文摘要
尽管被部署在许多关键应用程序中,但机器学习(ML)模型容易受到各种安全性和隐私攻击的影响。该领域中的一个主要隐私攻击是会员推理,在该推论中,对手的目的是确定目标数据样本是否是目标ML模型训练集的一部分。到目前为止,针对从头开始训练的ML模型,对当前的大多数会员推理攻击进行了评估。但是,通常在转移学习范式之后对现实世界中的ML模型进行了培训,在此范式下,模型所有者从其他数据集(即教师模型)中学习了一个验证的模型,并通过用自己的数据对教师模型进行微调来训练自己的学生模型。 在本文中,我们对针对转移学习模型的成员推理攻击进行了首次系统评估。我们采用影子模型培训的策略来得出培训会员推理分类器的数据。在四个现实世界图像数据集上进行的广泛实验表明,会员推理可以实现有效的性能。例如,在从RESNET20转移的CIFAR100分类器(由Caltech101预审到),我们的会员推理可实现$ 95 \%$ $攻击AUC。此外,我们表明,当目标模型的体系结构未知时,会员推理仍然有效。我们的结果阐明了实际上机器学习模型引起的会员风险的严重性。
While being deployed in many critical applications as core components, machine learning (ML) models are vulnerable to various security and privacy attacks. One major privacy attack in this domain is membership inference, where an adversary aims to determine whether a target data sample is part of the training set of a target ML model. So far, most of the current membership inference attacks are evaluated against ML models trained from scratch. However, real-world ML models are typically trained following the transfer learning paradigm, where a model owner takes a pretrained model learned from a different dataset, namely teacher model, and trains her own student model by fine-tuning the teacher model with her own data. In this paper, we perform the first systematic evaluation of membership inference attacks against transfer learning models. We adopt the strategy of shadow model training to derive the data for training our membership inference classifier. Extensive experiments on four real-world image datasets show that membership inference can achieve effective performance. For instance, on the CIFAR100 classifier transferred from ResNet20 (pretrained with Caltech101), our membership inference achieves $95\%$ attack AUC. Moreover, we show that membership inference is still effective when the architecture of target model is unknown. Our results shed light on the severity of membership risks stemming from machine learning models in practice.