论文标题
具有差异隐私的深度学习的钢化sigmoid激活
Tempered Sigmoid Activations for Deep Learning with Differential Privacy
论文作者
论文摘要
由于学习有时涉及敏感数据,因此已经扩展了机器学习算法以为培训数据提供隐私。实际上,这主要是事后的想法,通过使用其他优化器重新运行培训获得了隐私保护模型,但使用已经在非私人保护设置中表现良好的模型体系结构。正如我们在这里所展示的那样,这种方法导致不理想的隐私/公用事业权衡。取而代之的是,我们建议根据隐私保护培训明确选择模型体系结构。 为了根据差异隐私的黄金标准提供保证,必须尽可能严格地束缚单个培训点可能影响模型更新。在本文中,我们第一个观察到激活功能的选择对于界定了保护隐私深度学习的敏感性至关重要。我们在分析和实验上证明了一个界有界激活功能的一般家族,钢化sigmoids如何始终超过诸如relu之类的无界激活函数。使用此范式,我们在MNIST,FashionMnist和CIFAR10上实现了新的最新准确性,而无需对学习程序基础或差异隐私分析进行任何修改。
Because learning sometimes involves sensitive data, machine learning algorithms have been extended to offer privacy for training data. In practice, this has been mostly an afterthought, with privacy-preserving models obtained by re-running training with a different optimizer, but using the model architectures that already performed well in a non-privacy-preserving setting. This approach leads to less than ideal privacy/utility tradeoffs, as we show here. Instead, we propose that model architectures are chosen ab initio explicitly for privacy-preserving training. To provide guarantees under the gold standard of differential privacy, one must bound as strictly as possible how individual training points can possibly affect model updates. In this paper, we are the first to observe that the choice of activation function is central to bounding the sensitivity of privacy-preserving deep learning. We demonstrate analytically and experimentally how a general family of bounded activation functions, the tempered sigmoids, consistently outperform unbounded activation functions like ReLU. Using this paradigm, we achieve new state-of-the-art accuracy on MNIST, FashionMNIST, and CIFAR10 without any modification of the learning procedure fundamentals or differential privacy analysis.