论文标题
深度学习中的隐私:调查
Privacy in Deep Learning: A Survey
论文作者
论文摘要
在许多领域的深入学习的不断增长的进步,包括愿景,推荐系统,自然语言处理等,导致在生产系统中采用了深层神经网络(DNN)。大型数据集和高计算能力的可用性是这些进步的主要因素。数据集通常是众包,可能包含敏感信息。这构成了严重的隐私问题,因为可以通过各种漏洞滥用或泄漏此数据。即使信任云提供商和通信链接,仍然存在推理攻击的威胁,攻击者可以推测用于培训的数据的属性,或找到基础模型架构和参数。在这项调查中,我们回顾了深度学习带来的隐私问题,以及为解决这些问题而引入的缓解技术。我们还表明,文献中存在有关测试时间推理隐私的差距,并提出了未来的研究指示。
The ever-growing advances of deep learning in many areas including vision, recommendation systems, natural language processing, etc., have led to the adoption of Deep Neural Networks (DNNs) in production systems. The availability of large datasets and high computational power are the main contributors to these advances. The datasets are usually crowdsourced and may contain sensitive information. This poses serious privacy concerns as this data can be misused or leaked through various vulnerabilities. Even if the cloud provider and the communication link is trusted, there are still threats of inference attacks where an attacker could speculate properties of the data used for training, or find the underlying model architecture and parameters. In this survey, we review the privacy concerns brought by deep learning, and the mitigating techniques introduced to tackle these issues. We also show that there is a gap in the literature regarding test-time inference privacy, and propose possible future research directions.