论文标题

faceleaks:针对传输学习模型的推理攻击通过黑框查询

FaceLeaks: Inference Attacks against Transfer Learning Models via Black-box Queries

论文作者

Liew, Seng Pei, Takahashi, Tsubasa

论文摘要

转移学习是一个有用的机器学习框架,它允许人们使用大量数据预先训练的单个强大模型(教师模型)来构建特定于任务的模型(学生模型),而无需显着影响培训成本。教师模型可能包含私人数据,或与私人输入相互作用。我们研究是否可以直接与教师模型进行互动,可以泄漏或推断此类私人信息。我们在面部识别的背景下描述了这种推理攻击,这是对个人隐私高度敏感的转移学习的应用。 在黑框和现实的设置下,我们表明现有的推理技术是无效的,因为通过学生模型与单个培训实例进行互动并未透露有关教师的信息。然后,我们提出了新的策略来从汇总级别的信息中推断出来。因此,即使对手只能访问学生模型,对教师模型的会员推理攻击也是可能的。 我们进一步证明,即使在对手的辅助信息有限的情况下,也可以推断出敏感属性。最后,讨论和评估防御策略。我们的广泛研究表明,信息泄漏是对在现实生活中广泛使用的转移学习框架的真正隐私威胁。

Transfer learning is a useful machine learning framework that allows one to build task-specific models (student models) without significantly incurring training costs using a single powerful model (teacher model) pre-trained with a large amount of data. The teacher model may contain private data, or interact with private inputs. We investigate if one can leak or infer such private information without interacting with the teacher model directly. We describe such inference attacks in the context of face recognition, an application of transfer learning that is highly sensitive to personal privacy. Under black-box and realistic settings, we show that existing inference techniques are ineffective, as interacting with individual training instances through the student models does not reveal information about the teacher. We then propose novel strategies to infer from aggregate-level information. Consequently, membership inference attacks on the teacher model are shown to be possible, even when the adversary has access only to the student models. We further demonstrate that sensitive attributes can be inferred, even in the case where the adversary has limited auxiliary information. Finally, defensive strategies are discussed and evaluated. Our extensive study indicates that information leakage is a real privacy threat to the transfer learning framework widely used in real-life situations.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源