论文标题
面部:用于媒体身份验证的半碎片神经水印和对抗深击
FaceSigns: Semi-Fragile Neural Watermarks for Media Authentication and Countering Deepfakes
论文作者
论文摘要
由于逼真的图像和视频合成技术的最新进展,深层媒体和操纵媒体正在成为一个巨大的威胁。已经尝试使用机器学习分类器来打击深击。但是,这种分类器并不能很好地推广到黑框图像合成技术,并且已被证明容易受到对抗示例的影响。为了应对这些挑战,我们介绍了一种基于深度学习的半碎片水印技术,该技术可以通过验证图像像素中嵌入的无形秘密消息,从而允许媒体身份验证。我们建议不要使用视觉伪像识别和检测假媒体,而是建议将半碎片水印主动嵌入真实图像中,以便我们可以在需要时证明其真实性。我们的水印框架旨在易碎面部操作或篡改,同时对良性图像处理操作(例如图像压缩,缩放,饱和度,对比度调节等)都具有良好的良好性,这允许在Internet上共享图像,以保留可验证的水印,只要不应用任何其他深层修改技术。我们证明,面部可以将128位的秘密嵌入不可感知的图像水印,可以在几个压缩水平下以很高的恢复精度恢复,而当应用看不见的DeepFake操作时,它是不可恢复的。对于在我们的工作中研究的一系列看不见的良性和深层操作,面部可以可靠地检测到具有0.996的AUC得分的操纵内容,这显着高于先前的图像水印和隐照技术。
Deepfakes and manipulated media are becoming a prominent threat due to the recent advances in realistic image and video synthesis techniques. There have been several attempts at combating Deepfakes using machine learning classifiers. However, such classifiers do not generalize well to black-box image synthesis techniques and have been shown to be vulnerable to adversarial examples. To address these challenges, we introduce a deep learning based semi-fragile watermarking technique that allows media authentication by verifying an invisible secret message embedded in the image pixels. Instead of identifying and detecting fake media using visual artifacts, we propose to proactively embed a semi-fragile watermark into a real image so that we can prove its authenticity when needed. Our watermarking framework is designed to be fragile to facial manipulations or tampering while being robust to benign image-processing operations such as image compression, scaling, saturation, contrast adjustments etc. This allows images shared over the internet to retain the verifiable watermark as long as face-swapping or any other Deepfake modification technique is not applied. We demonstrate that FaceSigns can embed a 128 bit secret as an imperceptible image watermark that can be recovered with a high bit recovery accuracy at several compression levels, while being non-recoverable when unseen Deepfake manipulations are applied. For a set of unseen benign and Deepfake manipulations studied in our work, FaceSigns can reliably detect manipulated content with an AUC score of 0.996 which is significantly higher than prior image watermarking and steganography techniques.