论文标题

DeepHider:基于多任务学习的秘密NLP水印框架

DeepHider: A Covert NLP Watermarking Framework Based on Multi-task Learning

论文作者

Dai, Long, Mao, Jiarong, Fan, Xuefeng, Zhou, Xiaoyi

论文摘要

自然语言处理(NLP)技术在诸如情感分析之类的应用中显示出巨大的商业价值。但是,NLP模型容易受到盗版重新分配的威胁,从而损害了模型所有者的经济利益。数字水印技术是保护NLP模型的知识产权的有效手段。现有的NLP模型保护主要是通过改善安全性和鲁棒性目的设计水印方案,但是,这些方案的安全性和鲁棒性分别存在以下问题:(1)在验证过程中,很难通过对手检测并易于验证并因验证验证器而被验证和阻止欺诈过程。 (2)水印模型不能同时满足多个鲁棒性要求。为了解决上述问题,本文提出了基于深度模型过度参数和多任务学习理论的NLP模型的新型水印框架。具体而言,建立了一个秘密触发器集,以实现对水印模型的无知验证,而新颖的辅助网络旨在提高水印模型的鲁棒性和安全性。在两个基准数据集和三个主流NLP模型上评估了所提出的框架,结果表明,该框架可以通过100%验证精度和高级鲁棒性和安全性成功验证模型所有权,而不会损害主机模型性能。

Natural language processing (NLP) technology has shown great commercial value in applications such as sentiment analysis. But NLP models are vulnerable to the threat of pirated redistribution, damaging the economic interests of model owners. Digital watermarking technology is an effective means to protect the intellectual property rights of NLP model. The existing NLP model protection mainly designs watermarking schemes by improving both security and robustness purposes, however, the security and robustness of these schemes have the following problems, respectively: (1) Watermarks are difficult to defend against fraudulent declaration by adversary and are easily detected and blocked from verification by human or anomaly detector during the verification process. (2) The watermarking model cannot meet multiple robustness requirements at the same time. To solve the above problems, this paper proposes a novel watermarking framework for NLP model based on the over-parameterization of depth model and the multi-task learning theory. Specifically, a covert trigger set is established to realize the perception-free verification of the watermarking model, and a novel auxiliary network is designed to improve the robustness and security of the watermarking model. The proposed framework was evaluated on two benchmark datasets and three mainstream NLP models, and the results show that the framework can successfully validate model ownership with 100% validation accuracy and advanced robustness and security without compromising the host model performance.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源