论文标题

可扩展的白盒攻击基于树的模型

Scalable Whitebox Attacks on Tree-based Models

论文作者

Castiglione, Giuseppe, Ding, Gavin, Hashemi, Masoud, Srinivasa, Christopher, Wu, Ga

论文摘要

对抗性鲁棒性是确保机器学习模型可靠性的基本安全标准之一。尽管在过去的十年中引入了各种对抗性鲁棒性测试方法,但我们注意到其中大多数与非差异性模型(例如树团)不相容。由于树的合奏被广泛用于行业,因此揭示了对抗性鲁棒性研究与实际应用之间的关键差距。本文提出了一种新型的白盒对抗性鲁棒性测试方法,用于树集合模型。具体而言,所提出的方法通过温度控制的sigmoid函数使树的合奏平滑,从而实现基于梯度下降的对抗性攻击。通过利用采样和对数衍生技巧,建议的方法可以扩展到测试以前难以管理的任务。我们将方法与多个公共数据集(以及相应模型)上的随机扰动和黑框方法进行比较。我们的结果表明,所提出的方法可以1)成功揭示了树集合模型的对抗脆弱性,而不会引起测试的计算压力,而2)灵活平衡搜索性能和时间复杂性,以满足各种测试标准。

Adversarial robustness is one of the essential safety criteria for guaranteeing the reliability of machine learning models. While various adversarial robustness testing approaches were introduced in the last decade, we note that most of them are incompatible with non-differentiable models such as tree ensembles. Since tree ensembles are widely used in industry, this reveals a crucial gap between adversarial robustness research and practical applications. This paper proposes a novel whitebox adversarial robustness testing approach for tree ensemble models. Concretely, the proposed approach smooths the tree ensembles through temperature controlled sigmoid functions, which enables gradient descent-based adversarial attacks. By leveraging sampling and the log-derivative trick, the proposed approach can scale up to testing tasks that were previously unmanageable. We compare the approach against both random perturbations and blackbox approaches on multiple public datasets (and corresponding models). Our results show that the proposed method can 1) successfully reveal the adversarial vulnerability of tree ensemble models without causing computational pressure for testing and 2) flexibly balance the search performance and time complexity to meet various testing criteria.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源