论文标题
正则化可以减轻中毒攻击:基于多目标双层优化的新颖分析
Regularisation Can Mitigate Poisoning Attacks: A Novel Analysis Based on Multiobjective Bilevel Optimisation
论文作者
论文摘要
机器学习(ML)算法容易受到中毒攻击的影响,在这些训练数据中的一小部分被操纵以故意降低算法的性能。最佳的中毒攻击可以作为双层优化问题提出,有助于评估最坏情况下学习算法的鲁棒性。但是,当前针对使用超参数的算法的攻击通常假定这些超参数仍然持续忽略攻击对它们的影响。我们表明,这种方法导致对算法鲁棒性的过于悲观的看法。我们提出了一种新颖的最佳攻击公式,该公式通过将攻击作为多物镜双重优化问题来考虑攻击对超参数的影响。我们使用$ L_2 $正则化将这种新颖的攻击公式应用于ML分类器,并表明,与先前报道的结果相比,$ L_2 $正则化增强了学习算法的稳定性,并有助于减轻攻击。我们对不同数据集的经验评估证实了先前策略的局限性,这证明了使用$ L_2 $正则化来抑制中毒攻击的效果的好处,并显示正规化超参数如何随中毒点的分数增加。
Machine Learning (ML) algorithms are vulnerable to poisoning attacks, where a fraction of the training data is manipulated to deliberately degrade the algorithms' performance. Optimal poisoning attacks, which can be formulated as bilevel optimisation problems, help to assess the robustness of learning algorithms in worst-case scenarios. However, current attacks against algorithms with hyperparameters typically assume that these hyperparameters remain constant ignoring the effect the attack has on them. We show that this approach leads to an overly pessimistic view of the robustness of the algorithms. We propose a novel optimal attack formulation that considers the effect of the attack on the hyperparameters by modelling the attack as a multiobjective bilevel optimisation problem. We apply this novel attack formulation to ML classifiers using $L_2$ regularisation and show that, in contrast to results previously reported, $L_2$ regularisation enhances the stability of the learning algorithms and helps to mitigate the attacks. Our empirical evaluation on different datasets confirms the limitations of previous strategies, evidences the benefits of using $L_2$ regularisation to dampen the effect of poisoning attacks and shows how the regularisation hyperparameter increases with the fraction of poisoning points.