论文标题
公平性和解释性:弥合差距为公平模型解释
Fairness and Explainability: Bridging the Gap Towards Fair Model Explanations
论文作者
论文摘要
尽管机器学习模型在实际应用中取得了前所未有的成功,但它们可能会对特定的人群群体做出有偏见/不公平的决定,从而导致歧视结果。尽管研究工作已致力于测量和减轻偏见,但它们主要从结果的角度研究偏见,同时忽略了决策程序中编码的偏见。这导致他们无法捕获以程序为导向的偏差,因此限制了具有完全辩解方法的能力。幸运的是,随着可解释的机器学习的迅速发展,现在可以使用预测的解释来了解该过程。在这项工作中,我们通过基于解释的程序公平性展示了一个新颖的视角,弥合了公平与解释性之间的差距。我们通过测量不同组之间具有基于比率和基于价值的解释公平性的不同组之间的解释质量差距来确定基于程序的偏见。新的指标进一步促使我们设计一个优化目标,以减轻基于程序的偏见,我们观察到它也会减轻预测中的偏见。基于我们设计的优化目标,我们提出了一种全面的公平算法(CFA),该算法同时实现了多个目标 - 改善了传统的公平,满足解释公平并保持公用事业绩效。对现实世界数据集的广泛实验证明了我们提出的CFA的有效性,并强调了从解释性的角度考虑公平性的重要性。我们的代码可在https://github.com/yuyingzhao/fairexplanates-cfa上公开获取。
While machine learning models have achieved unprecedented success in real-world applications, they might make biased/unfair decisions for specific demographic groups and hence result in discriminative outcomes. Although research efforts have been devoted to measuring and mitigating bias, they mainly study bias from the result-oriented perspective while neglecting the bias encoded in the decision-making procedure. This results in their inability to capture procedure-oriented bias, which therefore limits the ability to have a fully debiasing method. Fortunately, with the rapid development of explainable machine learning, explanations for predictions are now available to gain insights into the procedure. In this work, we bridge the gap between fairness and explainability by presenting a novel perspective of procedure-oriented fairness based on explanations. We identify the procedure-based bias by measuring the gap of explanation quality between different groups with Ratio-based and Value-based Explanation Fairness. The new metrics further motivate us to design an optimization objective to mitigate the procedure-based bias where we observe that it will also mitigate bias from the prediction. Based on our designed optimization objective, we propose a Comprehensive Fairness Algorithm (CFA), which simultaneously fulfills multiple objectives - improving traditional fairness, satisfying explanation fairness, and maintaining the utility performance. Extensive experiments on real-world datasets demonstrate the effectiveness of our proposed CFA and highlight the importance of considering fairness from the explainability perspective. Our code is publicly available at https://github.com/YuyingZhao/FairExplanations-CFA .