论文标题

FAIRGBM:梯度提高限制

FairGBM: Gradient Boosting with Fairness Constraints

论文作者

Cruz, André F, Belém, Catarina, Jesus, Sérgio, Bravo, João, Saleiro, Pedro, Bizarro, Pedro

论文摘要

表格数据在许多高风险领域(例如金融服务或公共政策)中很普遍。由于其可扩展性,性能和较低的培训成本,因此在这些设置中梯度提升决策树(GBDT)很受欢迎。尽管这些领域的公平性是最重要的关注点,但现有的进行核对公平的ML方法要么与GBDT不兼容,要么遭受重大绩效损失,同时需要更长的时间进行训练。我们提出了FairgBM,这是一个在公平限制下培训GBDT的双层学习框架,与不受约束的GBDT相比,对预测性能几乎没有影响。由于观察公平度指标是不可差异的,因此我们建议使用``代理 - lagrangian''配方的平稳凸错误率代理,从而实现基于梯度的优化。我们的实施显示了相对于相关工作的训练时间的速度加速度,这是促进现实世界实践者广泛采用FairgBM的关键方面。

Tabular data is prevalent in many high-stakes domains, such as financial services or public policy. Gradient Boosted Decision Trees (GBDT) are popular in these settings due to their scalability, performance, and low training cost. While fairness in these domains is a foremost concern, existing in-processing Fair ML methods are either incompatible with GBDT, or incur in significant performance losses while taking considerably longer to train. We present FairGBM, a dual ascent learning framework for training GBDT under fairness constraints, with little to no impact on predictive performance when compared to unconstrained GBDT. Since observational fairness metrics are non-differentiable, we propose smooth convex error rate proxies for common fairness criteria, enabling gradient-based optimization using a ``proxy-Lagrangian'' formulation. Our implementation shows an order of magnitude speedup in training time relative to related work, a pivotal aspect to foster the widespread adoption of FairGBM by real-world practitioners.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源