Machine learning algorithms are becoming integrated into more and more high-stakes decision-making processes, such as in social welfare issues. Due to the need of mitigating the potentially disparate impacts from algorithmic predictions, many approaches have been proposed in the emerging area of fair machine learning. However, the fundamental problem of characterizing Bayes-optimal classifiers under various group fairness constraints has only been investigated in some special cases. Based on the classical Neyman-Pearson argument (Neyman and Pearson, 1933; Shao, 2003) for optimal hypothesis testing, this paper provides a unified framework for deriving Bayes-optimal classifiers under group fairness. This enables us to propose a group-based thresholding method we call FairBayes, that can directly control disparity, and achieve an essentially optimal fairness-accuracy tradeoff. These advantages are supported by thorough experiments.
翻译:机器学习算法正逐渐被纳入越来越多的高层决策程序,如社会福利问题。由于需要减轻算法预测的潜在不同影响,在新兴的公平机器学习领域提出了许多办法。然而,只在一些特殊案例中调查了在各种群体公平性限制下将贝亚斯-最佳分类者定性为各种群体公平性限制的根本问题。根据古典Neyman-皮尔逊论点(Neyman和Pearson,1933年;Shao,2003年)进行最佳假设测试,本文件为将贝亚斯-最佳分类者纳入群体公平性提供了一个统一框架。这使我们能够提出一种基于集团的门槛方法,即我们称之为FairBayes,可以直接控制差异,并实现基本上最佳的公平-准确性权衡。这些优势得到了彻底的实验的支持。