The issue of fairness in machine learning stems from the fact that historical data often displays biases against specific groups of people which have been underprivileged in the recent past, or still are. In this context, one of the possible approaches is to employ fair representation learning algorithms which are able to remove biases from data, making groups statistically indistinguishable. In this paper, we instead develop a fair representation learning algorithm which is able to map individuals belonging to different groups in a single group. This is made possible by training a pair of Normalizing Flow models and constraining them to not remove information about the ground truth by training a ranking or classification model on top of them. The overall, ``chained'' model is invertible and has a tractable Jacobian, which allows to relate together the probability densities for different groups and ``translate'' individuals from one group to another. We show experimentally that our methodology is competitive with other fair representation learning algorithms. Furthermore, our algorithm achieves stronger invariance w.r.t. the sensitive attribute.
翻译:机器学习的公平问题源于历史数据往往显示对最近过去处于劣势或仍然处于劣势的特定人群的偏见。 在这方面,一种可能的方法是采用公平的代表性学习算法,这种算法能够消除数据中的偏向,使各群体在统计上无法区分。在本文中,我们相反地发展一种公平的代表性学习算法,能够绘制属于不同群体的个人在单一群体中的分布图。这是通过培训一对正常流动模型,并限制他们不通过培训排名或分类模型来去除关于地面真相的信息。总体而言,“链式”模型是不可忽略的,具有可感人雅各语,可以将不同群体和“将个人从一个群体转到另一个群体”的概率密度联系起来。我们实验性地表明,我们的方法与其他公平的代表性学习算法相比是竞争性的。此外,我们的算法在敏感属性上更加强烈的不理解。