In a world increasingly reliant on artificial intelligence, it is more important than ever to consider the ethical implications of artificial intelligence on humanity. One key under-explored challenge is labeler bias, which can create inherently biased datasets for training and subsequently lead to inaccurate or unfair decisions in healthcare, employment, education, and law enforcement. Hence, we conducted a study to investigate and measure the existence of labeler bias using images of people from different ethnicities and sexes in a labeling task. Our results show that participants possess stereotypes that influence their decision-making process and that labeler demographics impact assigned labels. We also discuss how labeler bias influences datasets and, subsequently, the models trained on them. Overall, a high degree of transparency must be maintained throughout the entire artificial intelligence training process to identify and correct biases in the data as early as possible.
翻译:在一个日益依赖人工智能的世界中,现在比以往任何时候都更需要考虑人工智能对人类的伦理影响。一个未得到充分探讨的一个关键挑战是标签偏见,它可能为培训创造内在的偏差数据集,随后导致在保健、就业、教育和执法方面作出不准确或不公平的决定。因此,我们进行了一项研究,调查并衡量标签偏见的存在,在标签工作中使用不同族裔和性别的人的形象。我们的研究结果表明,参与者持有影响其决策过程的陈规定型观念,而且标签人口对所分配的标签产生影响。我们还讨论了标签偏见如何影响数据集以及随后培训的模型。总的来说,在整个人工智能培训过程中必须保持高度的透明度,以便尽早查明并纠正数据中的偏差。