Measuring Social Biases of Crowd Workers using Counterfactual Queries
Social biases based on gender, race, etc. have been shown to pollute machine learning (ML) pipeline predominantly via biased training datasets. Crowdsourcing, a popular cost-effective measure to gather labeled training datasets, is not immune to the inherent social biases of crowd workers. To ensure...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Social biases based on gender, race, etc. have been shown to pollute machine
learning (ML) pipeline predominantly via biased training datasets.
Crowdsourcing, a popular cost-effective measure to gather labeled training
datasets, is not immune to the inherent social biases of crowd workers. To
ensure such social biases aren't passed onto the curated datasets, it's
important to know how biased each crowd worker is. In this work, we propose a
new method based on counterfactual fairness to quantify the degree of inherent
social bias in each crowd worker. This extra information can be leveraged
together with individual worker responses to curate a less biased dataset. |
---|---|
DOI: | 10.48550/arxiv.2004.02028 |