Fairmod: making predictions fair in multiple protected attributes

Predictive models such as decision trees and neural networks may produce predictions with unfairness. Some algorithms have been proposed to mitigate unfair predictions when protected attributes are considered individually. However, mitigating unfair predictions becomes more difficult when the applic...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Knowledge and information systems 2024-03, Vol.66 (3), p.1861-1884
Hauptverfasser: Liu, Jixue, Li, Jiuyong, Liu, Lin, Le, Thuc, Ye, Feiyue, Li, Gefei
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Predictive models such as decision trees and neural networks may produce predictions with unfairness. Some algorithms have been proposed to mitigate unfair predictions when protected attributes are considered individually. However, mitigating unfair predictions becomes more difficult when the application involves multiple protected attributes that are expected to be enforced simultaneously. This issue has not been solved, and existing methods are not able to solve it. The paper aims to be the first to solve this problem and proposes a method for post-processing unfair predictions to achieve fair ones. The method considers multiple simultaneous protected attributes together with context attributes, such as position, profession and education, that describe contextual details of the application. Our method consists of two steps. The first step uses a nonlinear optimization problem to determine the best adjustment plan for meeting the requirements of multiple simultaneous protected attributes while better preserving the original predictions. This optimization guarantees the solution to handle the interaction among multiple protected attributes regarding fairness in the best manner. The second steps learns adjustment thresholds using the results of optimization. The proposed method is evaluated using real-world datasets, and the evaluation shows that the proposed method makes effective adjustments to meet fairness requirements.
ISSN:0219-1377
0219-3116
DOI:10.1007/s10115-023-02003-4