Enforcing Fairness via Constraint Injection with FaUCI

   page       attach   

The problem of fairness in AI can be tackled by minimising bias in the data (pre-processing), in the algorithms (in-processing), or in the results (post-processing). In the particular case of in-processing applied to supervised machine learning, state-of-the-art solutions rely on a few well-known fairness metrics – e.g., demographic parity, disparate impact, or equalised odds – optimised during training—which, however, mostly focus on binary attributes and their effects on binary classification problems. Accordingly, in this work we propose FaUCI as a general purpose framework for injecting fairness constraints into neural networks (or, any model trained via stochastic gradient descent), supporting attributes of many sorts—there including binary, discrete, or continuous features. To evaluate its effectiveness and efficiency, we test FaUCI against several sorts of features and fairness metrics. Furthermore, we compare FaUCI with state-of-the-art solutions for in-processing, demonstrating its superiority.

reference publication
page_white_acrobatEnforcing Fairness via Constraint Injection with FaUCI (paper in proceedings, 2024) — Matteo Magnini, Giovanni Ciatto, Roberta Calegari, Andrea Omicini
funding project
wrenchAEQUITAS — Assessment and Engineering of eQuitable, Unbiased, Impartial and Trustworthy Ai Systems (01/11/2022–31/10/2025)
works as
reference talk for
page_white_acrobatEnforcing Fairness via Constraint Injection with FaUCI (paper in proceedings, 2024) — Matteo Magnini, Giovanni Ciatto, Roberta Calegari, Andrea Omicini

cover