Disparate Impact Remover

Description

Disparate Impact Remover is a preprocessing technique that transforms feature values in a dataset to reduce statistical dependence between features and protected attributes (like race or gender). The method modifies non-protected features through mathematical transformations that preserve the utility of the data whilst reducing correlations that could lead to discriminatory outcomes. This approach specifically targets the '80% rule' disparate impact threshold by adjusting feature distributions to ensure more equitable treatment across demographic groups in downstream model predictions.

Example Use Cases

Fairness

Transforming features in a credit scoring dataset where variables like 'years of employment' and 'education level' are correlated with race, applying mathematical transformations to reduce these correlations whilst preserving the predictive value for creditworthiness assessment.

Preprocessing a recruitment dataset where features like 'previous job titles' and 'university attended' correlate with gender, modifying these features to ensure the '80% rule' is met whilst maintaining useful information for predicting job performance.

Transparency

Preprocessing financial lending data to provide transparent bias metrics showing the quantified reduction in correlation between protected attributes and creditworthiness features, enabling institutions to demonstrate compliance with the 80% rule and explain their fairness interventions to regulators.

Reliability

Ensuring consistent model performance across demographic groups in healthcare risk assessment by mathematically transforming features to reduce protected attribute correlations, improving reliability of predictions for minority populations who may have been systematically under-served.

Limitations

  • Feature transformations may reduce model accuracy by removing or distorting important predictive information during the debiasing process.
  • Only addresses measured protected attributes and cannot eliminate bias that operates through unmeasured proxy variables.
  • Effectiveness depends on the specific transformation method chosen and may not generalise well to different datasets or domains.
  • May create artificial feature distributions that don't reflect real-world data patterns, potentially causing issues in model deployment.

Resources

Software Packages

holisticai
Jan 26, 2023

This is an open-source tool to assess and improve the trustworthiness of AI systems.

AIF360
Aug 22, 2018

A comprehensive set of fairness metrics for datasets and machine learning models, explanations for these metrics, and algorithms to mitigate bias in datasets and models.

Documentations

Disparate Impact Remover — holisticai documentation
Holisticai DevelopersJan 1, 2023
aif360.algorithms.preprocessing.DisparateImpactRemover — aif360 ...
Aif360 DevelopersJan 1, 2018

Tags

Data Requirements:
Data Type:
Evidence Type:
Expertise Needed:
Fairness Approach:
Lifecycle Stage:
Technique Type: