SHapley Additive exPlanations
Description
SHAP explains model predictions by quantifying how much each input feature contributes to the outcome. It assigns an importance score to every feature, indicating whether it pushes the prediction towards or away from the average. The method systematically evaluates how predictions change as features are included or excluded, drawing on game theory concepts to ensure a fair distribution of contributions.
Example Use Cases
Explainability
Analysing a customer churn prediction model to understand why a specific high-value customer was flagged as likely to leave, revealing that recent support ticket interactions and declining purchase frequency were the main drivers.
Fairness
Auditing a loan approval model by comparing SHAP values for applicants from different demographic groups, ensuring that protected characteristics like race or gender do not have an undue influence on credit decisions.
Reliability
Validating a medical diagnosis model by confirming that its predictions are based on relevant clinical features (e.g., blood pressure, cholesterol levels) rather than spurious correlations (e.g., patient ID or appointment time), thereby improving model reliability.
Limitations
- Assumes feature independence, which can produce misleading explanations when features are highly correlated, as the model may attribute importance to features that are merely proxies for others.
- Computationally expensive for models with many features or large datasets, as the number of required predictions grows exponentially with the number of features.
- The choice of background dataset for generating explanations can significantly influence the results, requiring careful selection to ensure a representative baseline.
- Global explanations derived from averaging local SHAP values may obscure important heterogeneous effects where features impact subgroups of the population differently.