Ridge Regression Surrogates

Description

This technique approximates a complex model by training a ridge regression (a linear model with L2 regularization) on the original model's predictions. The ridge regression serves as a global surrogate that balances fidelity and interpretability, capturing the main linear relationships that the complex model learned while ignoring noise due to regularization.

Example Use Cases

Explainability

Approximating a complex ensemble model used for credit scoring with a ridge regression surrogate to identify the most influential features (income, credit history, debt-to-income ratio) and their linear relationships for regulatory compliance reporting.

Creating a ridge regression surrogate of a neural network used for medical diagnosis to understand which patient symptoms and biomarkers have the strongest linear predictive relationships with disease outcomes.

Transparency

Creating an interpretable approximation of a complex insurance pricing model for regulatory compliance, enabling stakeholders to understand and validate the decision-making process through transparent linear relationships.

Limitations

  • Linear approximation may miss important non-linear relationships and interactions captured by the original complex model.
  • Requires a representative dataset to train the surrogate model, which may not be available or may be expensive to generate.
  • Ridge regularisation may oversimplify the model by shrinking coefficients, potentially hiding important but less dominant features.
  • Surrogate fidelity depends on how well linear relationships approximate the original model's behaviour across the entire input space.

Resources

scikit-learn Ridge Regression Documentation
Documentation
Interpretable Machine Learning: A Guide for Making Black Box Models Explainable
Documentation

Tags

Applicable Models:
Data Requirements:
Data Type:
Evidence Type:
Expertise Needed:
Explanatory Scope:
Lifecycle Stage:
Technique Type: