Ridge Regression Surrogates
Description
This technique approximates a complex model by training a ridge regression (a linear model with L2 regularization) on the original model's predictions. The ridge regression serves as a global surrogate that balances fidelity and interpretability, capturing the main linear relationships that the complex model learned while ignoring noise due to regularization.
Example Use Cases
Explainability
Approximating a complex ensemble model used for credit scoring with a ridge regression surrogate to identify the most influential features (income, credit history, debt-to-income ratio) and their linear relationships for regulatory compliance reporting.
Creating a ridge regression surrogate of a neural network used for medical diagnosis to understand which patient symptoms and biomarkers have the strongest linear predictive relationships with disease outcomes.
Transparency
Creating an interpretable approximation of a complex insurance pricing model for regulatory compliance, enabling stakeholders to understand and validate the decision-making process through transparent linear relationships.
Limitations
- Linear approximation may miss important non-linear relationships and interactions captured by the original complex model.
- Requires a representative dataset to train the surrogate model, which may not be available or may be expensive to generate.
- Ridge regularisation may oversimplify the model by shrinking coefficients, potentially hiding important but less dominant features.
- Surrogate fidelity depends on how well linear relationships approximate the original model's behaviour across the entire input space.