Conformal Prediction
Description
Conformal prediction provides mathematically guaranteed uncertainty quantification by creating prediction sets that contain the true outcome with a specified probability (e.g., exactly 95% coverage). The technique works by measuring how 'strange' or 'nonconforming' new predictions are compared to calibration data - if a prediction seems unusual, it gets wider intervals. For example, in medical diagnosis, instead of saying 'likely cancer', it might say 'possible diagnoses: {cancer, benign tumour} with 95% confidence'. This distribution-free method works with any underlying model (neural networks, random forests, etc.) and requires no assumptions about data distribution, making it a robust framework for reliable uncertainty estimates in high-stakes applications.
Example Use Cases
Reliability
Creating prediction sets for drug discovery that guarantee 95% coverage, such as 'this compound will likely have activity against {target A, target B, target C}', ensuring reliable decision-making in costly experimental validation.
Transparency
Providing transparent multi-class predictions in judicial risk assessment by showing all plausible risk categories with guaranteed coverage, enabling judges to see the full range of possibilities rather than just a single point estimate.
Fairness
Ensuring fair uncertainty quantification across demographic groups in college admissions by verifying that prediction set sizes (number of possible outcomes) are consistent across protected groups, preventing discriminatory overconfidence for certain populations.
Limitations
- Prediction sets can be unnecessarily wide when nonconformity scores vary greatly across the feature space, leading to conservative intervals that reduce practical utility.
- Requires a held-out calibration set separate from training data, reducing the amount of data available for model training, which can impact performance on small datasets.
- Guarantees only hold under the exchangeability assumption - if test data distribution differs significantly from calibration data, coverage guarantees may be violated.
- For multi-class problems, prediction sets may include many classes when the model is uncertain, making decisions difficult when sets contain opposing outcomes.
- Computational cost increases with the number of calibration samples, and efficient implementation requires careful design for large-scale or real-time applications.
Resources
A tutorial on conformal prediction
Foundational tutorial introducing conformal prediction theory and applications by the method's creators
valeman/awesome-conformal-prediction
Curated collection of conformal prediction resources including videos, tutorials, books, papers, and open-source libraries
scikit-learn-contrib/MAPIE
Python library for uncertainty quantification using conformal prediction across regression, classification, and time series tasks