Reliability
79 techniques
Building AI systems that perform consistently and predictably.
79 techniques
| Goals | Models | Data Types | Description | |||
|---|---|---|---|---|---|---|
| SHapley Additive exPlanations | Algorithmic | Architecture/model Agnostic Requirements/black Box | Any | SHAP explains model predictions by quantifying how much each input feature contributes to the outcome. It assigns an... | ||
| Permutation Importance | Algorithmic | Architecture/model Agnostic Requirements/black Box | Any | Permutation Importance quantifies a feature's contribution to a model's performance by randomly shuffling its values and... | ||
| Mean Decrease Impurity | Algorithmic | Architecture/tree Based Paradigm/supervised +1 | Tabular | Mean Decrease Impurity (MDI) quantifies a feature's importance in tree-based models (e.g., Random Forests, Gradient... | ||
| Monte Carlo Dropout | Algorithmic | Architecture/neural Networks Paradigm/probabilistic +4 | Any | Monte Carlo Dropout estimates prediction uncertainty by applying dropout (randomly setting neural network weights to... | ||
| Out-of-Distribution Detector for Neural Networks | Algorithmic | Architecture/neural Networks Paradigm/discriminative +3 | Any | ODIN (Out-of-Distribution Detector for Neural Networks) identifies when a neural network encounters inputs significantly... | ||
| Permutation Tests | Algorithmic | Architecture/model Agnostic Requirements/black Box | Any | Permutation tests assess the statistical significance of observed results (such as model accuracy, feature importance,... | ||
| Synthetic Data Generation | Algorithmic | Architecture/neural Networks/generative/gan Architecture/neural Networks/generative/vae +5 | Any | Synthetic data generation creates artificial datasets that aim to preserve the statistical properties, distributions,... | ||
| Federated Learning | Algorithmic | Architecture/linear Models Architecture/neural Networks +4 | Any | Federated learning enables collaborative model training across multiple distributed parties (devices, organisations, or... | ||
| Prediction Intervals | Algorithmic | Architecture/model Agnostic Paradigm/supervised +1 | Any | Prediction intervals provide a range of plausible values around a model's prediction, expressing uncertainty as 'the... | ||
| Quantile Regression | Algorithmic | Architecture/linear Models/regression Architecture/neural Networks +4 | Any | Quantile regression estimates specific percentiles (quantiles) of the target variable rather than just predicting the... | ||
| Conformal Prediction | Algorithmic | Architecture/model Agnostic Requirements/black Box | Any | Conformal prediction provides mathematically guaranteed uncertainty quantification by creating prediction sets that... | ||
| Empirical Calibration | Algorithmic | Architecture/model Agnostic Paradigm/supervised +2 | Any | Empirical calibration adjusts a model's predicted probabilities to match observed frequencies. For example, if events... | ||
| Temperature Scaling | Algorithmic | Architecture/neural Networks Paradigm/discriminative +3 | Any | Temperature scaling adjusts a model's confidence by applying a single parameter (temperature) to its predictions. When a... | ||
| Deep Ensembles | Algorithmic | Architecture/neural Networks Paradigm/parametric +2 | Any | Deep ensembles combine predictions from multiple neural networks trained independently with different random... | ||
| Bootstrapping | Algorithmic | Architecture/model Agnostic Paradigm/supervised +2 | Any | Bootstrapping estimates uncertainty by repeatedly resampling the original dataset with replacement to create many new... | ||
| Jackknife Resampling | Algorithmic | Architecture/model Agnostic Paradigm/supervised +2 | Any | Jackknife resampling (also called leave-one-out resampling) assesses model stability and uncertainty by systematically... | ||
| Cross-validation | Algorithmic | Architecture/model Agnostic Paradigm/supervised +2 | Any | Cross-validation evaluates model performance and robustness by systematically partitioning data into multiple subsets... | ||
| Area Under Precision-Recall Curve | Algorithmic | Architecture/model Agnostic Paradigm/supervised +2 | Any | Area Under Precision-Recall Curve (AUPRC) measures model performance by plotting precision (the proportion of positive... | ||
| Safety Envelope Testing | Testing | Architecture/model Agnostic Requirements/black Box | Any | Safety envelope testing systematically evaluates AI system performance at the boundaries of its intended operational... | ||
| Internal Review Boards | Process | Architecture/model Agnostic Requirements/black Box | Any | Internal Review Boards (IRBs) provide independent, systematic evaluation of AI/ML projects throughout their lifecycle to... |
Rows per page
Page 1 of 4