Prototype and Criticism Models

Description

Prototype and Criticism Models provide data understanding by identifying two complementary sets of examples: prototypes represent the most typical instances that best summarise common patterns in the data, whilst criticisms are outliers or edge cases that are poorly represented by the prototypes. For example, in a dataset of customer transactions, prototypes might be the most representative buying patterns (frequent small purchases, occasional large purchases), whilst criticisms could be unusual behaviors (bulk buyers, one-time high-value customers). This dual approach reveals both what is normal and what is exceptional, helping understand data coverage and model blind spots.

Example Use Cases

Explainability

Analysing medical imaging datasets to identify prototype scans that represent typical healthy tissue patterns and criticism examples showing rare disease presentations, helping radiologists understand what the model considers 'normal' versus cases requiring special attention.

Evaluating credit scoring models by finding prototype borrowers who represent typical low-risk profiles and criticism cases showing unusual but legitimate financial patterns that the model might misclassify, ensuring fair treatment of edge cases.

Fairness

Evaluating representation bias in hiring datasets by examining whether prototypes systematically exclude certain demographic groups and criticisms disproportionately represent minorities, revealing data collection inequities.

Limitations

  • Selection of prototypes and criticisms is highly dependent on the choice of distance metric or similarity measure, which may not capture all meaningful relationships in the data.
  • Computational complexity can become prohibitive for very large datasets, as the method often requires pairwise comparisons or optimisation over the entire dataset.
  • The number of prototypes and criticisms to select is typically a hyperparameter that requires domain expertise to set appropriately.
  • Results may not generalise well if the training data distribution differs significantly from the deployment data distribution.

Resources

Examples are not Enough, Learn to Criticize! Criticism for Interpretability
Research PaperBeen Kim, Rajiv Khanna, and Oluwasanmi O. KoyejoDec 5, 2016
SeldonIO/alibi
Software Package
Prototype Selection for Interpretable Classification
Research PaperOscar Reyes, Carlos Morell, and Sebastian VenturaFeb 27, 2012
Alibi Explain Documentation
Documentation

Tags

Applicable Models:
Data Requirements:
Data Type:
Evidence Type:
Expertise Needed:
Explanatory Scope:
Lifecycle Stage:
Technique Type: