MCP Server
Setup guide and tool reference for the TEA Techniques MCP server
The Model Context Protocol (MCP) is an open standard that lets AI assistants like Claude connect to external data sources. The TEA Techniques MCP server gives Claude direct access to our knowledge graph—enabling conversational discovery of all TEA techniques.
Why Use the MCP Server?
Instead of manually browsing pages or writing API queries, you can ask an AI agent questions in natural language:
| Task | Manual Approach | With MCP |
|---|---|---|
| Find fairness techniques for neural networks | Browse categories, filter manually, read each page | "Find fairness techniques that work with neural networks" |
| Compare SHAP vs LIME | Open two tabs, compare side by side | "Compare SHAP and LIME and help me evaluate which is the best one for my use case" |
| Find evidence for an assurance claim | Search docs, map techniques to claims | "What techniques support the claim that our model treats all groups fairly?" |
| Explore related techniques | Follow links manually | "What techniques are related to SHAP?" |
How It Works
The MCP server runs locally on your machine and provides 10 tools that Claude can call. When you ask a question, Claude selects the appropriate tools, the server queries the knowledge graph, and Claude synthesises the results into a helpful response.
Quick Start
No installation needed — run directly with npx:
npx @chrisdburr/tea-techniques-mcp
Add to your project's .mcp.json:
{
"mcpServers": {
"tea-techniques": {
"type": "stdio",
"command": "npx",
"args": ["-y", "@chrisdburr/tea-techniques-mcp"]
}
}
}
After configuring, restart Claude. You should see 10 tools prefixed with mcp__tea-techniques__.
On first run, the server fetches the knowledge graph from GitHub Pages and caches it locally for 24 hours. Semantic search also downloads a lightweight ONNX model (~30MB) on first use.
Local development (from source)
If you want to contribute or run from the monorepo:
git clone https://github.com/alan-turing-institute/tea-techniques.git
cd tea-techniques/mcp-server
pnpm install
Then use npx tsx ./mcp-server/src/index.ts as the command in your MCP config, with --local to load data from project files instead of fetching remotely.
Workflow Examples
Discovering Techniques for a Use Case
Scenario: Building a credit scoring model that needs explainable decisions
You: I'm building a credit scoring model using gradient boosting. What techniques can help me explain individual decisions to applicants?
Claude uses find_techniques with goals=["explainability"], tags=["tree-based"]
Claude: Based on your needs, here are relevant techniques:
- SHAP — Quantifies each feature's contribution using game-theoretic principles.
- LIME — Creates simple local explanations by approximating the model.
- Permutation Importance — Measures feature importance by shuffling values.
You: How do SHAP and LIME compare?
Claude uses compare_techniques with slugs=["shapley-additive-explanations", "local-interpretable-model-agnostic-explanations"]
Claude: SHAP provides theoretically grounded explanations with consistency guarantees but is more computationally expensive. LIME is simpler and faster but may produce less stable explanations.
Building Assurance Evidence
Scenario: Demonstrating your model doesn't discriminate
You: What types of evidence can fairness techniques produce?
Claude uses find_evidence_types
Claude: Fairness techniques can produce:
- Fairness metrics (16 techniques) — Demographic parity, equalised odds
- Quantitative metrics (98 techniques) — Statistical measures of behaviour
- Qualitative reports (15 techniques) — Human review and documentation
You: Show me techniques that produce fairness metrics for neural networks.
Claude uses find_techniques with goals=["fairness"], tags=["neural-network", "quantitative-metric"]
Research and Exploration
Scenario: Understanding the AI assurance landscape
You: Give me an overview of the TEA Techniques knowledge graph.
Claude uses get_knowledge_graph_summary
Claude: The knowledge graph contains 121 techniques across 7 assurance goals, with 456 academic resources and 194 tags in a hierarchical taxonomy.
You: Which goal has the most techniques?
Claude uses coverage_statistics with dimension="goals"
Claude: Reliability has the most coverage (75 techniques), followed by Transparency (69) and Fairness (61). Privacy has the least (11).
Tool Reference
Discovery Tools
Tools for finding and exploring techniques.
find_techniques
Search and filter techniques by query text, assurance goals, and tags.
| Parameter | Type | Description |
|---|---|---|
query | string | Free-text search across names and descriptions |
goals | string[] | Filter by goal slugs (e.g. ["explainability", "fairness"]) |
tags | string[] | Filter by tag fragments (e.g. ["model-agnostic", "tabular"]) |
limit | number | Max results (default 20) |
get_technique
Get full details of a technique: goals, tags, use cases, limitations, related techniques, and resources.
| Parameter | Type | Description |
|---|---|---|
slug | string | Technique slug (e.g. shapley-additive-explanations) |
compare_techniques
Compare 2-5 techniques across goals, tags, and resources.
| Parameter | Type | Description |
|---|---|---|
slugs | string[] | Array of 2-5 technique slugs |
find_related
Find techniques related via explicit links, shared goals, or shared tags.
| Parameter | Type | Description |
|---|---|---|
slug | string | Technique slug to find relatives for |
depth | number | 1 = explicit + same-goal, 2 = also same-tag (default 1) |
Platform Integration Tools
Tools for connecting techniques to assurance arguments.
suggest_techniques_for_claim
Given an assurance claim, suggest relevant techniques using embedding-based semantic search with hybrid RRF ranking.
| Parameter | Type | Description |
|---|---|---|
claim | string | Assurance claim text (e.g. "the model treats all groups fairly") |
modelType | string | Model type context (e.g. neural-network) |
dataType | string | Data type context (e.g. tabular) |
lifecycleStage | string | Lifecycle stage (e.g. model-development) |
find_evidence_types
Explore what types of evidence techniques can produce, with example techniques for each type.
No parameters required.
Research Tools
Tools for analysing the dataset and finding resources.
explore_taxonomy
Navigate the hierarchical tag taxonomy. Call with no path to see top-level categories.
| Parameter | Type | Description |
|---|---|---|
path | string | Tag path to explore (e.g. applicable-models) |
includeTechniques | boolean | Include technique names for each tag |
coverage_statistics
Analyse dataset coverage across a chosen dimension.
| Parameter | Type | Description |
|---|---|---|
dimension | enum | One of: goals, tags, lifecycle, evidence, model-types, complexity, cross-goal |
search_resources
Search academic resources (papers, software, documentation).
| Parameter | Type | Description |
|---|---|---|
query | string | Free-text search across titles, abstracts, and authors |
type | enum | Filter by type: technical_paper, software_package, documentation, tutorial |
technique | string | Filter by technique slug to get its cited resources |
limit | number | Max results (default 20) |
get_knowledge_graph_summary
High-level statistics: entity counts and relationship counts across the entire graph.
No parameters required.
Data Source
The server fetches from the public knowledge graph:
https://alan-turing-institute.github.io/tea-techniques/data/ld/graph.jsonld
Data is cached locally in ~/.cache/tea-techniques-mcp/ for 24 hours. Use --local to load from project files during development:
npx tsx mcp-server/src/index.ts --local