Skip to main content

Context recall

Checks if your retrieved context contains the information needed to generate a known correct answer.

Use when: You have ground truth answers and want to verify your retrieval finds supporting evidence.

How it works: Breaks the expected answer into statements and checks if each can be attributed to the context. Score = attributable statements / total statements.

Example:

Expected: "Python was created by Guido van Rossum in 1991"
Context: "Python was released in 1991"
Score: 0.5 (year ✓, creator ✗)

Configuration

assert:
- type: context-recall
value: 'Python was created by Guido van Rossum in 1991'
threshold: 1.0 # Context must support entire answer

Required fields

  • value - Expected answer/ground truth
  • context - Retrieved text (in vars or via contextTransform)
  • threshold - Minimum score 0-1 (default: 0)

Full example

tests:
- vars:
query: 'Who created Python?'
context: 'Guido van Rossum created Python in 1991.'
assert:
- type: context-recall
value: 'Python was created by Guido van Rossum in 1991'
threshold: 1.0

Dynamic context extraction

For RAG systems that return context with their response:

# Provider returns { answer: "...", context: "..." }
assert:
- type: context-recall
value: 'Expected answer here'
contextTransform: 'output.context' # Extract context field
threshold: 0.8

Limitations

  • Binary attribution (no partial credit)
  • Works best with factual statements
  • Requires known correct answers

Further reading