RAG Evaluation: Measuring Answer Grounding and Retrieval Quality in Agentic AI
RAG Evaluation: Measuring Answer Grounding and Retrieval Quality
Two levels of evaluation
- Retrieval quality: did we fetch the right chunks?
- Answer quality: did the model use them correctly?
Practical metrics
- Hit rate@k
- Context precision
- Groundedness (claim supported by evidence)

