Output guardrails help control what a model says or does. The key metrics to check are accuracy for correctness, precision to avoid wrong or harmful outputs, and recall to ensure important or safe outputs are not missed. For example, in a chatbot, precision helps avoid wrong answers, while recall ensures it answers all questions well.
Output guardrails in Prompt Engineering / GenAI - Model Metrics & Evaluation
| Predicted Safe | Predicted Unsafe |
|----------------|------------------|
| True Safe (TN) | False Unsafe (FP)|
| False Safe (FN)| True Unsafe (TP) |
TP: Model correctly blocks unsafe content.
FP: Model wrongly blocks safe content.
FN: Model wrongly outputs unsafe content.
TN: Model correctly outputs safe content.
Metrics use these counts to measure how well guardrails work.
High precision means the model rarely outputs unsafe content (few false unsafe outputs). This is important to keep users safe.
High recall means the model catches most unsafe content (few unsafe outputs slip through). This is also critical for safety.
But improving one can hurt the other. For example, strict guardrails may block many safe outputs (low recall), while loose guardrails may let unsafe outputs through (low precision).
Finding the right balance depends on the use case and risk tolerance.
- Good: Precision and recall both above 90%, meaning most unsafe outputs are blocked and safe outputs are allowed.
- Bad: Precision below 70%, meaning many unsafe outputs get through, or recall below 70%, meaning many safe outputs are blocked.
- Accuracy alone can be misleading if unsafe content is rare.
- Accuracy paradox: If unsafe outputs are rare, a model that always says safe can have high accuracy but fail safety.
- Data leakage: If test data leaks into training, metrics look better but real safety is worse.
- Overfitting: Guardrails tuned too tightly on test data may fail on new inputs.
- Ignoring context: Metrics must consider context to judge if output is truly safe or unsafe.
Your model has 98% accuracy but only 12% recall on unsafe outputs. Is it good for production?
Answer: No. The model misses 88% of unsafe outputs, which is dangerous. High accuracy here is misleading because unsafe outputs are rare. You need higher recall to catch unsafe content reliably.