Experiment - Output guardrails
Problem:You have a text generation model that sometimes produces unsafe or irrelevant outputs. This can confuse or upset users.
Current Metrics:Safety violations: 15% of outputs contain unsafe content. Relevance score: 70%.
Issue:The model outputs are not reliably safe or relevant, which reduces user trust and satisfaction.