Bird
0
0

Which metric is best suited for evaluating the quality of a summary generated by a model?

easy📝 Conceptual Q2 of 15
NLP - Text Generation
Which metric is best suited for evaluating the quality of a summary generated by a model?
ABLEU
BAccuracy
CROUGE
DPrecision
Step-by-Step Solution
Solution:
  1. Step 1: Identify metric purpose

    ROUGE is designed to evaluate summaries by measuring overlap of sequences like longest common subsequence.
  2. Step 2: Compare with other metrics

    BLEU is better for translation; accuracy and precision are classification metrics, not for summaries.
  3. Final Answer:

    ROUGE -> Option C
  4. Quick Check:

    Summary evaluation uses ROUGE = C [OK]
Quick Trick: Use ROUGE for summaries, BLEU for translations [OK]
Common Mistakes:
MISTAKES
  • Using BLEU for summary evaluation
  • Confusing classification metrics with text evaluation
  • Ignoring ROUGE's design for summaries

Want More Practice?

15+ quiz questions · All difficulty levels · Free

Free Signup - Practice All Questions
More NLP Quizzes