For summarization tasks, the key metrics are ROUGE scores. ROUGE measures how much the model's summary overlaps with a human-written summary. It checks matching words and phrases to see if the summary captures the important points. ROUGE-1 counts matching single words, ROUGE-2 counts matching pairs of words, and ROUGE-L looks at the longest matching sequence. These metrics matter because summarization is about keeping the main ideas, not just any words.
Summarization with Hugging Face in NLP - Model Metrics & Evaluation
Summarization is a generation task, so confusion matrices don't apply directly. Instead, we use ROUGE scores as a way to compare summaries.
ROUGE-1 (unigram overlap): 0.45
ROUGE-2 (bigram overlap): 0.22
ROUGE-L (longest common subsequence): 0.40
These scores mean the model's summary shares 45% of single words, 22% of word pairs, and 40% of longest sequences with the reference summary.
ROUGE metrics have precision and recall parts:
- Precision: How many words in the model's summary appear in the reference summary? High precision means the summary is focused and mostly relevant.
- Recall: How many words from the reference summary appear in the model's summary? High recall means the summary covers most important points.
Example:
- If a summary is very short but only uses correct words, it has high precision but low recall.
- If a summary is long and covers many points but includes extra unrelated words, it has high recall but lower precision.
Good summarization balances both to keep important info without extra noise.
Good ROUGE scores depend on dataset and task, but generally:
- Good: ROUGE-1 > 0.4, ROUGE-2 > 0.2, ROUGE-L > 0.4 means the summary captures key info well.
- Bad: ROUGE scores below 0.2 suggest the summary misses many important points or is very different from the reference.
Very high scores near 1.0 are rare and may indicate copying the reference summary exactly, which is not always desired.
- Overfitting: Model memorizes training summaries, leading to high ROUGE on training but poor real-world summaries.
- Data leakage: If test summaries appear in training, ROUGE scores will be unrealistically high.
- Ignoring fluency: ROUGE measures overlap but not if the summary reads well or makes sense.
- Length bias: Very short or very long summaries can skew precision or recall.
Your summarization model has ROUGE-1 = 0.65 but ROUGE-2 = 0.10. Is this good? Why or why not?
Answer: The model captures many single words well (high ROUGE-1), but few word pairs (low ROUGE-2). This means it may list important words but not in meaningful phrases. The summary might be disjointed or miss context. So, it is not fully good; improving phrase-level coherence is needed.