Summarization methods can be broadly categorized into extractive and abstractive. Which statement correctly describes their main difference?
Think about whether the summary uses original sentences or creates new ones.
Extractive summarization picks important sentences from the original text unchanged. Abstractive summarization rewrites or generates new sentences to convey the main ideas.
Given the following Python code that extracts the first sentence as a summary, what will be printed?
text = "Machine learning helps computers learn from data. It is widely used in AI. Summarization is one application." summary = text.split('.')[0] + '.' print(summary)
Look at how the text is split and which part is selected.
The code splits the text at periods and selects the first part, then adds a period. So it prints the first sentence only.
You want to build an abstractive summarization system that generates new sentences. Which model architecture is most appropriate?
Consider models that can generate sequences from input sequences.
Sequence-to-sequence models with attention can generate new text based on input sequences, making them ideal for abstractive summarization.
When evaluating how good a summary is compared to a reference summary, which metric is most commonly used?
Think about metrics that compare overlap of words or phrases.
ROUGE measures overlap of n-grams between generated and reference summaries, making it standard for summarization evaluation.
Consider a trained abstractive summarization model that often repeats the same phrase multiple times in its output. What is the most likely cause?
Think about how models generate sequences and avoid repeating themselves.
Repetitive output often happens when the decoding strategy does not include mechanisms like coverage or diversity to avoid repeating the same phrases.