Recall & Review
beginner
What is the main purpose of evaluating a fine-tuned model?
To check how well the model performs on new, unseen data after training, ensuring it learned useful patterns and can make accurate predictions.
Click to reveal answer
beginner
Name two common metrics used to evaluate classification models.
Accuracy and F1-score are common metrics. Accuracy measures the percentage of correct predictions, while F1-score balances precision and recall.
Click to reveal answer
intermediate
Why is it important to use a separate test set when evaluating a fine-tuned model?
Using a separate test set helps measure how the model performs on data it has never seen before, preventing overly optimistic results from training data.
Click to reveal answer
intermediate
What does overfitting mean in the context of fine-tuned models?
Overfitting happens when a model learns the training data too well, including noise, and performs poorly on new data.
Click to reveal answer
intermediate
How can you visually inspect a fine-tuned model's performance on classification tasks?
By using a confusion matrix, which shows correct and incorrect predictions for each class, helping identify where the model makes mistakes.
Click to reveal answer
Which metric is best to use when classes are imbalanced?
✗ Incorrect
F1-score balances precision and recall, making it better for imbalanced classes than accuracy.
What does a high training accuracy but low test accuracy usually indicate?
✗ Incorrect
High training accuracy but low test accuracy means the model learned training data too well but fails on new data, which is overfitting.
Which dataset is used to tune model parameters during fine-tuning?
✗ Incorrect
The training set is used to adjust model parameters during fine-tuning.
What is the role of the test set in model evaluation?
✗ Incorrect
The test set is used only to assess the final performance of the model on unseen data.
Which visualization helps understand classification errors?
✗ Incorrect
A confusion matrix shows how many predictions were correct or wrong for each class.
Explain why evaluating a fine-tuned model on unseen data is crucial and describe common metrics used.
Think about how to know if the model learned well beyond just memorizing.
You got /5 concepts.
Describe overfitting in fine-tuned models and how you can detect it using evaluation results.
Consider what happens when a model performs great on training but poorly on new data.
You got /3 concepts.