Recall & Review
beginner
Why is it important to split your dataset into training, validation, and test sets?
Splitting the dataset helps to train the model on one part (training), tune parameters on another (validation), and finally evaluate performance on unseen data (test) to avoid overfitting and get a realistic measure of how the model will perform in real life.
Click to reveal answer
beginner
What does 'overfitting' mean in model evaluation?
Overfitting happens when a model learns the training data too well, including noise and details that don't generalize. This causes poor performance on new, unseen data.
Click to reveal answer
intermediate
What is the purpose of using metrics like accuracy, precision, recall, and F1-score in computer vision?
These metrics help measure how well the model predicts. Accuracy shows overall correctness, precision measures how many predicted positives are true, recall shows how many actual positives were found, and F1-score balances precision and recall.
Click to reveal answer
intermediate
Why should you use cross-validation in model evaluation?
Cross-validation splits data into multiple parts and trains/tests the model several times. This gives a better estimate of model performance by reducing bias from a single train-test split.
Click to reveal answer
beginner
What is the difference between validation data and test data?
Validation data is used during model training to tune parameters and make decisions. Test data is kept separate and used only once at the end to evaluate the final model's performance.
Click to reveal answer
What is the main goal of splitting data into training and test sets?
✗ Incorrect
Splitting data allows us to test the model on data it hasn't seen before, which shows how well it will perform in real situations.
Which metric balances precision and recall in classification tasks?
✗ Incorrect
F1-score combines precision and recall into a single metric to balance false positives and false negatives.
What does overfitting cause in a model?
✗ Incorrect
Overfitting means the model fits training data too closely and fails to generalize to new data.
Why is cross-validation useful?
✗ Incorrect
Cross-validation tests the model multiple times on different data splits to give a more reliable performance estimate.
When should you use the test dataset?
✗ Incorrect
Test data is used only once after training to check how well the model performs on unseen data.
Explain why splitting data into training, validation, and test sets is important in model evaluation.
Think about how each set helps the model learn and be tested fairly.
You got /5 concepts.
Describe the difference between precision and recall and why both are important in evaluating a computer vision model.
Consider how mistakes in predictions affect model usefulness.
You got /4 concepts.