Imagine you are training a model to recognize cats in photos. Why does the quality of the annotations (labels) matter?
Think about what happens if the model learns from wrong labels.
Good annotations help the model learn correct patterns. Bad labels cause confusion and reduce accuracy.
You have two annotators labeling images for object detection. Which metric best measures how much they agree on the labels?
Look for a metric that measures agreement between annotators.
Cohen's Kappa measures agreement beyond chance between two annotators.
What is the output of this Python code that calculates annotation agreement?
from sklearn.metrics import cohen_kappa_score labels_annotator1 = [1, 0, 1, 1, 0] labels_annotator2 = [1, 0, 0, 1, 0] kappa = cohen_kappa_score(labels_annotator1, labels_annotator2) print(round(kappa, 2))
Calculate agreement considering chance agreement.
The Cohen's Kappa score for these labels is 0.62, indicating moderate agreement.
You have a dataset with some incorrect labels. Which model approach is best to reduce the impact of noisy annotations?
Think about models that can avoid overfitting noisy data.
Dropout and early stopping help neural networks avoid fitting noise, improving robustness to bad labels.
After training a classifier, you notice low accuracy. You suspect annotation errors. Which step below will best help identify if annotation quality caused the problem?
Think about how to verify if labels are correct.
Manually reviewing annotations helps find label mistakes that may cause low accuracy.