0
0
NLPml~20 mins

Evaluation metrics (accuracy, F1, confusion matrix) in NLP - Practice Problems & Coding Challenges

Choose your learning style9 modes available
Challenge - 5 Problems
🎖️
Evaluation Metrics Master
Get all challenges correct to earn this badge!
Test your skills under time pressure!
Predict Output
intermediate
2:00remaining
Calculate accuracy from predictions and true labels
Given the true labels and predicted labels below, what is the accuracy of the model?
NLP
true_labels = [1, 0, 1, 1, 0, 0, 1]
pred_labels = [1, 0, 0, 1, 0, 1, 1]
correct = sum(t == p for t, p in zip(true_labels, pred_labels))
accuracy = correct / len(true_labels)
print(round(accuracy, 2))
A0.85
B0.57
C0.71
D0.43
Attempts:
2 left
💡 Hint
Count how many predictions match the true labels, then divide by total labels.
🧠 Conceptual
intermediate
1:30remaining
Understanding the F1 score
Which statement best describes the F1 score in classification tasks?
AIt is the average of accuracy and recall.
BIt measures the ratio of correct predictions to total predictions.
CIt counts the number of true negatives in the confusion matrix.
DIt is the harmonic mean of precision and recall, balancing both.
Attempts:
2 left
💡 Hint
Think about how F1 combines two important metrics to balance false positives and false negatives.
Predict Output
advanced
2:00remaining
Confusion matrix values from predictions
Given the true and predicted labels below, what is the value of True Positives (TP) in the confusion matrix?
NLP
true_labels = [0, 1, 1, 0, 1, 0, 1, 1]
pred_labels = [0, 1, 0, 0, 1, 1, 1, 0]
TP = sum(1 for t, p in zip(true_labels, pred_labels) if t == 1 and p == 1)
print(TP)
A3
B2
C4
D1
Attempts:
2 left
💡 Hint
Count how many times the true label and predicted label are both 1.
Model Choice
advanced
1:30remaining
Choosing metric for imbalanced data
You have a dataset where 95% of samples belong to class A and 5% to class B. Which metric is best to evaluate your model's performance on class B?
AF1 score
BLoss value
CAccuracy
DTotal number of predictions
Attempts:
2 left
💡 Hint
Think about which metric balances false positives and false negatives well for rare classes.
Metrics
expert
2:30remaining
Calculate macro F1 score from class-wise precision and recall
Given the following precision and recall for three classes, what is the macro F1 score? Class 1: precision=0.8, recall=0.6 Class 2: precision=0.7, recall=0.7 Class 3: precision=0.9, recall=0.5
A0.72
B0.67
C0.70
D0.75
Attempts:
2 left
💡 Hint
Calculate F1 for each class, then average them.