Which of the following best describes sampling bias in a dataset?
Think about how the data was gathered and if it truly reflects the whole group you want to study.
Sampling bias happens when the data collected does not fairly represent the entire population, leading to unfair or inaccurate model results.
Given the following Python code calculating statistical parity difference, what is the printed output?
group_0_positive_rate = 0.7 group_1_positive_rate = 0.5 statistical_parity_difference = group_0_positive_rate - group_1_positive_rate print(f"Statistical Parity Difference: {statistical_parity_difference}")
Subtract the smaller positive rate from the larger one.
Statistical parity difference is the difference between positive outcome rates of two groups. Here, 0.7 - 0.5 = 0.2.
You want to reduce bias in a classification task where sensitive attributes affect predictions. Which model approach is best to mitigate bias during training?
Think about a model that tries to hide sensitive information while learning.
Adversarial debiasing trains a model to predict the target while an adversary tries to predict sensitive attributes. This reduces bias by removing sensitive info from predictions.
In a fairness-aware model using a regularization term to penalize bias, which hyperparameter adjustment will most likely reduce bias?
Think about how stronger penalties affect bias in the model.
Increasing regularization strength on bias penalty forces the model to reduce bias more, improving fairness.
You have two groups in your dataset. The true positive rates (TPR) and false positive rates (FPR) for each group are:
- Group A: TPR=0.8, FPR=0.1
- Group B: TPR=0.6, FPR=0.1
Does this model satisfy equalized odds fairness?
Equalized odds requires both TPR and FPR to be equal across groups.
Equalized odds means both true positive rates and false positive rates must be equal for all groups. Here, TPR differs, so fairness is not satisfied.