Which of the following is the most common source of bias in generative AI models?
Think about what influences the model's learned patterns the most.
Bias in generative models mainly comes from the training data. If the data has stereotypes or is unbalanced, the model learns and reproduces those biases.
Given a generative model trained on biased data, what is the most likely output of this prompt?
Prompt: "The nurse said that"
def generate_text(prompt): # Simulated biased output based on stereotypical training data if prompt == "The nurse said that": return "she will take care of you soon." else: return "No output." output = generate_text("The nurse said that")
Consider common gender stereotypes in training data.
The model reflects gender bias by associating nurses with female pronouns, showing bias in output.
Which metric is best suited to measure bias in a generative model's output across different demographic groups?
Look for a metric that compares output distributions between groups.
Demographic Parity Difference measures if the model's outputs are equally likely across groups, indicating bias.
Consider a generative model that produces the following outputs for the prompt "The CEO is":
["a man", "a woman", "a man", "a man", "a woman"]
What is the bias amplification issue here?
Think about how the model exaggerates existing biases.
Bias amplification means the model increases the imbalance seen in training data, here showing more male CEOs than expected.
You want to build a generative model that reduces bias in text generation. Which approach is most effective?
Consider methods that actively reduce bias during learning.
Conditional generation with fairness constraints helps the model learn to avoid biased outputs by design.