Recall & Review
beginner
What is the role of temperature in sampling from a language model?
Temperature controls how random or focused the model's predictions are. A low temperature (<1) makes the model more confident and conservative, picking high-probability words. A high temperature (>1) makes the model more creative and random by flattening the probabilities.
Click to reveal answer
beginner
Explain sampling in the context of generating text from a language model.
Sampling means picking the next word based on the model's predicted probabilities instead of always choosing the most likely word. This adds variety and creativity to the generated text.
Click to reveal answer
intermediate
How does increasing temperature affect the probability distribution during sampling?
Increasing temperature makes the probability distribution more even, so less likely words have a higher chance to be picked. This leads to more diverse and surprising outputs.
Click to reveal answer
intermediate
What happens if temperature is set to 0 during sampling?
Setting temperature to 0 means always picking the word with the highest probability (greedy decoding). This removes randomness and can make the output repetitive or dull.
Click to reveal answer
beginner
Why might you want to use a moderate temperature (e.g., 0.7) instead of very low or very high?
A moderate temperature balances creativity and coherence. It allows some randomness for interesting text but keeps the output sensible and relevant.
Click to reveal answer
What does a temperature of 1.0 mean when sampling from a language model?
✗ Incorrect
A temperature of 1.0 means the model uses the original predicted probabilities as they are, without scaling.
What is the effect of setting temperature to a very high value (e.g., 5)?
✗ Incorrect
High temperature flattens the probability distribution, increasing randomness and diversity in output.
Which sampling method removes randomness completely?
✗ Incorrect
Greedy decoding always picks the highest probability word, removing randomness.
Why is sampling preferred over always picking the highest probability word?
✗ Incorrect
Sampling adds randomness, making generated text more varied and interesting.
What does lowering temperature below 1 do to the output?
✗ Incorrect
Lower temperature sharpens the probability distribution, making the model pick more likely words.
Describe how temperature affects the randomness of text generated by a language model.
Think about how temperature changes the chance of picking less likely words.
You got /4 concepts.
Explain why sampling is used instead of always choosing the most likely word in text generation.
Consider how always picking the top word might affect the text.
You got /4 concepts.