Experiment - Top-p and top-k sampling
Problem:You have a language model that generates text by picking the next word from a list of possible words with probabilities. Currently, it picks the word with the highest probability every time (greedy sampling). This makes the text boring and repetitive.
Current Metrics:Text diversity is low, with many repeated phrases. Perplexity is 15.0 on the validation set.
Issue:The model's output is too predictable and lacks creativity because it always picks the most likely next word.