In simple terms, what is the main task of a language model?
Think about what comes next when you read a sentence.
A language model learns to guess the next word by looking at the words before it. This helps computers understand and generate text.
Given the following bigram probabilities, what is the predicted next word after 'I'?
{'I': {'am': 0.6, 'like': 0.3, 'have': 0.1}}Choose the word with the highest probability after 'I'.
The word 'am' has the highest probability (0.6) following 'I', so it is the predicted next word.
You want to build a language model that understands long sentences and context. Which model type is best?
Think about models that remember previous words in a sequence.
RNNs can remember information from earlier words, making them good for understanding long sentences and context.
What is a likely effect of increasing the vocabulary size in a language model?
Think about what happens when the model has to learn more words.
More vocabulary means more parameters and data needed to learn word patterns well, increasing memory and data requirements.
A language model has a perplexity score of 20 on a test set. What does this mean?
Perplexity measures how surprised the model is by the test data.
Perplexity of 20 means the model is as uncertain as picking one word out of 20 equally likely choices on average.