0
0
NLPml~5 mins

FastText embeddings in NLP - Cheat Sheet & Quick Revision

Choose your learning style9 modes available
Recall & Review
beginner
What are FastText embeddings?
FastText embeddings are word representations that include information about subwords (small parts of words). This helps capture meanings of rare or new words by looking at their parts.
Click to reveal answer
intermediate
How does FastText differ from traditional word embeddings like Word2Vec?
Unlike Word2Vec, which treats each word as a single unit, FastText breaks words into smaller pieces called n-grams. This helps it understand words it has never seen before by combining these pieces.
Click to reveal answer
beginner
Why are subword embeddings useful in FastText?
Subword embeddings help FastText handle rare words, misspellings, and new words by learning from smaller parts of words. This makes the model more flexible and accurate.
Click to reveal answer
intermediate
What is the role of n-grams in FastText embeddings?
N-grams are sequences of characters inside words. FastText learns embeddings for these n-grams and combines them to form the word's embedding, capturing internal word structure.
Click to reveal answer
advanced
How can FastText embeddings improve performance on languages with rich morphology?
Because FastText uses subword information, it can better understand word variations and endings common in languages with rich morphology, improving representation and downstream task results.
Click to reveal answer
What is the main advantage of FastText embeddings over Word2Vec?
AThey ignore word order
BThey require less training data
CThey only work with fixed vocabulary
DThey use subword information to handle rare words
In FastText, what are n-grams?
AComplete sentences
BSequences of characters inside words
CIndividual words
DParagraphs
Why does FastText perform well on misspelled words?
ABecause it uses a dictionary lookup
BBecause it ignores spelling
CBecause it uses subword parts to build embeddings
DBecause it trains on misspelled words only
Which of these is NOT a feature of FastText embeddings?
AIgnoring word context completely
BHandling out-of-vocabulary words
CUsing character n-grams
DCapturing subword information
FastText embeddings are especially useful for which type of languages?
ALanguages with rich morphology
BLanguages with no grammar
CLanguages with only short words
DLanguages without alphabets
Explain how FastText embeddings use subword information to represent words.
Think about how smaller parts of words help understand the whole word.
You got /4 concepts.
    Describe why FastText embeddings can improve performance on languages with many word forms.
    Consider how word parts change in different forms.
    You got /4 concepts.