NLP - Word EmbeddingsWhich characteristic of FastText embeddings enables them to generate vectors for words not present in the training data?AThey use one-hot encoding for each wordBThey represent words as bags of character n-gramsCThey rely solely on word frequency countsDThey use fixed-length word hashingCheck Answer
Step-by-Step SolutionSolution:Step 1: Understand FastText's approachFastText breaks words into character n-grams, allowing it to capture subword information.Step 2: Compare with traditional embeddingsTraditional embeddings treat words as atomic units, so unseen words have no vectors.Final Answer:They represent words as bags of character n-grams -> Option BQuick Check:Subword modeling enables embeddings for unseen words [OK]Quick Trick: FastText uses subword units for unseen words [OK]Common Mistakes:MISTAKESAssuming FastText uses one-hot encodingThinking FastText ignores character-level infoBelieving FastText relies only on word frequency
Master "Word Embeddings" in NLP9 interactive learning modes - each teaches the same concept differentlyLearnWhyDeepModelTryChallengeExperimentRecallMetrics
More NLP Quizzes Sentiment Analysis Advanced - Hybrid approaches - Quiz 2easy Sequence Models for NLP - Attention mechanism basics - Quiz 13medium Sequence Models for NLP - Embedding layer usage - Quiz 10hard Sequence Models for NLP - Bidirectional LSTM - Quiz 5medium Text Generation - Temperature and sampling - Quiz 8hard Text Similarity and Search - Jaccard similarity - Quiz 5medium Topic Modeling - LDA with Gensim - Quiz 7medium Topic Modeling - Why topic modeling discovers themes - Quiz 6medium Topic Modeling - LDA with scikit-learn - Quiz 15hard Word Embeddings - Word2Vec (CBOW and Skip-gram) - Quiz 12easy