0
0
Agentic AIml~20 mins

Embedding models for semantic search in Agentic AI - Practice Problems & Coding Challenges

Choose your learning style9 modes available
Challenge - 5 Problems
🎖️
Semantic Search Embedding Master
Get all challenges correct to earn this badge!
Test your skills under time pressure!
🧠 Conceptual
intermediate
1:30remaining
Understanding Embedding Vectors
Which statement best describes what an embedding vector represents in semantic search?
AA one-hot encoded vector indicating the presence of specific keywords
BA numeric representation capturing the meaning of text in a multi-dimensional space
CA binary vector representing the frequency of words in a document
DA sequence of characters representing the original text
Attempts:
2 left
💡 Hint
Think about how semantic similarity is measured between texts.
Predict Output
intermediate
2:00remaining
Output of Cosine Similarity Calculation
What is the output of this Python code calculating cosine similarity between two embedding vectors?
Agentic AI
import numpy as np

def cosine_similarity(vec1, vec2):
    return np.dot(vec1, vec2) / (np.linalg.norm(vec1) * np.linalg.norm(vec2))

vec_a = np.array([1, 2, 3])
vec_b = np.array([4, 5, 6])
result = cosine_similarity(vec_a, vec_b)
print(round(result, 2))
A0.77
B0.67
C0.97
D0.87
Attempts:
2 left
💡 Hint
Recall the cosine similarity formula and calculate dot product and norms.
Model Choice
advanced
2:30remaining
Choosing an Embedding Model for Semantic Search
You want to build a semantic search system for short product descriptions. Which embedding model is best suited?
AA pretrained sentence-transformer model fine-tuned on product descriptions
BA large transformer-based model trained on general web text
CA small word2vec model trained on product reviews
DA bag-of-words model with TF-IDF weighting
Attempts:
2 left
💡 Hint
Consider model size, training data relevance, and ability to capture sentence meaning.
Hyperparameter
advanced
2:00remaining
Effect of Embedding Dimension Size
What is a likely effect of increasing the embedding vector dimension size in a semantic search model?
AAlways reduces overfitting and improves generalization
BDecreases model accuracy but speeds up similarity calculations
CHas no effect on model performance or computational cost
DImproves semantic representation but increases computational cost and risk of overfitting
Attempts:
2 left
💡 Hint
Think about trade-offs between detail captured and resource use.
🔧 Debug
expert
2:30remaining
Debugging Semantic Search with Embeddings
Given this code snippet for semantic search, what error will it raise when run?
Agentic AI
import numpy as np

embeddings = {'doc1': np.array([0.1, 0.2]), 'doc2': np.array([0.3, 0.4])}
query = np.array([0.1, 0.2, 0.3])

scores = {doc: np.dot(vec, query) for doc, vec in embeddings.items()}
print(scores)
AValueError due to mismatched vector sizes in np.dot
BKeyError because 'query' is not in embeddings
CTypeError because np.dot expects lists, not arrays
DNo error; prints similarity scores
Attempts:
2 left
💡 Hint
Check the shapes of vectors used in dot product.