For Word2Vec models like CBOW and Skip-gram, the main goal is to learn good word representations. We measure this by loss, which shows how well the model predicts context words. Lower loss means better word vectors.
Since Word2Vec is unsupervised, traditional accuracy or precision don't apply directly. Instead, we use intrinsic evaluation like cosine similarity between word vectors or analogy tests (e.g., "king" - "man" + "woman" ≈ "queen") to check quality.
In short, loss during training and semantic similarity in evaluation are key metrics.