When using spaCy models for tasks like text classification or named entity recognition, the key metrics to watch are Precision, Recall, and F1-score. These metrics tell us how well the model finds the right information in text.
Precision shows how many of the model's positive predictions are actually correct. Recall shows how many of the actual positive cases the model found. F1-score balances both precision and recall into one number.
We focus on these because spaCy models often work with unbalanced data, like rare entities or categories, where accuracy alone can be misleading.