In few-shot learning with prompts, the model sees very few examples before making predictions. Because of this, accuracy is important to see how often the model gets the right answer. But accuracy alone can be misleading if classes are unbalanced.
Therefore, precision and recall are also key. Precision tells us how many predicted answers are actually correct, and recall tells us how many correct answers the model finds out of all possible correct ones.
Since few-shot learning often deals with limited data, F1 score is very useful. It balances precision and recall into one number, showing overall quality.
For tasks like classification, a confusion matrix helps visualize where the model makes mistakes.