0
0
Prompt Engineering / GenAIml~5 mins

Evaluation of fine-tuned models in Prompt Engineering / GenAI - Cheat Sheet & Quick Revision

Choose your learning style9 modes available
Recall & Review
beginner
What is the main purpose of evaluating a fine-tuned model?
To check how well the model performs on new, unseen data after training, ensuring it learned useful patterns and can make accurate predictions.
Click to reveal answer
beginner
Name two common metrics used to evaluate classification models.
Accuracy and F1-score are common metrics. Accuracy measures the percentage of correct predictions, while F1-score balances precision and recall.
Click to reveal answer
intermediate
Why is it important to use a separate test set when evaluating a fine-tuned model?
Using a separate test set helps measure how the model performs on data it has never seen before, preventing overly optimistic results from training data.
Click to reveal answer
intermediate
What does overfitting mean in the context of fine-tuned models?
Overfitting happens when a model learns the training data too well, including noise, and performs poorly on new data.
Click to reveal answer
intermediate
How can you visually inspect a fine-tuned model's performance on classification tasks?
By using a confusion matrix, which shows correct and incorrect predictions for each class, helping identify where the model makes mistakes.
Click to reveal answer
Which metric is best to use when classes are imbalanced?
AAccuracy
BF1-score
CTraining loss
DEpoch count
What does a high training accuracy but low test accuracy usually indicate?
AGood generalization
BUnderfitting
CData leakage
DOverfitting
Which dataset is used to tune model parameters during fine-tuning?
ATraining set
BValidation set
CTest set
DUnlabeled data
What is the role of the test set in model evaluation?
ATo assess final model performance
BTo tune hyperparameters
CTo train the model
DTo generate synthetic data
Which visualization helps understand classification errors?
AScatter plot
BLoss curve
CConfusion matrix
DHistogram
Explain why evaluating a fine-tuned model on unseen data is crucial and describe common metrics used.
Think about how to know if the model learned well beyond just memorizing.
You got /5 concepts.
    Describe overfitting in fine-tuned models and how you can detect it using evaluation results.
    Consider what happens when a model performs great on training but poorly on new data.
    You got /3 concepts.