You have a general language model and want it to perform well on a specific task with limited data. Which approach is best to start with?
Think about the cost and data needed for fine-tuning versus prompt engineering.
Prompt engineering is usually the first step when data is limited because it does not require changing the model weights and is faster and cheaper. Fine-tuning is better when you have enough data and need more specialized behavior.
Which situation best justifies fine-tuning a large language model instead of relying on prompt engineering?
Consider when changing the model weights is beneficial.
Fine-tuning is preferred when you have enough task-specific data and want the model to consistently perform well on that task, as it adjusts the model weights to specialize it.
You fine-tune a model and also try prompt engineering on the base model. You measure accuracy on a test set. Which metric result indicates fine-tuning improved performance?
base_accuracy = 0.75 fine_tuned_accuracy = 0.82 prompt_engineered_accuracy = 0.78
Compare the accuracy numbers carefully.
Fine-tuning accuracy (0.82) is higher than both base (0.75) and prompt engineered (0.78), showing it improved performance the most.
You fine-tuned a model on a small dataset but test accuracy dropped. What is the most likely cause?
Think about what happens when fine-tuning with little data.
Fine-tuning on a small dataset can cause the model to memorize training data and perform worse on new data, known as overfitting.
You want to build a chatbot that handles many topics without retraining often. Which approach is best?
Consider flexibility and maintenance effort for many topics.
Prompt engineering with a single large model allows dynamic adaptation to many domains without retraining multiple models, making it more scalable and efficient.