Model Pipeline - Pre-training and fine-tuning concept
This pipeline shows how a large model first learns general knowledge from a big dataset (pre-training), then adapts to a specific task with a smaller dataset (fine-tuning).
This pipeline shows how a large model first learns general knowledge from a big dataset (pre-training), then adapts to a specific task with a smaller dataset (fine-tuning).
Loss
2.3 |*
1.5 | *
0.9 | *
0.6 | *
0.5 | *
0.3 | *
----------------
1 5 10 15 16 20
Epochs| Epoch | Loss ↓ | Accuracy ↑ | Observation |
|---|---|---|---|
| 1 | 2.3 | 0.10 | High loss, low accuracy as model starts learning |
| 5 | 1.5 | 0.45 | Loss decreasing, accuracy improving steadily |
| 10 | 0.9 | 0.75 | Model captures general language patterns well |
| 15 | 0.6 | 0.85 | Pre-training converging with good accuracy |
| 16 | 0.5 | 0.88 | Fine-tuning starts on task-specific data |
| 20 | 0.3 | 0.95 | Fine-tuning improves task accuracy significantly |