Experiment - Gradient accumulation
Problem:Training a neural network on limited GPU memory causes small batch sizes, leading to unstable training and slower convergence.
Current Metrics:Training loss decreases slowly; validation accuracy plateaus around 70% after 10 epochs with batch size 16.
Issue:Batch size is too small due to memory limits, causing noisy gradient updates and slower learning.