Experiment - Model optimization (distillation, quantization)
Problem:You have a large NLP model that performs well on text classification but is too slow and large for deployment on mobile devices.
Current Metrics:Training accuracy: 95%, Validation accuracy: 93%, Model size: 500MB, Inference time per sample: 500ms
Issue:The model is too large and slow, making it unsuitable for mobile deployment despite good accuracy.