Which of the following is NOT a common reason why machine learning models degrade in production?
Think about what usually causes models to perform worse, not operational mistakes.
Model degradation usually happens because the data or environment changes, not because the model code is deleted. Code deletion is an operational error, not a degradation cause.
You run a data drift detection command on your production data. The tool outputs:
Drift detected: Feature 'age' distribution changed significantly (p-value=0.01)
What does this output mean?
Data drift means the input data changes from what the model expects.
A low p-value indicates a significant change in the 'age' feature distribution, which can cause model performance to drop.
Which sequence of steps correctly describes how to handle model degradation caused by data drift?
Think about detecting the problem first, then fixing it step-by-step.
First detect drift, then analyze features, retrain model with new data, and finally deploy the updated model.
Your model's accuracy suddenly dropped in production. Logs show no code changes and data pipeline is running fine. What is the most likely cause?
Think about what can change without code or pipeline errors.
Data distribution shift is a common cause of sudden accuracy drops when code and pipelines are unchanged.
Which practice is the best way to prevent model degradation due to changing data in production?
Think about adapting the model to new data over time.
Regular retraining with fresh data and monitoring helps keep the model accurate as data changes.