Which of the following best describes a common source of technical debt in machine learning pipelines?
Think about what causes maintenance problems over time in ML systems.
Using outdated data schemas causes components to break or behave unexpectedly, creating technical debt.
What is the output of the following command when run in an MLflow tracking server with two registered model versions?
mlflow models list-versions --model-name my_model
Consider what MLflow shows when models have versions in different stages.
The command lists all versions of the model with their stages, showing production and staging versions.
Given this simplified ML deployment workflow, which step introduces the most technical debt?
Steps:
- Data collection
- Manual feature engineering without documentation
- Model training with fixed hyperparameters
- Deployment without automated monitoring
Think about what makes future changes and debugging harder.
Manual feature engineering without documentation makes it hard to maintain or update features, increasing technical debt.
An ML system uses automated drift detection but fails to alert when the input data distribution changes. What is the most likely cause?
Consider what would prevent drift detection from noticing changes.
If drift detection monitors wrong features, it won't detect actual distribution changes, causing missed alerts.
Which practice most effectively reduces technical debt in ML systems over time?
Think about automation and testing to catch issues early.
CI/CD with automated tests ensures changes are validated and reduces hidden errors, lowering technical debt.