Which statement best explains why horizontal scaling requires different strategies than vertical scaling in machine learning operations?
Think about how adding more machines versus upgrading one machine affects system design.
Horizontal scaling means adding more machines that must work together, requiring strategies for coordination and data sharing. Vertical scaling means making one machine stronger, which doesn't need coordination.
What is the output of this command when scaling a deployment named 'ml-model' to 3 replicas?
kubectl scale deployment ml-model --replicas=3Scaling a deployment changes the number of pods, not creating or deleting the deployment.
The command scales the existing deployment to the specified number of replicas and confirms with 'deployment.apps/ml-model scaled'.
Put these steps in the correct order to set up auto-scaling for a machine learning model deployment.
Think about deploying first, then monitoring and configuring scaling.
You first deploy the service, then define what to monitor, configure scaling rules, and finally test the setup.
You tried to scale a deployment but the pods remain at the old count. Which issue is most likely causing this?
Think about cluster limits that prevent scaling.
A resource quota can limit how many pods can run, preventing scaling even if the command succeeds.
Which approach is best when scaling a stateful machine learning service that stores user session data?
Think about how to keep user data consistent when scaling.
Sticky sessions ensure users connect to the same pod, preserving session data and avoiding data loss during scaling.