Which statement best describes the main advantage of using spot instances in cloud-based machine learning workflows?
Think about cost versus reliability trade-offs in cloud compute options.
Spot instances are cheaper but can be stopped by the cloud provider at any time. They are best for jobs that can handle interruptions, like batch training, not for critical real-time tasks.
Given the command kubectl top pods --namespace=ml-training, what output will you see?
kubectl top pods --namespace=ml-training
Consider what 'top pods' command shows in Kubernetes.
The command shows resource usage (CPU and memory) for pods in the specified namespace. If metrics-server is missing, it errors out.
You want to reduce costs by automatically scaling your training cluster based on GPU usage. Which Kubernetes resource should you configure?
Think about scaling nodes, not just pods, when GPU resources are involved.
Cluster Autoscaler adjusts the number of nodes in the cluster based on resource requests like GPUs, helping optimize costs by adding/removing nodes as needed.
Your cloud bill suddenly increased after deploying a new ML pipeline. Which tool can help you identify which pipeline step caused the cost spike?
Think about tools that analyze spending by resource usage or tags.
Cost explorer tools let you break down cloud spending by tags or resources, helping pinpoint which pipeline step caused the spike.
Which practice best helps prevent unexpected high costs in a large-scale MLOps environment?
Think about proactive cost management and limits.
Budget alerts notify teams when spending approaches limits, and resource quotas prevent teams from using more resources than allowed, controlling costs effectively.