This visual execution shows how Kubernetes manages ML workloads by deploying containerized models as pods. First, the ML model is packaged into a container image. Then a Deployment YAML specifies how many replicas to run. Applying this YAML creates a Deployment resource. The Kubernetes scheduler assigns pods to nodes, and kubelets start the containers. A Service load balances prediction requests to pods. The Horizontal Pod Autoscaler monitors load and scales pods up or down automatically. When a new model version is available, updating the Deployment triggers a rolling update, replacing old pods with new ones without downtime. Variables like pod count and model version change step-by-step, helping beginners understand the process clearly.