Concept Flow - Google Dataproc overview
User submits job to Dataproc
Dataproc creates a cluster
Cluster runs Apache Spark job
Job processes data on cluster nodes
Results returned to user
User deletes cluster to save cost
This flow shows how a user submits a Spark job to Google Dataproc, which creates a cluster, runs the job, processes data, returns results, and then the cluster is deleted.