Complete the code to create a Dataproc cluster using the Google Cloud SDK.
gcloud dataproc clusters create [1] --region=us-central1 --single-nodeThe command requires the cluster name after 'create'. 'my-cluster' is a valid example name.
Complete the code to submit a PySpark job to the Dataproc cluster.
gcloud dataproc jobs submit pyspark [1] --cluster=my-cluster --region=us-central1The PySpark job file must be a Python script, so 'job.py' is correct.
Fix the error in the code to list all Dataproc clusters in a region.
gcloud dataproc clusters list --region=[1]The correct region format is 'us-central1' with a hyphen.
Fill both blanks to create a Dataproc cluster with 3 worker nodes and specify the machine type.
gcloud dataproc clusters create my-cluster --region=us-central1 --num-workers=[1] --worker-machine-type=[2]
To create 3 workers, use '3'. For a common machine type, 'n1-standard-4' is suitable.
Fill all three blanks to create a cluster with autoscaling enabled and specify the autoscaling policy.
gcloud dataproc clusters create my-cluster --region=us-central1 --autoscaling-policy=[1] --num-preemptible-workers=[2] --worker-machine-type=[3]
Use your autoscaling policy name, set 2 preemptible workers, and choose a machine type like 'n1-standard-2'.