Default: null
Name
wb-resource-create-dataproc-cluster - Add a controlled GCP Dataproc cluster resource with Jupyter. For a detailed explanation of parameters, see https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.clusters#Cluster
Synopsis
wb resource create dataproc-cluster [--quiet] [--autoscaling-policy=<autoscalingPolicy>] [--bucket=<configBucket>] [--cluster-id=<clusterId>] [--description=<description>] [--format=<format>] [--idle-delete-ttl=<idleDeleteTtl>] [--image-version=<imageVersion>] [--region=<region>] [--software-framework=<softwareFrameworkType _ >_] [--temp-bucket=<tempBucket>] [--workspace=<id>] [--components=<components>[, <components>…]]… [--initialization-actions=<initializationAct _ ions>_[,<initializationActions>…]]… [-M=<String=String>[, <String=String>…]]… [--properties=<String=String>[, <String=String>…]]… (--id=<id>) [[--manager-machine-type=<machineType>] [--manager-image-uri=<imageUri>] [[--manager-accelerator-type=<type>] [--manager-accelerator-count=<count>]] [[--manager-boot-disk-type=<bootDiskType>] [--manager-boot-disk-size=<bootDiskSizeGb>] [--manager-num-local-ssds=<numLocalSsds>] [--manager-local-ssd-interface=<localSsdInte _ rface>_]]] [[--num-workers=<numNodes>] [--worker-machine-type=<machineType>] [--worker-image-uri=<imageUri>] [[--worker-accelerator-type=<type>] [--worker-accelerator-count=<count>]] [[--worker-boot-disk-type=<bootDiskType>] [--worker-boot-disk-size=<bootDiskSizeGb>] [--worker-num-local-ssds=<numLocalSsds>] [--worker-local-ssd-interface=<localSsdInter _ face>_]]] [[--num-secondary-workers=<numNodes>] [--secondary-worker-machine-type=<machineTyp _ e>_] [--secondary-worker-image-uri=<imageUri>] [--secondary-worker-type=<type>] [[--secondary-worker-accelerator-type=<type> ] [--secondary-worker-accelerator-count=<count _ >_]] [[--secondary-worker-boot-disk-type=<bootDis _ kType>_] [--secondary-worker-boot-disk-size=<bootDisk _ SizeGb>_] [--secondary-worker-num-local-ssds=<numLocal _ Ssds>_] [--secondary-worker-local-ssd-interface=<loc _ alSsdInterface>_]]]
Description
Add a controlled GCP Dataproc cluster resource with Jupyter. For a detailed explanation of parameters, see https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.clusters#Cluster
Options
- --id=<id>
-
ID of the resource, scoped to the workspace. Only use letters, numbers, dashes, and underscores.
- --description=<description>
-
Description of the resource.
- --workspace=<id>
-
Workspace id to use for this command only.
- --format=<format>
-
Set the format for printing command output: JSON, TEXT. Defaults to the config format property.
- --quiet
-
Suppress interactive prompt.
- --cluster-id=<clusterId>
-
The unique name to give to the dataproc cluster. Cannot be changed later. The instance name must be 1 to 52 characters long and contain only lowercase letters, numeric characters, and dashes. The first character must be a lowercase letter and the last character cannot be a dash. If not specified, a value will be auto-generated for you.
- --region=<region>
-
The Google Cloud region of the cluster.
- --image-version=<imageVersion>
-
The dataproc cluster image version containing versions of its software components. See https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-version-clusters for the full list of image versions and their bundled software components.
- --initialization-actions=<initializationActions>[,<initializationActions>…]
-
A comma separated list of initialization scripts to run during cluster creation.The path must be a URL or Cloud Storage path, e.g. 'gs://path-to-file/file-name'.
- --components=<components>[,<components>…]
-
Comma-separated list of components.
- --properties=<String=String>[,<String=String>…]
-
Properties in the format key=value.
- --software-framework=<softwareFrameworkType>
-
Software framework for the cluster. Available frameworks are: NONE, HAIL.
Default: NONE
- --bucket=<configBucket>
-
Resource name of the cluster staging bucket. If not specified, a default staging bucket will be created.
- --temp-bucket=<tempBucket>
-
Resource name of the cluster temp bucket. If not specified, a default temp bucket will be created.
- --autoscaling-policy=<autoscalingPolicy>
-
Autoscaling policy id to attach to the cluster.
- -M, --metadata=<String=String>[,<String=String>…]
-
Custom metadata to apply to this cluster.
specify multiple metadata in the format of --metadata="key1=value1" -Mkey2=value2.
It allows multiple metadata entries split by "," like --metadata=key1=value1,key2=value2
By default, set Workbench CLI server terra-cli-server=[CLI_SERVER_ID]
and the Workbench workspace id (terra-workspace-id=[WORKSPACE_ID]).
- --idle-delete-ttl=<idleDeleteTtl>
-
Time-to-live after which the resource becomes idle and is deleted.
Manager node configurations
- --manager-machine-type=<machineType>
-
The machine type of the manager node.
Default: n2-standard-2
- --manager-image-uri=<imageUri>
-
The image URI for the manager node.
- --manager-accelerator-type=<type>
-
The type of accelerator for the manager.
- --manager-accelerator-count=<count>
-
The count of accelerators for the manager.
Default: 0
- --manager-boot-disk-type=<bootDiskType>
-
The type of boot disk for the manager node.
- --manager-boot-disk-size=<bootDiskSizeGb>
-
The size of the boot disk in GB for the manager node.
Default: 500
- --manager-num-local-ssds=<numLocalSsds>
-
The number of local SSDs for the manager node.
Default: 0
- --manager-local-ssd-interface=<localSsdInterface>
-
The interface type of local SSDs for the manager node.
Default: scsi
Worker node configurations
- --num-workers=<numNodes>
-
The number of worker nodes.
Default: 2
- --worker-machine-type=<machineType>
-
The machine type of the worker node.
Default: n2-standard-2
- --worker-image-uri=<imageUri>
-
The image URI for the worker node.
- --worker-accelerator-type=<type>
-
The type of accelerator for the worker.
- --worker-accelerator-count=<count>
-
The count of accelerators for the worker.
Default: 0
- --worker-boot-disk-type=<bootDiskType>
-
The type of boot disk for the worker node.
- --worker-boot-disk-size=<bootDiskSizeGb>
-
The size of the boot disk in GB for the worker node.
Default: 500
- --worker-num-local-ssds=<numLocalSsds>
-
The number of local SSDs for the worker node.
Default: 0
- --worker-local-ssd-interface=<localSsdInterface>
-
The interface type of local SSDs for the worker node.
Default: scsi
Secondary worker node configurations
- --num-secondary-workers=<numNodes>
-
The number of secondary worker nodes.
Default: 0
- --secondary-worker-machine-type=<machineType>
-
The machine type of the secondary worker node.
Default: n2-standard-2
- --secondary-worker-image-uri=<imageUri>
-
The image URI for the secondary worker node.
- --secondary-worker-type=<type>
-
The type of the secondary worker. Valid values are preemptible, non-preemptible, and spot.
Default: spot
- --secondary-worker-accelerator-type=<type>
-
The type of accelerator for the secondary worker.
- --secondary-worker-accelerator-count=<count>
-
The count of accelerators for the secondary worker.
Default: 0
- --secondary-worker-boot-disk-type=<bootDiskType>
-
The type of boot disk for the secondary worker node.
- --secondary-worker-boot-disk-size=<bootDiskSizeGb>
-
The size of the boot disk in GB for the secondary worker node.
Default: 500
- --secondary-worker-num-local-ssds=<numLocalSsds>
-
The number of local SSDs for the secondary worker node.
Default: 0
- --secondary-worker-local-ssd-interface=<localSsdInterface>
-
The interface type of local SSDs for the secondary worker node.
Default: scsi
Last Modified: 1 January 0001