Name

wb-resource-create-dataproc-cluster - Add a controlled GCP Dataproc cluster resource with Jupyter. For a detailed explanation of parameters, see https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.clusters#Cluster

Synopsis

wb resource create dataproc-cluster [--quiet] [--autoscaling-policy=<autoscalingPolicy>] [--bucket=<configBucket>] [--cluster-id=<clusterId>] [--description=<description>] [--format=<format>] [--idle-delete-ttl=<idleDeleteTtl>] [--image-version=<imageVersion>] [--region=<region>] [--software-framework=<softwareFrameworkType _ >_] [--temp-bucket=<tempBucket>] [--workspace=<id>] [--components=<components>[, <components>…​]]…​ [--initialization-actions=<initializationAct _ ions>_[,<initializationActions>…​]]…​ [-M=<String=String>[, <String=String>…​]]…​ [--properties=<String=String>[, <String=String>…​]]…​ (--id=<id>) [[--manager-machine-type=<machineType>] [--manager-image-uri=<imageUri>] [[--manager-accelerator-type=<type>] [--manager-accelerator-count=<count>]] [[--manager-boot-disk-type=<bootDiskType>] [--manager-boot-disk-size=<bootDiskSizeGb>] [--manager-num-local-ssds=<numLocalSsds>] [--manager-local-ssd-interface=<localSsdInte _ rface>_]]] [[--num-workers=<numNodes>] [--worker-machine-type=<machineType>] [--worker-image-uri=<imageUri>] [[--worker-accelerator-type=<type>] [--worker-accelerator-count=<count>]] [[--worker-boot-disk-type=<bootDiskType>] [--worker-boot-disk-size=<bootDiskSizeGb>] [--worker-num-local-ssds=<numLocalSsds>] [--worker-local-ssd-interface=<localSsdInter _ face>_]]] [[--num-secondary-workers=<numNodes>] [--secondary-worker-machine-type=<machineTyp _ e>_] [--secondary-worker-image-uri=<imageUri>] [--secondary-worker-type=<type>] [[--secondary-worker-accelerator-type=<type> ] [--secondary-worker-accelerator-count=<count _ >_]] [[--secondary-worker-boot-disk-type=<bootDis _ kType>_] [--secondary-worker-boot-disk-size=<bootDisk _ SizeGb>_] [--secondary-worker-num-local-ssds=<numLocal _ Ssds>_] [--secondary-worker-local-ssd-interface=<loc _ alSsdInterface>_]]]

Description

Add a controlled GCP Dataproc cluster resource with Jupyter. For a detailed explanation of parameters, see https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.clusters#Cluster

Options

--id=<id>

ID of the resource, scoped to the workspace. Only use letters, numbers, dashes, and underscores.

--description=<description>

Description of the resource.

--workspace=<id>

Workspace id to use for this command only.

--format=<format>

Set the format for printing command output: JSON, TEXT. Defaults to the config format property.

Default: null
--quiet

Suppress interactive prompt.

--cluster-id=<clusterId>

The unique name to give to the dataproc cluster. Cannot be changed later. The instance name must be 1 to 52 characters long and contain only lowercase letters, numeric characters, and dashes. The first character must be a lowercase letter and the last character cannot be a dash. If not specified, a value will be auto-generated for you.

--region=<region>

The Google Cloud region of the cluster.

--image-version=<imageVersion>

The dataproc cluster image version containing versions of its software components. See https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-version-clusters for the full list of image versions and their bundled software components.

--initialization-actions=<initializationActions>[,<initializationActions>…​]

A comma separated list of initialization scripts to run during cluster creation.The path must be a URL or Cloud Storage path, e.g. 'gs://path-to-file/file-name'.

--components=<components>[,<components>…​]

Comma-separated list of components.

--properties=<String=String>[,<String=String>…​]

Properties in the format key=value.

--software-framework=<softwareFrameworkType>

Software framework for the cluster. Available frameworks are: NONE, HAIL.

Default: NONE
--bucket=<configBucket>

Resource name of the cluster staging bucket. If not specified, a default staging bucket will be created.

--temp-bucket=<tempBucket>

Resource name of the cluster temp bucket. If not specified, a default temp bucket will be created.

--autoscaling-policy=<autoscalingPolicy>

Autoscaling policy id to attach to the cluster.

-M, --metadata=<String=String>[,<String=String>…​]

Custom metadata to apply to this cluster.

specify multiple metadata in the format of --metadata="key1=value1" -Mkey2=value2.

It allows multiple metadata entries split by "," like --metadata=key1=value1,key2=value2

By default, set Workbench CLI server terra-cli-server=[CLI_SERVER_ID]

and the Workbench workspace id (terra-workspace-id=[WORKSPACE_ID]).

--idle-delete-ttl=<idleDeleteTtl>

Time-to-live after which the resource becomes idle and is deleted.

Manager node configurations

--manager-machine-type=<machineType>

The machine type of the manager node.

Default: n2-standard-2
--manager-image-uri=<imageUri>

The image URI for the manager node.

--manager-accelerator-type=<type>

The type of accelerator for the manager.

--manager-accelerator-count=<count>

The count of accelerators for the manager.

Default: 0
--manager-boot-disk-type=<bootDiskType>

The type of boot disk for the manager node.

--manager-boot-disk-size=<bootDiskSizeGb>

The size of the boot disk in GB for the manager node.

Default: 500
--manager-num-local-ssds=<numLocalSsds>

The number of local SSDs for the manager node.

Default: 0
--manager-local-ssd-interface=<localSsdInterface>

The interface type of local SSDs for the manager node.

Default: scsi

Worker node configurations

--num-workers=<numNodes>

The number of worker nodes.

Default: 2
--worker-machine-type=<machineType>

The machine type of the worker node.

Default: n2-standard-2
--worker-image-uri=<imageUri>

The image URI for the worker node.

--worker-accelerator-type=<type>

The type of accelerator for the worker.

--worker-accelerator-count=<count>

The count of accelerators for the worker.

Default: 0
--worker-boot-disk-type=<bootDiskType>

The type of boot disk for the worker node.

--worker-boot-disk-size=<bootDiskSizeGb>

The size of the boot disk in GB for the worker node.

Default: 500
--worker-num-local-ssds=<numLocalSsds>

The number of local SSDs for the worker node.

Default: 0
--worker-local-ssd-interface=<localSsdInterface>

The interface type of local SSDs for the worker node.

Default: scsi

Secondary worker node configurations

--num-secondary-workers=<numNodes>

The number of secondary worker nodes.

Default: 0
--secondary-worker-machine-type=<machineType>

The machine type of the secondary worker node.

Default: n2-standard-2
--secondary-worker-image-uri=<imageUri>

The image URI for the secondary worker node.

--secondary-worker-type=<type>

The type of the secondary worker. Valid values are preemptible, non-preemptible, and spot.

Default: spot
--secondary-worker-accelerator-type=<type>

The type of accelerator for the secondary worker.

--secondary-worker-accelerator-count=<count>

The count of accelerators for the secondary worker.

Default: 0
--secondary-worker-boot-disk-type=<bootDiskType>

The type of boot disk for the secondary worker node.

--secondary-worker-boot-disk-size=<bootDiskSizeGb>

The size of the boot disk in GB for the secondary worker node.

Default: 500
--secondary-worker-num-local-ssds=<numLocalSsds>

The number of local SSDs for the secondary worker node.

Default: 0
--secondary-worker-local-ssd-interface=<localSsdInterface>

The interface type of local SSDs for the secondary worker node.

Default: scsi

Last Modified: 1 January 0001