Compute profile configuration options

Options for configuring the compute profile of a cloud environment

Prior reading: Cloud environments overview

Purpose: This document provides detailed instructions for customizing the compute resources allocated to a cloud environment through the web UI.

Note: This article does not apply to Dataproc cluster cloud environments. To learn more about configuring a Dataproc cluster environment, please see Use Dataproc (Spark) and Hail on Verily Workbench.



Introduction

The cost of running your cloud environment primarily depends on the following factors:

  • virtual machine type
  • number and type of CPUs
  • number and type of GPUs
  • memory
  • disk storage

These values are configurable via the Workbench UI, the Workbench CLI, and/or the Google Cloud console.

If you’re creating a new cloud environment via the UI, you can view the Cost estimate on the right side of the Creating cloud environment dialog. The cost will dynamically change as you update the number of CPUs and/or GPUs.

Screenshot of the Choose configuration step in the Creating cloud environment dialog.
View the dynamic Cost estimate depending on the selected configuration.

Once a cloud environment has been created, you can use the Google Cloud console to make configuration changes and view pricing summaries.

Virtual machines

A virtual machine (VM) is an emulation of a physical computer and can perform the same functions, such as running applications and other software. VM instances are created from pools of resources in cloud data centers. You can specify your VM’s geographic region, compute, and storage resources to meet your job’s requirements.

You can create and destroy virtual machines at will to run applications of interest to you, such as interactive analysis environments or data processing algorithms. Virtual machines underlie Verily Workbench’s cloud environments.

By default, Workbench uses N1 high-memory machine types. To learn more about them, click here.

If you create a cloud environment via the wb resource create CLI command, you can specify a different VM type via the --machine-type option.

You can also change the VM type via the Google Cloud console after creating a cloud environment. Note that your VM options may be limited if your environment uses GPUs.

Screenshot of the Hardware tab in Google Cloud console.
View the different types of VMs available and cost estimates in the Google Cloud console.

CPUs (central processing units)

The central processing unit (CPU), or simply processor, can be considered the “brain” of a computer. Every computational machine will have at least one CPU, which is connected to every part of the computer system. It’s the operational center that receives, executes, and delegates the instructions received from programs. CPUs also handle computational calculations and logic. Increasing the number of CPUs accelerates the processing of these tasks. Other types of processors (GPUs or TPUs) may be better suited for processing specialized tasks, such as parallel computing and machine learning.

New cloud environments based off of JupyterLab, R Analysis Environment, and Visual Studio Code apps will include an N1 high-memory VM, two CPUs, and 13 GB of total memory by default.

If you think you’ll need more than two CPUs, you can add them via the Workbench UI at the time of creation.

If you need to add additional CPUs after you’ve created the cloud environment, you can update the compute profile via Google Cloud console as described above.

GPUs (graphical processing units)

A graphical processing unit (GPU) is a specialized processor that excels at parallel computing, which means processing many tasks at once. While a central processing unit (CPU) must process tasks one by one, a GPU can split complex tasks into many pieces and work through them simultaneously. GPUs were traditionally used to accelerate video rendering, but their parallel computing ability also makes them ideal for tasks such as sequence alignment, AI, and machine learning.

If you’re creating a cloud environment via the Workbench UI, there are several types of GPUs available for cloud environments using PyTorch or TensorFlow Enterprise environment images:

  • NVIDIA Tesla P4
  • NVIDIA Tesla T4
  • NVIDIA Tesla V100

GPUs are not included in new cloud environments by default. However, they can be added when you create your cloud environment via the Workbench UI or Workbench CLI. See Create a notebook environment for creating an environment via the Workbench CLI. You can also add GPUs via the Google Cloud console after the environment’s been created.

You can learn more about GPUs on Google Compute Engine here, and see more details about GPU pricing here.

Memory

Memory, also known as random access memory (RAM), is where programs and data that are currently in use are temporarily stored. The central processing unit (CPU) receives instructions and data from programs, which are kept in the computer’s memory while being used. Once the instructions are completed, or the program is no longer in use, the memory is freed up. If the computer system doesn’t have enough memory for all of the CPU’s instructions, the system’s performance will diminish and slow down. While the CPU is commonly thought of as a computer’s brain, you can think of memory as the attention span.

The memory amount automatically scales when the number of CPUs is changed.

Disk storage

By default, your cloud environment’s data disk size will be 100 GB for Vertex AI Workbench instances and 500 GB for Compute Engine instances.

You can change the size when creating a new cloud environment via the UI. The data disk can be as small as 10 GB, but the minimum recommended size is 100 GB. The maximum allowed is 64,000 GB (64 TB).

If you select a shared configuration template in the Choose configuration step, the environment will inherit the specified data disk size. However, this amount can be updated in the Customize step.

Screenshot of Choose configuration dialog showing two configuration template options.
View the specifications for a shared configuration template.

If you’re creating a new cloud environment via the CLI, you can change this value via the --data-disk-size option.

Autostop

Your cloud environment includes an autostop feature that will automatically stop running environments after a set idle time. The default idle time is 4 hours, but you can change it to any length between 1 hour and 14 days. This can be configured at the time of cloud environment creation or updated after creation via the Workbench UI or Workbench CLI. You can also choose to disable the autostop feature.

Screenshot of Customize dialog showing environment image options, cloud compute profile, data disk size, and autostop.
Set the environment image type, CPUs, GPUs, data disk size, and autostop during cloud environment creation.

Last Modified: 16 September 2024