Dataflow support for TPUs

Google Cloud TPUs are custom-designed AI accelerators created by Google that are optimized for training and using of large AI models. They are designed to scale cost-efficiently for a wide range of AI workloads and provide versatility to accelerate inference workloads on AI frameworks, including PyTorch, JAX, and TensorFlow. For more details about TPUs, see Introduction to Google Cloud TPU.

Prerequisites for using TPUs in Dataflow

  • Your Google Cloud projects must be approved to use this GA offering.

Limitations

This offering is subject to the following limitations:

  • Only single-host TPU accelerators are supported: The Dataflow TPU offering supports only single-host TPU configurations where each Dataflow worker manages one or many TPU devices that are not interconnected with TPUs managed by other workers.
  • Only homogenous TPU worker pools are supported: Features like Dataflow right fitting and Dataflow Prime don't support TPU workloads.

Pricing

Dataflow jobs that use TPUs are billed for worker TPU chip-hours consumed and are not billed for worker CPU and memory. For more information, see the Dataflow pricing page.

Availability

The following TPU accelerators and processing regions are available.

Supported TPU accelerators

The supported TPU accelerator combinations are identified by the tuple (TPU type, TPU topology).

  • TPU type refers to the model of the TPU device.
  • TPU topology refers to the number and physical arrangement of the TPU chips in a slice.

To configure the type and topology of TPUs for Dataflow workers, use the worker_accelerator pipeline option formatted as type:TPU_TYPE;topology:TPU_TOPOLOGY.

The following TPU configurations are supported with Dataflow:

TPU type Topology Required worker_machine_type
tpu-v5-lite-podslice 1x1 ct5lp-hightpu-1t
tpu-v5-lite-podslice 2x2 ct5lp-hightpu-4t
tpu-v5-lite-podslice 2x4 ct5lp-hightpu-8t
tpu-v6e-slice 1x1 ct6e-standard-1t
tpu-v6e-slice 2x2 ct6e-standard-4t
tpu-v6e-slice 2x4 ct6e-standard-8t
tpu-v5p-slice 2x2x1 ct5p-hightpu-4t

Regions

For information about available regions and zones for TPUs, see TPU regions and zones in the Cloud TPU documentation.

What's next