Design an optimal storage strategy for your cloud workload

Last reviewed 2024-10-22 UTC

This guide helps you assess the storage requirements of your cloud workload, understand the available storage options in Google Cloud, and design a storage strategy that provides optimal business value.

For a visual summary of the main design recommendations, see the decision tree diagram.

For information about selecting storage services for AI and ML workloads, see Design storage for AI and ML workloads in Google Cloud.

Overview of the design process

As a cloud architect, when you plan storage for a cloud workload, you need to first consider the functional characteristics of the workload, security constraints, resilience requirements, performance expectations, and cost goals. Next, you need to review the available storage services and features in Google Cloud. Then, based on your requirements and the available options, you select the storage services and features that you need. The following diagram shows this three-phase design process:

Phased approach to designing storage for cloud workloads.

Define your requirements

Use the questionnaires in this section to define the key storage requirements of the workload that you want to deploy in Google Cloud.

Guidelines for defining storage requirements

When answering the questionnaires, consider the following guidelines:

  • Define requirements granularly

    For example, if your application needs Network File System (NFS)-based file storage, identify the required NFS version.

  • Consider future requirements

    For example, your current deployment might serve users in countries within Asia, but you might plan to expand the business to other continents. In this case, consider any storage-related regulatory requirements of the new business territories.

  • Consider cloud-specific opportunities and requirements

    • Take advantage of cloud-specific opportunities.

      For example, to optimize the storage cost for data stored in Cloud Storage, you can control the storage duration by using data retention policies and lifecycle configurations.

    • Consider cloud-specific requirements.

      For example, the on-premises data might exist in a single data center, and you might need to replicate the migrated data across two Google Cloud locations for redundancy.

Questionnaires

The questionnaires that follow are not exhaustive checklists for planning. Use them as a starting point to systematically analyze all the storage requirements of the workload that you want to deploy to Google Cloud.

Assess your workload's characteristics

  • What kind of data do you need to store?

    • Static website content
    • Backups and archives for disaster recovery
    • Audit logs for compliance
    • Large data objects that users download directly
    • Transactional data
    • Unstructured, and heterogeneous data

  • How much capacity do you need? Consider your current and future requirements.

  • Should capacity scale automatically with usage?

  • What are the access requirements? For example, should the data be accessible from outside Google Cloud?

  • What are the expected read-write patterns?

    • Frequent writes and reads
    • Frequent writes, but occasional reads
    • Occasional writes and reads
    • Occasional writes, but frequent reads

  • Does the workload need file-based access, using NFS for example?

  • Should multiple clients be able to read or write data simultaneously?

Identify security constraints

  • What are your data-encryption requirements? For example, do you need to use keys that you control?

  • Are there any data-residency requirements?

Define data-resilience requirements

  • Does your workload need low-latency caching or scratch space?
  • Do you need to replicate the data in the cloud for redundancy?
  • Do you need strict read-write consistency for replicated datasets?

Set performance expectations

  • What is the required I/O rate?

  • What levels of read and write throughput does your application need?

  • What environments do you need storage for? For a given workload, you might need high-performance storage for the production environment, but could choose a lower-performance option for the non-production environments.

Review the storage options

Google Cloud offers storage services for all the key storage formats: block, file, and object. Review and evaluate the features, design options, and relative advantages of the services available for each storage format.

Overview

The data that you store in block storage is divided into chunks, each stored as a separate block with a unique address. Applications access data by referencing the appropriate block addresses. Block storage is optimized for high-IOPS workloads, such as transaction processing. It's similar to on-premises storage area network (SAN) and directly attached storage (DAS) systems.

The block storage options in Google Cloud are a part of the Compute Engine service.

Option Overview
Persistent Disk Dedicated hard-disk drives (HDD) and solid-state drives (SSD) for enterprise and database applications deployed to Compute Engine VMs and Google Kubernetes Engine (GKE) clusters.
Google Cloud Hyperdisk Fast and redundant network storage for Compute Engine VMs, with configurable performance and volumes that can be dynamically resized.
Local SSD Ephemeral, locally attached block storage for high-performance applications.

Data is organized and represented in a hierarchy of files that are stored in folders, similar to on-premises network-attached storage (NAS). File systems can be mounted on clients using protocols such as NFS and Server Message Block (SMB). Applications access data using the relevant filename and directory path.

Google Cloud provides a range of fully managed and third-party solutions for file storage.

Solution Overview
Filestore

File-based storage using NFS file servers for Compute Engine VMs and Google Kubernetes Engine clusters.

You can choose a service tier (Basic, Zonal, or Regional) that suits your use case.

Parallelstore

Low-latency parallel file system for AI, high performance computing (HPC), and data-intensive applications.

NetApp Volumes

File-based storage using NFS or SMB.

You can choose a service level (Flex, Standard, Premium, or Extreme) that suits your use case.

More options See Summary of file server options.

Data is stored as objects in a flat hierarchy of buckets. Each object is assigned a globally unique ID. Objects can have system-assigned and user-defined metadata, to help you organize and manage the data. Applications access data by referencing the object IDs, using REST APIs or client libraries.

Cloud Storage provides low-cost, highly durable, no-limit object storage for diverse data types. The data you store in Cloud Storage can be accessed from anywhere, within and outside Google Cloud. Optional redundancy across regions provides maximum reliability. You can select a storage class that suits your data-retention and access-frequency requirements.

Comparative analysis

The following table lists the key capabilities of the storage services in Google Cloud.

Persistent Disk Hyperdisk Local SSD Filestore Parallelstore NetApp Volumes Cloud Storage
Capacity

10 GiB to 64 TiB per disk

257 TiB per VM

4 GiB to 64 TiB per disk

512 TiB per VM

10 TiB to 1 PiB per storage pool

375 GiB per disk

12 TiB per VM

1-100 TiB per instance 12-100 TiB

1 TiB to 10 PiB per storage pool

1 GiB to 100 TiB per volume

No lower or upper limit
Scaling
Scale up Not scalable
  • Basic: scale up
  • Zonal and Regional: scale up and down
Not scalable Scale up and down Scales automatically based on usage
Sharing
Supported Supported Not shareable Mountable on multiple Compute Engine VMs, remote clients, and GKE clusters Mountable on multiple Compute Engine VMs and GKE clusters. Mountable on multiple Compute Engine VMs and GKE clusters
  • Read/write from anywhere
  • Integrates with Cloud CDN and third-party CDNs
Encryption key options
  • Google-managed encryption keys
  • Customer-managed
  • Customer-supplied
  • Google-managed encryption keys
  • Customer-managed
  • Customer-supplied
Google-managed encryption keys
  • Google-managed encryption keys
  • Customer-managed (Zonal and Regional tiers)
Google-managed encryption keys
  • Google-managed encryption keys
  • Customer-managed
  • Google-managed encryption keys
  • Customer-managed
  • Customer-supplied
Persistence
Lifetime of the disk Lifetime of the disk Ephemeral (data is lost when the VM is stopped or deleted) Lifetime of the Filestore instance Ephemeral (data is lost when the instance is deleted) Lifetime of the volume Lifetime of the bucket
Availability
  • Zonal
  • Disk cloning
  • Zonal
    • Regional or zonal based on tier
    • Snapshots for Zonal and Regional tiers
    • Backups
    Zonal
    Performance
    Linear scaling with disk size and CPU count Dynamic scaling persistent storage High-performance scratch storage Linear scaling with provisioned capacity

    Scalable performance

    Expectations depend on the service level

    Autoscaling read-write rates, and dynamic load redistribution
    Management
    Manually format and mount Manually format and mount Manually format, stripe, and mount Fully managed Fully managed Fully managed Fully managed

    The following table lists the workload types that each Google Cloud storage option is appropriate for:

    Storage option Workload types
    Persistent Disk
    • IOPS-intensive or latency-sensitive applications
    • Databases
    • Shared read-only storage
    • Rapid, durable VM backups
    Hyperdisk
    • Performance-intensive workloads
    • Scale-out analytics
    Local SSD
    • Flash-optimized databases
    • Hot-caching for analytics
    • Scratch disk
    Filestore
    • Lift-and-shift on-premises file systems
    • Shared configuration files
    • Common tooling and utilities
    • Centralized logs
    Parallelstore
    • AI and ML workloads
    • HPC
    NetApp Volumes
    • Lift-and-shift on-premises file systems
    • Shared configuration files
    • Common tooling and utilities
    • Centralized logs
    • Windows workloads
    Cloud Storage
    • Streaming videos
    • Media asset libraries
    • High-throughput data lakes
    • Backups and archives
    • Long-tail content

    Choose a storage option

    There are two parts to selecting a storage option:

    • Deciding which storage services you need.
    • Choosing the required features and design options in a given service.
      • Deployment region and zone
      • Regional replication
      • Disk type, size, and IOPS (for Extreme Persistent Disk)
      • Encryption keys: Google-owned and Google-managed, customer-managed, or customer-supplied
      • Snapshot schedule
      • Deployment zone
      • Disk type, size, throughput (for Hyperdisk Throughput) and IOPS (for Hyperdisk Extreme)
      • Encryption keys: Google-owned and Google-managed, customer-managed, or customer-supplied
      • Snapshot schedule
      • Deployment region and zone
      • Instance tier
      • Capacity
      • IP range: auto-allocated or custom
      • Access control
      • Deployment region
      • Service level for the storage pool
      • Pool and volume capacity
      • Volume protocol
      • Volume export rules
      • Location: multi-region, dual-region, single region
      • Storage class: Standard, Nearline, Coldline, Archive
      • Access control: uniform or fine-grained
      • Encryption keys: Google-owned and Google-managed, customer-managed, or customer-supplied
      • Retention policy

    Storage recommendations

    Use the following recommendations as a starting point to choose the storage services and features that meet your requirements. For guidance that's specific to AI and ML workloads, see Design storage for AI and ML workloads in Google Cloud.

    General storage recommendations are also presented as a decision tree later in this document.

    • For applications that need a parallel file system, use Parallelstore.

    • For applications that need file-based access, choose a suitable file storage service based on your requirements for access protocol, availability, and performance.

      Access protocol Recommendation
      NFS
      • If you need regional availability and high performance that scales with capacity, use Filestore Regional.
      • If zonal availability is sufficient, but you need high performance that scales with capacity, use Filestore Zonal or NetApp Volumes Premium or Extreme.
      • Otherwise, use either Filestore Basic or NetApp Volumes.

      For information about the differences between the Filestore service tiers, see Service tiers.

      SMB Use NetApp Volumes.

    • For workloads that need primary storage with high performance, use local SSD, Persistent Disk, or Hyperdisk depending on your requirements.

      Requirement Recommendation
      Fast scratch disk or cache

      Use local SSD disks (ephemeral).

      Sequential IOPS Use Persistent Disks with the pd-standard disk type.
      IOPS-intensive workload Use Persistent Disks with the pd-extreme or pd-ssd disk type.
      Balance between performance and cost Use Persistent Disks with the pd-balanced disk type.
      Scale performance and capacity dynamically

      Use Hyperdisk.

      Choose a suitable Hyperdisk type:

      • Hyperdisk Balanced is recommended for general-purpose workloads and highly available applications that need shared write access.
      • Hyperdisk Extreme is recommended for workloads that need high I/O, such as high-performance databases.
      • Hyperdisk Throughput is recommended for scale-out analytics, data drives for cost-sensitive apps, and for cold storage.
      • Hyperdisk ML is recommended for ML workloads that need high throughput to multiple VMs in read-only mode.

      For more information, see About Google Cloud Hyperdisk.

      • Depending on your redundancy requirements, choose between zonal and regional disks.
        Requirement Recommendation
        Redundancy within a single zone in a region Use zonal Persistent Disks or Hyperdisks.
        Redundancy across multiple zones within a region Use regional Persistent Disks.
        For a detailed comparative analysis, see Persistent Disk options.
    • For unlimited-scale and globally available storage, use Cloud Storage.

      Depending on the data-access frequency and the storage duration, choose a suitable Cloud Storage class.

      Requirement Recommendation>
      Access frequency varies, or the data-retention period is unknown or not predictable. Use the Autoclass feature to automatically transition objects in a bucket to appropriate storage classes based on each object's access pattern.
      Storage for data that's accessed frequently, including for high-throughput analytics, data lakes, websites, streaming videos, and mobile apps.

      Use the Standard storage class.

      To cache frequently accessed data and serve it from locations that are close to the clients, use Cloud CDN.

      Low-cost storage for infrequently accessed data that can be stored for at least 30 days (for example, backups and long-tail multimedia content). Use the Nearline storage class.
      Low-cost storage for infrequently accessed data that can be stored for at least 90 days (for example, disaster recovery). Use the Coldline storage class.
      Lowest-cost storage for infrequently accessed data that can be stored for at least 365 days, including regulatory archives. Use the Archive storage class.

      For a detailed comparative analysis, see Cloud Storage classes.

    Data transfer options

    After you choose appropriate Google Cloud storage services, to deploy and run workloads, you need to transfer your data to Google Cloud. The data that you need to transfer might exist on-premises or on other cloud platforms.

    You can use the following methods to transfer data to Google Cloud:

    • Transfer data online by using Storage Transfer Service: Automate the transfer of large amounts of data between object and file storage systems, including Cloud Storage, Amazon S3, Azure storage services, and on-premises data sources.
    • Transfer data offline by using Transfer Appliance: Transfer and load large amounts of data offline to Google Cloud in situations where network connectivity and bandwidth are unavailable, limited, or expensive.
    • Upload data to Cloud Storage: Upload data online to Cloud Storage buckets by using the Google Cloud console, gcloud CLI, Cloud Storage APIs, or client libraries.

    When you choose a data transfer method, consider factors like the data size, time constraints, bandwidth availability, cost goals, and security and compliance requirements. For information about planning and implementing data transfers to Google Cloud, see Migrate to Google Cloud: Transfer your large datasets.

    Storage options decision tree

    The following decision tree diagram guides you through the Google Cloud storage recommendations discussed earlier. For guidance that's specific to AI and ML workloads, see Design storage for AI and ML workloads in Google Cloud.

    View a larger image

    Decision tree to select a storage strategy.

    What's next

    Contributors

    Author: Kumar Dhanagopal | Cross-Product Solution Developer

    Other contributors: