Move your workload to a new compute instance


In certain situations, you might want to move your workload from an existing virtual machine instance (VM) to a newer VM. Reasons to move to a new VM include the following:

  • Take advantage of the new machine types for faster storage or networking speeds. For example, upgrading from C2 to H3 for improved networking bandwidth.
  • Benefit from greater price performance relative to the source VM instance. For example, upgrading from N1 to N4 for greater value on the 5th Generation Intel Xeon processor.
  • Use features only available on the new VM instance. For example, upgrading from N4 to C4 to take advantage of additional performance and maintenance options.
  • Change a virtual machine (VM) instance to a bare metal instance.
  • Add Local SSD disks to your C3 or C3D VM instance.

When upgrading to the newest generation machine series, you might be able to use the simpler procedure described in Edit the machine type of a compute instance if the following conditions are met by the current (source) VM:

  • The operation system (OS) version is supported by the new machine series.
  • The disk type of the boot disk attached to the source VM is supported by the new machine series.
  • The VM doesn't use Local SSD storage.
  • Your VM with attached GPUs uses a G2 machine type. See Add or remove GPUs for details .
  • The VM is using only features that are supported by the new machine series.
  • The VM isn't part of a managed instance group (MIG).
  • The VM uses the gVNIC network interface.

Before you begin

  • If you haven't already, set up authentication. Authentication is the process by which your identity is verified for access to Google Cloud services and APIs. To run code or samples from a local development environment, you can authenticate to Compute Engine as follows.

    Select the tab for how you plan to use the samples on this page:

    Console

    When you use the Google Cloud console to access Google Cloud services and APIs, you don't need to set up authentication.

    gcloud

    1. Install the Google Cloud CLI, then initialize it by running the following command:

      gcloud init
    2. Set a default region and zone.

    REST

    To use the REST API samples on this page in a local development environment, you use the credentials you provide to the gcloud CLI.

      Install the Google Cloud CLI, then initialize it by running the following command:

      gcloud init

    For more information, see Authenticate for using REST in the Google Cloud authentication documentation.

Required roles

To get the permissions that you need to edit or change a VM, ask your administrator to grant you the following IAM roles on the project:

For more information about granting roles, see Manage access to projects, folders, and organizations.

These predefined roles contain the permissions required to edit or change a VM. To see the exact permissions that are required, expand the Required permissions section:

Required permissions

The following permissions are required to edit or change a VM:

  • To change the machine type:
    • compute.instances.stop on the project
    • compute.instances.create on the project
    • compute.instances.start on the project
    • compute.instances.setMachineType on the instance
  • To create a snapshot of the disk:
    • compute.snapshots.create on the project
    • compute.disks.createSnapshot on the disk
  • To create a new disk:
    • compute.disks.list on the project
    • compute.disks.create on the project
    • compute.disks.update on the project
  • To attach a disk to a VM:
    • compute.instances.attachDisk on the instance
    • compute.disks.use on the disk
  • To delete a disk: compute.disks.delete on the project
  • To make changes to the network type:
    • compute.networks.list on the project
    • compute.networks.update on the project

You might also be able to get these permissions with custom roles or other predefined roles.

Evaluate VM migration options

Migrating from one machine type to another is dependent upon several factors, including: regional availability of the new machine type, and the compatibility of the storage options and network interfaces with respect to the guest OS of the source and the new machine series.

Compute requirements

Review the following requirements for your current instance and the new machine type:

  • Explore the machine family resource documentation to identify what machine types are suitable for your workload. Consider whether your application requires specific hardware (GPUs), high performance, or lower costs.
  • Review the features of the disk types supported by the new machine type. Most of the features of Persistent Disk, but not all, are supported by Hyperdisk. However, Hyperdisk provides additional features that aren't available with Persistent Disk.
  • Review the features for the prospective machine series. The new machine series might not support the same features that you use with your current machine series, such as custom machine types, Local SSD, or Shielded VM.
  • Review the regions and zones to ensure the new machine series is available in all the regions as your current VM. You might need to adjust your deployment, high availability, and disaster recovery plans.
  • Review your OS migration plan:
    • If the new VM requires a newer version of the OS, verify that your applications are compatible with the newer OS version.
    • If you're moving to Arm and an Arm image is not available for your current OS version, choose a new OS or an OS version to run your applications on and verify that your applications are compatible with the new OS.
  • It's possible to migrate from a C3 VM instance to a C3 bare metal instance, as long as the source C3 VM instance uses a supported operating system and network driver.
  • If you're moving from a machine series other than C3 to a bare metal instance, you must create a new instance. You might have to run your own hypervisor; however, you can also run any C3 metal supported operating system as long as the IDPF driver is enabled. Bare metal instances use the IDPF network interface presented as only a physical function, not a virtual function.

Storage requirements

Review the following storage requirements for your current instance and the new instance type:

  • Review the supported storage types and the supported storage interfaces for the new machine series.
    • By default, first and second generation machine series use only the Persistent Disk storage type and the VirtIO-SCSI interfaces.
    • Third generation and newer machine series (like M3, C3, and N4) support only the NVMe interface, and some support only Hyperdisk and Local SSD storage types.
    • Bare metal instances (like C3 and X4) support only Hyperdisk.
  • Disk compatibility:
    • If the boot disk uses a disk type that isn't supported by the new machine series, for example pd-standard, then you must create a new boot disk for the new VM.
    • If you are upgrading the OS to a new version, and the operating system doesn't support in-place upgrades, then you must create a new boot disk. All data on the source boot disk is lost unless you copy it to a temporary non-boot disk. Next, you create a new boot disk and copy the data stored on the temporary non-boot disk to the new boot disk.
    • If you aren't upgrading the OS version, then you can take a snapshot of your current boot disk and restore it to the new, supported disk type. When you create a VM, you can then use this new disk as the boot disk.
    • If a non-boot disk uses a disk type that isn't supported by the new machine series, you can use a snapshot to change the source disk to a new disk type, as described in Change the disk type.
  • Local SSD disks can't be moved to a new VM. You can attach a disk large enough to store all the Local SSD data to your current VM, and then use a snapshot to change the source disk to a new disk type, as described in Change the disk type. After you create a VM with attached Local SSD disks, then you can copy the data back to the Local SSD disks.
  • If your current VM instance uses disks in a storage pool, but you are moving your workload to a VM in a different region, then you must recreate the disks and storage pool in the new region.
  • If the new machine series uses a different disk interface (for example, NVMe instead of SCSI), then the disk device names in the guest OS are different. Make sure your applications and scripts use either persistent device names or symlinks when referencing the attached disks.

Networking requirements

Review the following networking requirements for your current instance and the new instance type:

  • Review the supported networking interfaces for the new VM.

    • By default, first and second generation machine series use only the VirtIO network interface.
    • Third generation and newer machine series (like M3, C3, and N4) support only the gVNIC network interface.
    • Bare metal instances support only the IDPF network interface.
  • Make sure your application and operating system support the interfaces available for the machine series.

  • Review your network configuration for your VM to determine if you need to keep the assigned IP addresses. If so, you must promote the IP addresses to static IP addresses.

  • If you use per VM Tier_1 networking performance with your current VM, make sure it is available or needed with the new machine series. For example, you can use Tier_1 networking with a C2 machine type, but it is not needed with an H3 VM.

To determine the network interface type of your current VM, use the gcloud compute instances describe command to view the VM's nic-type:

  gcloud compute instances describe VM_NAME --zone=ZONE

If your VM has a nic-type set to VIRTIO, you can't change the network interface type. You must create a new VM and set the network interface type to gVNIC.

Prepare to move your existing VMs

After you've completed the evaluation section, the next step is to prepare to move your VM instances by requesting resources for the new VM instance and preparing backups of the source VM instance.

Prepare compute resources

Complete the following steps to prepare for moving your current instance to a new instance:

  1. Request quota in the region and zones where you plan to move your resources. If you have existing quota for a machine type, you can request to move that quota. The process takes a few days to complete.
  2. Create a reservation for the new VM instances to ensure the machine resources are available in the new region and zones. Make sure you understand how reserved resources are consumed and test that you can consume reserved resources.
  3. Extend your high availability and disaster recovery plans to include the new region.
  4. If needed, upgrade the OS on the current VM.
    1. If supported by the operating system vendor, perform an in-place upgrade of your OS to a version that is supported by the new machine series and verify that your workload is performing as expected on the new OS version.
    2. If an in-place upgrade of the OS isn't supported, then, when you create a new VM, you must create a new boot disk. Determine what information you need to copy from the current boot disk, and copy it to a temporary location on a non-boot disk so it can be transferred to the new VM. If you don't have any non-boot disks attached to your current VM:

Prepare storage resources

Complete the following steps to prepare for moving the data in the disks attached to your current instance to a new instance:

  1. On Linux systems, test your updated applications and scripts to make sure they work with persistent device names or symlinks instead of the disk device names.
  2. If you're migrating from a VM that runs Microsoft Windows:
  3. If your new VM doesn't support the same disk types as your current VM, you might need to update your deployment scripts or instance templates to support the new machine series.
  4. If your current VM uses a disk type for the boot disk that isn't supported by the new machine series, and you are migrating multiple VMs with the same configuration, create a custom image to use when creating the new VMs:
    1. Create a snapshot of the pd-standard boot disk of your current VM.
    2. Create a custom image using the disk snapshot as the source.
  5. If you need to move Local SSD information, create a blank disk large enough to backup your Local SSD disks.
    1. If possible, use a disk type that is supported by the new VM.
    2. If there are no disk types that are supported by both the current VM and the new VM, then create a temporary disk using a disk type supported by the current VM.
    3. Attach the new disk to the current VM, then format and mount the disk.
    4. Copy the data from the Local SSD disks attached to the current VM to this temporary disk.
  6. Change the disk type of any disks attached to the current VM that use a disk type that isn't supported by the new VM. To move the disk data to new disks, create snapshots of the disks. You can alternatively transfer files from one VM to the other.

    1. You can take the snapshots while the VM is running, but any data written to the disks after you take the snapshot is not captured. Because snapshots are incremental, you can take a second snapshot after you stop the VM to capture all the most recent changes. This approach should minimize the length of time the VM is unavailable while you switch to a new VM.
    2. Alternatively, you can take all the disk snapshots after you stop the VM. We recommend that you create a snapshot of all the disks attached to your VM, even if the disk type is supported by the new machine series. Include any temporary disks that contain the copied Local SSD data.
    3. The amount of time it takes to snapshot a disk is dependent upon multiple factors, such as the disk size and amount of data contained on the disk. For example, if you take a snapshot of a 1 TiB disk that is 85% full, it might take 5 minutes for the snapshot to complete. But, if you take a snapshot of a 100 TiB disk that is 85% full, it might take 11 minutes to complete. We recommend you perform test snapshots of your disks before you start the migration process to understand how long snapshotting takes.
  7. If you have a disk that can be taken offline, you can use the following approach to move the data to a new disk while the source VM is still available:

    1. Detach the disk from your VM.
    2. Take a snapshot of the disk.
    3. Use the snapshot to create a new disk using a disk type that is supported by the new machine series. The new disk must be either the same size or larger than the source disk.

Prepare network resources

Complete the following steps to update the network configuration used by your current instance to support the new instance:

  1. If your current VM doesn't use gVNIC, you must create a new instance with a network interface that uses gVNIC. Review Overview of using gVNIC with Compute Engine VMs to understand the steps you need to take when creating a new instance.
  2. If you're creating a VM in a new region, create a VPC network and subnets in the new region.
  3. If you configured custom NIC queue counts, see Queue allocations and changing the machine type.
  4. If you want to keep the IP addresses used by the source VM, promote the IP addresses to static IP addresses.
  5. Unassign static IP address before you stop your source VM.

Move your workload to the new VM

After preparing your VMs for migration, the next step is to move your workload to the new VM.

If you are moving your VMs from the first generation to the second generation machine series, read the instructions on the Edit the machine type of a VM page. If you want to change the name of your existing VM, review the information at Rename a VM.

Permissions required for this task

To perform this task, you must have the following permissions:

  • compute.instances.setMachineType on the VM

This section describes how to move your workload from a first or second generation VM to a third (or newer) generation VM. During this procedure, you create a new VM instance, then move your workloads to the new VM.

  1. When creating the new VM, choose one of the supported disk types for the boot disk, for example, Hyperdisk Balanced.

Create the new VM

When moving your workloads from first or second generation VMs (N1 or N2 for example) to third generation or later, you must first create a new VM and then move your workloads.

  1. If the source VM uses non-boot disks with a disk type that is supported by the new machine series, detach the disks from the VM.
  2. Stop your source VM.
  3. Create snapshots of all disks that are still attached to the source VM.
  4. Create a new compute VM instance using either a public image or a custom image that is configured to use gVNIC. When creating the new VM, choose the following options:
    • Select the machine type from the machine series that you have chosen.
    • Select a supported OS image, or use a custom image that you created previously.
    • Select a supported disk type for the boot disk.
    • If you created new disks from snapshots of the original disks, include those new disks.
    • Specify the new VPC network, if you're creating the instance in a different region.
    • If both VirtIO and gVNIC are supported for the new instance, select gVNIC.
    • Specify the static IP addresses, if you promoted the ephemeral IP addresses of the source VM.
  5. Start the new VM.

After the instance starts

Now that the new instance has been created and started, complete the following steps to finish the configuration of the new instance and copy over all the data from the source instance.

  1. Attach the disks you detached from the source VM to the new VM.
  2. For any disks attached to the source VM that use a disk type not supported by the new VM, Create a disk from a snapshot and attach it to the new instance. When creating the new disk, select a disk type that is supported by the new VM and specify a size that is at least as large as the original disk.
  3. If you created the new VM using a public OS image, and not a custom image, then do the following:
    1. Configure the necessary users, drivers, packages, and file directories on the new instance to support your workload.
    2. Install your modified applications and programs on the new VM. Recompile the programs on the new OS or architecture, if required.
  4. Optional: If you moved the contents of Local SSD disks to a temporary disk, and the new VM has attached Local SSD storage, after you format and mount the disks, you can move the data from the temporary disk to the Local SSD disks.
  5. Reassign any static IP addresses associated with the source VM to the new VM.
  6. Complete any additional tasks required to make your new VM highly available, such as configuring load balancers and updating the forwarding rules.
  7. Optional: Update the DNS entries, if needed, for the new VM.
  8. Recommended: Schedule disk backups for the new disks.

If you encounter issues when moving your workload, contact your Technical Account Manager (TAM) or the Google Professional Services Organization (PSO) for assistance.

Migration example of n1-standard-8 to n4-standard-8

The following example is a migration of an n1-standard-8 VM to an n4-standard-8 VM. The n1-standard-8 VM has a PD-SSD boot disk running an Ubuntu1804 image and a PD-SSD data disk. You must use the CLI or REST API for this procedure.

There are two options available to upgrade your N1 VM to an N4 VM:

Option 1: If your N1 VM uses the VirtIO network interface, then you must create a new N4 VM. N4 supports only the gvnic network interface, and Hyperdisk Balanced disks. You create a snapshot of your Persistent Disk boot and data disks, create Hyperdisk Balanced disks from those snapshots, attach the Hyperdisk Balanced disks, and create the new N4 VM with the Hyperdisk Balanced disks.

You can also choose to create a new Hyperdisk Balanced boot disk using a more recent version of the Ubuntu OS. In this scenario, you can create a new Hyperdisk Balanced disk from the boot disk snapshot, but you attach that disk as a non-boot disk to the N4 VM. Then you can copy non-system data from the restored snapshot to the new boot disk.

Option 2: If your N1 VM uses the gvnic network interface, the operating system has an NVMe storage device driver, doesn't have any attached Local SSD disks or GPUs, and isn't part of a managed instance group (MIG), then you can change the machine type from N1 to N4, but you still must change your Persistent Disk disk types to Hyperdisk Balanced disks. You must first detach your Persistent Disk boot and data disks, create snapshots of the disks, create Hyperdisk Balanced disks using the snapshots as the source, then attach the new Hyperdisk Balanced disks to your N4 VM after you change the machine type. If your VM has attached GPUs, you must detach them first.

The time to snapshot a disk is dependent upon multiple factors such as the total number of TBs on a disk. For example, if you take a snapshot of a 1 TB disk that is 85% full, it might take 5 minutes for the snapshot to complete. But, if you take a snapshot of a 100 TB disk that is 85% full, it might take 11 minutes to complete. Google recommends you perform test snapshots of your disks before you start the migration process to understand how long snapshotting takes.

gcloud

Option 1: Create a new N4 VM with snapshotted disks:

  1. Stop the VM by using gcloud compute instances stop:

    gcloud compute instances stop VM_NAME \
      --zone=ZONE \
    

    Replace the following:

    • VM_NAME The name of your current n1-standard-8 VM.
    • ZONE: The zone where the VM is located.
  2. Snapshot your disks. Use the gcloud compute snapshots create to create a snapshot of both the Persistent Disk boot disk and data disk attached to the VM.

    gcloud compute snapshots create SNAPSHOT_NAME \
        --source-disk=SOURCE_DISK_NAME \
        --source-disk-zone=SOURCE_DISK_ZONE
    

    Replace the following:

    • SNAPSHOT_NAME: The name of the snapshot you want to create.
    • SOURCE_DISK_NAME: The name of your source disk.
    • SOURCE_DISK_ZONE: The zone of your source disk.
  3. Create a new Hyperdisk Balanced disk for the data disk by repeating the previous step and specifying the data disk information instead of the boot disk. gcloud compute disks create:

    gcloud compute disks create DISK_NAME \
        --project=PROJECT_NAME \
        --type=DISK_TYPE \
        --size=DISK_SIZE \
        --zone=ZONE \
        --source-snapshot=SNAPSHOT_NAME \
        --provisioned-iops=PROVISIONED_IOPS \
        --provisioned-throughput=PROVISIONED_THROUGHPUT
    
    

    Replace the following:

    • DISK_NAME: The name of the new disk you are creating from the snapshotted disk.
    • PROJECT_NAME: The name of your project.
    • DISK_TYPE: The new disk type—in this example, it's a Hyperdisk Balanced disk.
    • DISK_SIZE: The size of the disk (example: 100GB).
    • ZONE: The zone where the new disk is located.
    • SNAPSHOT_NAME: The name of the snapshot source disk.
    • Optional: PROVISIONED_IOPS: The IOPS performance for the disk (example: 3600).
    • Optional: PROVISIONED_THROUGHPUT: The throughput performance to provision the disk (example: 290).
  4. Repeat the previous step for each snapshotted disk.

  5. Create the n4-standard-8 VM and attach the Hyperdisk Balanced disks using the gcloud compute instances create:

    gcloud compute instances create VM_NAME \
        --project=PROJECT_NAME \
        --zone=ZONE \
        --machine-type=NEW_MACHINE_TYPE \
        --boot-disk-device-name=BOOT_DISK_NAME \
        --disk=name=NON_BOOT_DISK_NAME, boot=no \
        --network-interface=nic-type=GVNIC \
    

    Replace the following:

    • VM_NAME: The name of the new VM instance.
    • PROJECT_NAME: The name of your project.
    • ZONE: The zone where the new VM is located.
    • NEW_MACHINE_TYPE: The machine type, in this example it's n4-standard-8.
    • BOOT_DISK_NAME The name of the Hyperdisk Balanced boot disk you created from the source disk snapshot attached to the n1-standard-8 VM.
    • NON_BOOT_DISK_NAME The name of the Hyperdisk Balanced data disk you created from the source snapshot disk attached to the n1-standard-8 VM.
  6. Start the n4-standard-8 VM using the gcloud compute instances start:

    gcloud compute instances start VM_NAME
    

    Replace VM_NAME with name of the new VM.

Option 2: Perform an in-place machine upgrade:

This option is only available if your N1 VM uses the gvnic network interface, the operating system has an NVMe storage device driver, doesn't have any attached Local SSD disks or GPUs, and isn't a part of a managed instance group (MIG). Performing this procedure with an N1 VM with a VirtIO network interface generates a VM incompatibility error.

  1. Stop the VM.
  2. Detach the disks from the VM.
  3. Create a snapshot of the boot and data disks.
  4. Create Hyperdisk Balanced boot and data disks using a disk snapshot as the source for each disk.
  5. Set the machine type to an N4 VM.
  6. Attach the Hyperdisk Balanced boot disk and the Hyperdisk Balanced data disk.
  7. Start the N4 VM.

REST

Option 1: Create a new N4 VM with snapshotted disks:

  1. Stop the VM by using the instances.stop method:

     POST https://compute.googleapis.com/compute/v1/projects/PROJECT_NAME/zones/ZONE/instances/VM_NAME/stop
    

    Replace the following:

    • PROJECT_NAME: The project ID.
    • ZONE: The zone containing the VM.
    • VM_NAME: The name of your current n1-standard-8 VM.
  2. Snapshot your disks using the disks.createSnapshot method to create a snapshot of both the Persistent Disk boot disk and data disk attached to the instance.

    POST https://compute.googleapis.com/compute/v1/projects/PROJECT_NAME/zones/ZONE/disks/DISK_NAME/createSnapshot
    

    In the body of the request, include the name for your new snapshotted Persistent Disk disk.

    For example:

    {
        "name": "SNAPSHOT_NAME"
    }
    

    Replace the following:

    • PROJECT_NAME: The name of your project.
    • ZONE: The zone where your disk is located.
    • DISK_NAME: The disk you plan to snapshot.
    • SNAPSHOT_NAME: A name for the snapshot, such as hdb-boot-disk or hdb-data-disk.
  3. Create a Hyperdisk Balanced disk using the disks.insert method. You perform this step two times: once to include the name of your Hyperdisk Balanced boot disk; and a second time to include the name of your data disks. Use the sourceSnapshot for the new Hyperdisk Balanced boot and data disks, the type of disk, Hyperdisk Balanced, and sizeGB of the disk in the request body.

    POST https://compute.googleapis.com/compute/v1/projects/PROJECT_NAME/zones/ZONEdisks
    

    Replace the following:

    • PROJECT_NAME: The name of your project.
    • ZONE: The zone where your disk is located.

    In the body of the request, include the following:

    For example:

    {
        "name": "my-hdb-boot-disk" or "my-hdb-data-disk",
        "sourceSnapshot": "projects/your-project/global/snapshots/SNAPSHOT_NAME",
        "type": "projects/your-project/zones/us-central1-a/diskTypes/hyperdisk-balanced",
        "sizeGb": "100"
    }'
    
  4. Use the instances.insert method to create the new N4 VM.

    
    POST https://compute.googleapis.com/compute/v1/projects/PROJECT_NAME/zones/ZONE/instances
    
    

    Replace the following:

    • PROJECT_NAME: The name of your project.
    • ZONE: The zone where your disk is located.

    In the body of the request, include the following:

    
      {
        "machineType":"projects/your-project/zones/us-central1-a/machineTypes/n4-standard-8" "name":"VM_NAME", 
        "disks": [
          { 
            "boot": true,
            "deviceName": "my-hdb-boot-disk",
            "source": "projects/your-project/zones/us-central1-a/disks/my-hdb-boot-disk",
            "type": "PERSISTENT"
          },
    
          {
            "boot": false,
            "deviceName": "my-hdb-data-disk",
            "source": "projects/your-project/zones/us-central1-a/disks/my-hdb-data-disk",
            "type": "PERSISTENT"
          }
          ],
            "networkInterfaces":[
              {
                "network":"global/networks/NETWORK_NAME", 
                "subnetwork":"regions/REGION/subnetworks/SUBNET_NAME",
                "nicType": "GVNIC"
              }
           ]
         }
    
    

    Replace the following:

    • VM_NAME: The name of the VM.
    • NETWORK_NAME: The name of the network.
    • REGION: The name of the region.
    • SUBNET_NAME: The name of the subnet.
  5. Start the VM by using the instances.start method:

    POST https://compute.googleapis.com/compute/v1/projects/PROJECT_NAME/zones/ZONE/instances/VM_NAME/start
    

    Replace the following:

    • PROJECT_NAME: The name of your project.
    • ZONE: The zone where your VM is located.
    • VM_NAME: The name of the VM.

Option 2: Perform an in-place machine upgrade:

This option is only available if your N1 VM uses the gvnic network interface, doesn't have any attached Local SSD disks or GPUs, and isn't a part of a managed instance group (MIG). Performing this procedure with an N1 VM with a VirtIO network interface generates a VM incompatibility error.

  1. Stop the VM by using the instances.stop method.

  2. Detach the disks by using the instances.detachDisk method method to detach the original Persistent Disk boot disk from the N1 VM. You also must detach any data disks from the VM.

    https://compute.googleapis.com/compute/v1/projects/PROJECT_NAME/zones/ZONE/instances/VM_NAME/detachDisk?deviceName=DISK_NAME
    

    Replace the following:

    • PROJECT_NAME: The name of your project.
    • ZONE: The zone where your disk is located.
    • VM_NAME: The name of the source VM with the pd-ssd disk attached to it.
    • DISK_NAME: The disk you want to detach.
  3. Snapshot the disks. Use the disks.createSnapshot method to create a snapshot of both the Persistent Disk boot disk and data disks attached to the instance.

  4. Create a Hyperdisk Balanced boot and data disks using the disks.insert method include the name of your Hyperdisk Balanced disk, sourceSnapshot for the new Hyperdisk Balanced disk, the type of disk, Hyperdisk Balanced, and sizeGB of the disk in the request body.

  5. Perform an in-place machine type upgrade using the instances.setMachineType method include the machineType in the request body:

    POST  https://compute.googleapis.com/compute/v1/projects/PROJECT_NAME/zones/ZONEinstances/VM_NAME/setMachineTypeMACHINE_TYPE
    

    Replace the following:

    • PROJECT_NAME: The name of your project.
    • ZONE: The zone where your disk is located.
    • VM_NAME: The name of the VM to upgrade.
    • MACHINE_TYPE: The new machine type.

    In the request body, include the following:

    
    {
     "machineType": "projects/PROJECT_NAME/zones/ZONE/machineTypes/MACHINE_TYPE",
    }
    
    
  6. Use the instances.attachDisk method to attach the new Hyperdisk Balanced boot disk and the Hyperdisk Balanced data disks to the N4 VM.

    POST https://compute.googleapis.com/compute/v1/projects/PROJECT_NAME/zones/ZONE/instancesVM_NAMEattachDiskDISK_NAME
    

    Replace the following:

    • PROJECT_NAME: The name of your project.
    • ZONE: The zone where your disk is located.
    • VM_NAME: The name of the source VM instance with the pd-ssd disk attached to it.
    • DISK_NAME The disk you want to attach.

    In the request body, include the following:

    {
    "source": "projects/your-project/zones/us-central1-a/disks/my-hdb-boot-disk",
    "deviceName":"my-hdb-boot-disk","boot":true
    }
    
    {
    "source": "projects/your-project/zones/us-central1-a/disks/my-hdb-data-disk",
    "deviceName":"my-hdb-data-disk","boot":false
    }
    
  7. Start the N4 VM by using the instances.start method.

    POST https://compute.googleapis.com/compute/v1/projects/PROJECT_NAME/zones/ZONEinstances/VM_NAME/start
    

    Replace the following:

    • PROJECT_NAME: The name of your project.
    • ZONE: The zone where your disk is located.
    • VM_NAME: The name of the VM.

Clean up

After you verify that you can connect to the new VM, and that your workload is running as expected on the new VM, you can remove the resources that are no longer needed:

  1. The snapshots you created for the disks attached to the source VM.
  2. Any snapshot schedules for the disks that were attached to the source VM.
  3. The temporary disk created to copy the Local SSD data to the new VM.
  4. The source VM and all attached disks.

What's next