Overview
This page shows you how to migrate a version 1.30 or higher admin cluster to these recommended features:
The load balancer configuration:
The integrated F5 BIG-IP load balancer configuration to
ManualLB
.or
The bundled Seesaw load balancer to MetalLB.
Migrate to a high availability (HA) admin cluster from a non-HA admin cluster. Availability is significantly improved with an HA admin cluster while using the same number of VMs. A non-HA admin cluster has one control-plane node and two add-on nodes. An HA admin cluster's three nodes are all control-plane nodes with no add-on nodes.
This page is for IT administrators and Operators who manage the lifecycle of the underlying tech infrastructure. To learn more about common roles and example tasks that we reference in Google Cloud content, see Common GKE Enterprise user roles and tasks.
For more information about migration planning, see Plan cluster migration to recommended features.
Best practices
If you have multiple environments such as test, development, and production, we recommend that you first migrate the least critical environment, such as test. After you verify that the migration was successful, repeat this process for each environment, migrating your production environment last. This lets you validate each migration's success and ensure that your workloads are running properly, before moving to the next more critical environment.
Requirements
- The admin cluster must be at version 1.30 or higher.
- All the user clusters managed by the admin cluster must already be migrated to recommended features, as described in the Migrate a user cluster to recommended features.
Plan for downtime during migration
For the migration, plan for some limited control-plane downtime. The Kubernetes API access is unavailable for non-HA admin clusters for about 20 minutes, but the Kubernetes control plane remains available for HA admin clusters with F5. During migrations, the Kubernetes data plane continues to work in a stable state.
From | To | Kubernetes API Access | User workloads |
---|---|---|---|
HA admin clusters with F5 BIG-IP |
HA admin clusters with |
Not affected |
Not affected |
non-HA admin clusters with |
HA admin clusters with the same kind of load balancer |
Affected |
Not affected |
non-HA admin clusters with F5 BIG-IP |
HA admin clusters with |
Affected |
Not affected |
non-HA admin clusters with Seesaw |
HA admin clusters with MetalLB |
Affected |
Not affected |
- Affected: There is a noticeable service disruption during the migration.
- Not affected: There is either no service disruption or it is almost unnoticeable.
Prepare for the migration
If your admin cluster is non-HA, prepare to migrate to an HA admin cluster by following the steps in this section. If your admin cluster is already HA, skip to the next section, Prepare for the load balancer migration.
Allocate additional IP addresses
When migrating the admin cluster from non-HA to HA, allocate four additional IP addresses. Ensure these IP addresses are in the same VLAN as the existing admin cluster nodes and aren't already used by any existing nodes:
- Allocate one IP address for the new control plane VIP,
for the
loadBalancer.vips.controlPlaneVIP
field in the admin cluster configuration file. - Allocate a new IP addresses for each of the three control-plane nodes,
for the
network.controlPlaneIPBlock
section in the admin cluster configuration file.
Update firewall rules
When migrating the admin cluster from non-HA to HA, update the firewall rules on your admin cluster. This ensures that the newly allocated IP addresses for the control-plane nodes can reach all required APIs and other destinations, as described in Firewall rules for admin clusters.
Prepare for the load balancer migration
If your admin cluster is using the integrated F5 BIG-IP configuration or the bundled Seesaw load balancer, follow the steps in this section to make the to make the necessary changes to the admin cluster configuration file. Otherwise, skip to the next section, Prepare to migrate from non-HA to HA.
F5 BIG-IP
If your admin cluster is using the integrated F5 BIG-IP configuration, make the following changes to the admin cluster configuration file:
- Set the
loadBalancer.kind
field to"ManualLB"
. - Set or keep the value of the
loadBalancer.vips.controlPlaneVIP
field. If your admin cluster is already HA, keep the same value. If you are migrating from a non-HA admin cluster to an HA admin cluster, change the value of theloadBalancer.vips.controlPlaneVIP
field to the IP address that you allocated. - Delete the entire
loadBalancer.f5BigIP
section.
The following example admin cluster configuration file shows these changes:
loadBalancer: vips: controlPlaneVIP: 192.0.2.6 kind:"F5BigIP""ManualLB"f5BigIP: address: "203.0.113.20" credentials: fileRef: path: ""my-config-folder/user-creds.yaml" entry: "f5-creds" partition: "my-f5-user-partition"
Seesaw
If your admin cluster uses the Seesaw load balancer, make the following changes to the admin cluster configuration file:
- Set the
loadBalancer.kind
field to "MetalLB". - Keep the
network.hostConfig
section. - Set or keep the value of the
loadBalancer.vips.controlPlaneVIP
]5 field. If your admin cluster is already HA, you can keep the same value. If migrating from a non-HA admin cluster to an HA admin cluster, change the value of theloadBalancer.vips.controlPlaneVIP
field to the IP address that you allocated. - Remove the
loadBalancer.seesaw
section.
The following example admin cluster configuration file shows these changes:
network: hostConfig: dnsServers: - "203.0.113.1" - "203.0.113.2" ntpServers: - "203.0.113.3" loadBalancer: vips: controlPlaneVIP: 192.0.2.6 kind: "MetalLB""Seesaw"seesaw: ipBlockFilePath: "user-cluster-1-ipblock.yaml" vrid: 1 masterIP: "" cpus: 4 memoryMB: 3072
Prepare to migrate from non-HA to HA
If your admin cluster is non-HA, prepare to migrate to HA by following the steps in this section.
If your admin cluster is already HA, skip to the next section, Migrate the admin cluster.
If your admin cluster version is 1.29.0-1.29.600 or 1.30.0-1.30.100, and if always-on secrets encryption was enabled in the admin cluster at version 1.14 or earlier, you must rotate the encryption key before starting the migration. Otherwise, the new HA admin cluster will be unable to decrypt secrets.
To check whether the cluster could be using an old encryption key:
kubectl --kubeconfig ADMIN_CLUSTER_KUBECONFIG get secret -n kube-system admin-master-component-options -o jsonpath='{.data.data}' | base64 -d | grep -oP '"GeneratedKeys":\[.*?\]'
If the output shows an empty key, such as in the following example, you must rotate the encryption key using the steps in this known issue.
"GeneratedKeys":[{"KeyVersion":"1","Key":""}]
Update the admin cluster configuration file
Make these changes to the admin cluster configuration file:
- Fill in the
network.controlPlaneIPBlock
with the three IP addresses that you allocated for the control-plane nodes. - Ensure that you have filled the
network.hostConfig
section. This section holds information about NTP servers, DNS servers, and DNS search domains used by the VMs that are your cluster nodes. - Ensure that you have replaced the value of
loadBalancer.vips.controlPlaneVIP
with the IP address that you allocated. - Set
adminMaster.replicas
to 3. - Remove the
vCenter.dataDisk
field. For an HA admin cluster, the paths for the three data disks used by control-plane nodes are automatically generated under the root directoryanthos
in the datastore. - If
loadBalancer.kind
is set to"ManualLB"
, setloadBalancer.manualLB.controlPlaneNodePort
to 0.
The following example admin cluster configuration file shows these changes:
vCenter: address: "my-vcenter-server.my-domain.example" datacenter: "my-data-center"dataDisk: "xxxx.vmdk"... network: hostConfig: dnsServers: - 203.0.113.1 - 203.0.113.2 ntpServers: - 203.0.113.3 ... controlPlaneIPBlock: netmask: "255.255.255.0" gateway: "198.51.100.1" ips: - ip: "192.0.2.1" hostname: "admin-cp-hostname-1" - ip: "192.0.2.2" hostname: "admin-cp-hostname-2" - ip: "192.0.2.3" hostname: "admin-cp-hostname-3" ... ... loadBalancer: vips: controlPlaneVIP:192.0.2.6192.0.2.50 kind: ManualLB manualLB:controlPlaneNodePort: 300030 ... adminMaster: replicas: 3 cpus: 4 memoryMB: 8192 ...
Adjust mappings in your load balancer if needed
If your admin cluster has been using manual load balancing, complete the step in this section.
If you are migrating from integrated F5 BIG-IP to manual load balancing, or if you are migrating to MetalLB, skip to the next section, Migrate the admin cluster.
For each of the three new control-plane node IP addresses that you specified in
the network.controlPlaneIPBlock
section, configure this mapping in your
external load balancer (such as F5 BIG-IP or Citrix):
(old controlPlaneVIP:443) -> (NEW_NODE_IP_ADDRESS:old controlPlaneNodePort)
This ensures that the old control-plane VIP continues working during the migration.
Migrate the admin cluster
Carefully review all the changes that you made to the admin cluster configuration file. All the settings are immutable except when updating the cluster for the migration.
Update the cluster:
gkectl update admin --kubeconfig ADMIN_CLUSTER_KUBECONFIG \
--config ADMIN_CLUSTER_CONFIG
Replace the following
:
ADMIN_CLUSTER_KUBECONFIG
: the path of the admin cluster kubeconfig file.ADMIN_CLUSTER_CONFIG
: the path of the admin cluster configuration file.
The command displays the progress of the migration.
When prompted, enter Y
to continue.
During the migration from non-HA to HA, the older control-plane VIP continues to function and can be used to access the new HA admin cluster. When the migration completes, the admin cluster kubeconfig file is automatically updated to use the new control-plane VIP.
After the migration
After the update completes, verify that the admin cluster is running:
kubectl get nodes --kubeconfig ADMIN_CLUSTER_KUBECONFIG
The expected output is similar to the following:
Load balancer migration
If you migrated the load balancer, verify that the load balancer components are running successfully.
MetalLB
If you migrated to MetalLB, verify that the MetalLB components are running successfully using the following command:
kubectl --kubeconfig ADMIN_CLUSTER_KUBECONFIG get pods \
--namespace kube-system --selector app=metallb
The output shows Pods for the MetalLB controller and speaker. For example:
metallb-controller-744884bf7b-rznr9 1/1 Running
metallb-speaker-6n8ws 1/1 Running
metallb-speaker-nb52z 1/1 Running
metallb-speaker-rq4pp 1/1 Running
After a successful migration, delete the powered-off Seesaw VMs for the admin
cluster. You can find the Seesaw VM names in the vmnames
section of the
seesaw-for-gke-admin.yaml
file in your configuration directory.
ManualLB
After you update your clusters to use manual load balancing, traffic to your clusters isn't interrupted. This is because the existing F5 resources still exist, as you can see by running the following command:
kubectl --kubeconfig ADMIN_CLUSTER_KUBECONFIG \
The expected output is similar to the following:
Warning: v1 ComponentStatus is deprecated in v1.19+
NAMESPACE NAME TYPE DATA AGE
kube-system secret/bigip-login-xt697x Opaque 4 13h
NAMESPACE NAME SECRETS AGE
kube-system serviceaccount/bigip-ctlr 0 13h
kube-system serviceaccount/load-balancer-f5 0 13h
NAMESPACE NAME READY UP-TO-DATE AVAILABLE AGE
kube-system deployment.apps/k8s-bigip-ctlr-deployment 1/1 1 1 13h
kube-system deployment.apps/load-balancer-f5 1/1 1 1 13h
NAME ROLE AGE
clusterrolebinding.rbac.authorization.k8s.io/bigip-ctlr-clusterrole-binding ClusterRole/bigip-ctlr-clusterrole 13h
clusterrolebinding.rbac.authorization.k8s.io/load-balancer-f5-clusterrole-binding ClusterRole/load-balancer-f5-clusterrole 13h
NAME CREATED AT
clusterrole.rbac.authorization.k8s.io/bigip-ctlr-clusterrole 2024-03-25T04:37:34Z
clusterrole.rbac.authorization.k8s.io/load-balancer-f5-clusterrole 2024-03-25T04:37:34Z