This page describes how to restore Cassandra in a single region.
In a single region deployment, Apigee hybrid is deployed in a single data center or a region. If you
have multiple Apigee organizations in your deployment, the restore process restores data for all the organizations.
In a multi-organization setup, you cannot restore a specific organization.
Restoring a region from a backup
Update the Cassandra restore details in the overrides.yaml file:
Namespace for restore. Use the same namespace as in your original cluster.
cassandra:hostNetwork
hostNetwork is required and should always be set to
false.
restore:enabled
Restore is disabled by default. You must set this
property to true.
restore:serviceAccountPath
SA_JSON_FILE_PATH
The path on your filesystem to the service account you created
for the backup.
restore:dbStorageBucket
CLOUD_STORAGE_BUCKET_NAME
The name of a Google Cloud Storage bucket that stores backup archives to be used
for data restoration.
restore:cloudProvider
GCP
The cloudProvider: "GCP" property is required.
restore:snapshotTimestamp
TIMESTAMP
The timestamp of the backup snapshot to restore. To check what timestamps can be used,
go to the dbStorageBucket and look at the files that are present in the
bucket. Each file name contains a timestamp value. For example,
backup_20210203213003_apigee-cassandra-default-0.tgz
Where 20210203213003 is the snapshotTimestamp value you would
use if you wanted to restore the backups created at that point in time.
backup:enabled
You should set this property to false in case it had been previously set
to true.
In case you do not have a clean cluster to start out with, follow the
Decommission a hybrid region for helm
documentation to bring your existing Hybrid installation into a clean state
(you can leave the Cert Manager installed). This would bring you to an equal state
as if you would have followed Helm runtime setup manual
until the beginning of Step 11.
Verify there are no pods remaining in the Apigee namespace:
kubectl get pods -n APIGEE_NAMESPACE
If you are using CSI backup, make sure that you can see
the volumesnapshots you want to use for the restoration process by running:
kubectl get volumesnapshot -n APIGEE_NAMESPACE
Install all Hybrid components one by one as described in
Step 10: Install Apigee hybrid using Helm . Note that the apigee-cassandra-restore pod will get
created once you run the command to install the datastore, but it will only
go into running state after you install the apigee-org component.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-08-29 UTC."],[],[],null,["# Restoring in a single region\n\nThis page describes how to restore Cassandra in a single region.\n\nIn a single region deployment, Apigee hybrid is deployed in a single data center or a region. If you\nhave multiple Apigee organizations in your deployment, the restore process restores data for all the organizations.\nIn a multi-organization setup, you cannot restore a specific organization.\n| **Note** : Before you begin restoring a single region, consider whether the following prerequisite steps are applicable:\n|\n| - If you want to preserve an existing setup for troubleshooting and root cause analysis (RCA), you should delete all the `org` and `env` components from the Kubernetes cluster *except* the Apigee controller, and then retain the cluster. The cluster will contain the existing Apigee datastore (Cassandra) which you can use for troubleshooting. Create a new Kubernetes cluster and then restore Cassandra in the new cluster.\n| - If your hybrid installation was set up with multiple organizations, get the overrides files for each organization before proceeding with restore in a single region. You can add the restore configuration as described in [Step 3](#step3) to any **one** of the overrides files. Do not add the restore configuration to any other overrides file.\n\nRestoring a region from a backup\n--------------------------------\n\n1. Update the Cassandra restore details in the `overrides.yaml` file:\n\n ```actionscript-3\n namespace: YOUR_RESTORE_NAMESPACE # Use the same namespace as in your original cluster.\n cassandra:\n hostNetwork: false\n ...\n restore:\n enabled: true\n serviceAccountPath: \"\u003cvar translate=\"no\"\u003eSA_JSON_FILE_PATH\u003c/var\u003e\"\n dbStorageBucket: \"\u003cvar translate=\"no\"\u003eCLOUD_STORAGE_BUCKET_NAME\u003c/var\u003e\"\n cloudProvider: \"GCP\" # required verbatim \"GCP\" (all caps)\n snapshotTimestamp: \"\u003cvar translate=\"no\"\u003eTIMESTAMP\u003c/var\u003e\"\n ...\n backup:\n enabled: false\n ...\n ```\n\n\n Where:\n\n | **Note:** In case you are using **CSI backup** , please follow the **example restore config** in the [CSI backup and restore](/apigee/docs/hybrid/v1.15/cassandra-csi-backup-restore#example-restore-config) documentation.\n2. In case you do not have a clean cluster to start out with, follow the\n [Decommission a hybrid region for helm](/apigee/docs/hybrid/v1.15/decommission-region#helm)\n documentation to bring your existing Hybrid installation into a clean state\n (you can leave the **Cert Manager** installed). This would bring you to an equal state\n as if you would have followed [Helm runtime setup manual](/apigee/docs/hybrid/v1.15/install-create-cluster)\n until the beginning of Step 11.\n\n3. Verify there are no pods remaining in the Apigee namespace:\n\n ```\n kubectl get pods -n APIGEE_NAMESPACE\n ```\n4. If you are using CSI backup, make sure that you can see\n the volumesnapshots you want to use for the restoration process by running:\n\n ```\n kubectl get volumesnapshot -n APIGEE_NAMESPACE\n \n ```\n5. Install all Hybrid components one by one as described in\n [Step 10: Install Apigee hybrid using Helm](/apigee/docs/hybrid/v1.15/install-helm-charts). Note that the `apigee-cassandra-restore` pod will get\n created once you run the command to install the `datastore`, but it will only\n go into `running` state after you install the `apigee-org` component.\n\nSee [Cassandra backup overview](/apigee/docs/hybrid/v1.15/cassandra-backup-overview) for more details on Cassandra backup and restore.\n\nVerify the restoration job progress and confirm that `apigeeds` and all the other pods are up:\n\n1. Check `apigeeds`: \n\n ```\n kubectl get apigeeds -n APIGEE_NAMESPACE\n ```\n2. Check all other pods: \n\n ```\n kubectl get pods -n APIGEE_NAMESPACE\n ```\n\nUpon successful completion of the restore and confirmation that the runtime components are healthy,\nwe recommend configuring a backup on the cluster:\n\n1. Remove the `restore` configuration from the `overrides-restore.yaml` file.\n2. Add the `backup` configuration to the `overrides-restore.yaml` file.\n3. Apply the `backup` configuration with the following command: \n\n ```\n helm upgrade datastore apigee-datastore/ \\\n --namespace APIGEE_NAMESPACE \\\n --atomic \\\n -f overrides-restore.yaml\n ```\n\n\u003cbr /\u003e\n\n| **Warning:** See [known issue 391861216](/apigee/docs/release/known-issues#391861216) affecting Apigee Hybrid and Cassandra after restore using a remote server or Cloud Storage."]]