Starting April 29, 2025, Gemini 1.5 Pro and Gemini 1.5 Flash models are not available in projects that have no prior usage of these models, including new projects. For details, see Model versions and lifecycle.
Stay organized with collections
Save and categorize content based on your preferences.
Google was the first in the industry to publish an
AI/ML Privacy Commitment,
which outlines our belief that customers should have the highest level of
security and control over their data that is stored in the cloud. That commitment
extends to Google Cloud's generative AI products. Google ensures that its
teams are following these commitments through robust data governance practices,
which include reviews of the data that Google Cloud uses in the development of
its products. More details about how Google processes data can also be found in
Google's Cloud Data Processing Addendum (CDPA).
Training restriction
As outlined in Section 17 "Training Restriction" in the Service Terms section of
Service Specific Terms,
Google won't use your data to train or fine-tune any AI/ML models without your
prior permission or instruction. This applies to all managed models on
Vertex AI, including GA and pre-GA models.
Customer data retention and achieving zero data retention
Customer data is retained in Vertex AI for Google models for limited
periods of time in the following scenarios and conditions. To achieve zero data retention, customers must take specific actions within each of these areas:
Data caching for Google models: By default, Google foundation models cache
inputs for Gemini models. This is done to reduce latency
and accelerate responses to subsequent prompts from the customer. Cached
contents are stored for up to 24 hours in the data center where the request
was served. Data caching is enabled or disabled at the Google Cloud project
level, and project-level privacy is enforced for cached data. The same cache
settings for a Google Cloud project apply to all regions. To achieve zero data retention, you must disable data caching. See
Enabling and disabling data caching.
Prompt logging for abuse monitoring for Google models: As outlined in
Section 4.3 "Generative AI Safety and Abuse" of
Google Cloud Platform Terms of Service,
Google may log prompts to detect potential abuse and violations of its
Acceptable Use Policy and
Prohibited Use Policy
as part of providing generative AI services to customers. Only customers
whose use of Google Cloud is governed by the
Google Cloud Platform Terms of Service
and who don't have an
Invoiced Cloud Billing account
are subject to prompt logging for abuse monitoring. If you are in scope for prompt logging for abuse monitoring and want zero data retention, you can request an exception for abuse monitoring. See
Abuse monitoring.
Grounding with Google Search: As outlined in Section
19 "Generative AI Services: Grounding with Google Search" of the
Service Specific Terms,
Google stores prompts and contextual information that customers may provide,
and generated output for thirty (30) days for the purposes of creating
grounded results and search suggestions, and this stored information may be
used for debugging and testing of systems that support grounding with
Google Search. There is no way to disable the storage of this information if you use Grounding with Google Search.
Session resumption for Gemini Live API: This feature is disabled by
default. It must be enabled by the user every time they call the API by
specifying the field in the API request, and project-level privacy is
enforced for cached data. Enabling Session Resumption allows the user to
reconnect to a previous session within 24 hours by storing cached data,
including text, video, and audio prompt data and model outputs, for up to 24
hours. To achieve zero data retention, do not enable this feature. For more
information about this feature, including how to enable it, see
Live API.
This applies to all managed models on Vertex AI, including GA and
pre-GA models.
Enabling and disabling data caching
You can use the following curl commands to get
caching status, disable caching, or re-enable caching.
When you disable or re-enable caching, the change
applies to all Google Cloud regions. For more information about using
Identity and Access Management to grant permissions required to enable or disable caching, see
Vertex AI access control with IAM.
Expand the following sections to learn how to get the current cache setting, to
disable caching, and to enable caching.
Get current caching setting
Run the following command to determine if caching is enabled or disabled for a
project. To run this command, a user must be granted one of the following
roles: roles/aiplatform.viewer, roles/aiplatform.user, or
roles/aiplatform.admin.
PROJECT_ID=PROJECT_ID
# Setup project_id
$ gcloud config set project PROJECT_ID
# GetCacheConfig
$ curl -X GET -H "Authorization: Bearer $(gcloud auth application-default print-access-token)" -H "Content-Type: application/json" https://us-central1-aiplatform.googleapis.com/v1/projects/PROJECT_ID/cacheConfig
# Response if caching is enabled (caching is enabled by default).
{
"name": "projects/PROJECT_ID/cacheConfig"
}
# Response if caching is disabled.
{
"name": "projects/PROJECT_ID/cacheConfig"
"disableCache": true
}
Disable caching
Run the following curl command to disable caching for a Google Cloud project. To run
this command, a user must be granted the Vertex AI administrator role,
roles/aiplatform.admin.
If you disabled caching for a Google Cloud project and want re-enable it, run the
following curl command. To run this command, a user must be granted the
Vertex AI administrator role, roles/aiplatform.admin.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-08-29 UTC."],[],[],null,["# Generative AI and zero data retention\n\nGoogle was the first in the industry to publish an\n[AI/ML Privacy Commitment](https://cloud.google.com/blog/products/ai-machine-learning/google-cloud-unveils-ai-and-ml-privacy-commitment),\nwhich outlines our belief that customers should have the highest level of\nsecurity and control over their data that is stored in the cloud. That commitment\nextends to Google Cloud's generative AI products. Google ensures that its\nteams are following these commitments through robust data governance practices,\nwhich include reviews of the data that Google Cloud uses in the development of\nits products. More details about how Google processes data can also be found in\nGoogle's [Cloud Data Processing Addendum (CDPA)](https://cloud.google.com/terms/data-processing-addendum).\n\nTraining restriction\n--------------------\n\nAs outlined in Section 17 \"Training Restriction\" in the Service Terms section of\n[Service Specific Terms](https://cloud.google.com/terms/service-terms),\nGoogle won't use your data to train or fine-tune any AI/ML models without your\nprior permission or instruction. This applies to all managed models on\nVertex AI, including GA and pre-GA models.\n\nCustomer data retention and achieving zero data retention\n---------------------------------------------------------\n\nCustomer data is retained in Vertex AI for Google models for limited\nperiods of time in the following scenarios and conditions. To achieve zero data retention, customers must take specific actions within each of these areas:\n\n- **Data caching for Google models** : By default, Google foundation models cache inputs for Gemini models. This is done to reduce latency and accelerate responses to subsequent prompts from the customer. Cached contents are stored for up to 24 hours in the data center where the request was served. Data caching is enabled or disabled at the Google Cloud project level, and project-level privacy is enforced for cached data. The same cache settings for a Google Cloud project apply to all regions. To achieve zero data retention, you must disable data caching. See [Enabling and disabling data caching](#enabling-disabling-caching).\n- **Prompt logging for abuse monitoring for Google models** : As outlined in Section 4.3 \"Generative AI Safety and Abuse\" of [Google Cloud Platform Terms of Service](https://cloud.google.com/terms), Google may log prompts to detect potential abuse and violations of its [Acceptable Use Policy](https://cloud.google.com/terms/aup) and [Prohibited Use Policy](https://policies.google.com/terms/generative-ai/use-policy) as part of providing generative AI services to customers. Only customers whose use of Google Cloud is governed by the [Google Cloud Platform Terms of Service](https://cloud.google.com/terms) and who don't have an [Invoiced Cloud Billing account](/billing/docs/concepts#billing_account_types) are subject to prompt logging for abuse monitoring. If you are in scope for prompt logging for abuse monitoring and want zero data retention, you can request an exception for abuse monitoring. See [Abuse monitoring](/vertex-ai/generative-ai/docs/learn/abuse-monitoring).\n- **Grounding with Google Search** : As outlined in Section 19 \"Generative AI Services: Grounding with Google Search\" of the [Service Specific Terms](https://cloud.google.com/terms/service-terms), Google stores prompts and contextual information that customers may provide, and generated output for thirty (30) days for the purposes of creating grounded results and search suggestions, and this stored information may be used for debugging and testing of systems that support grounding with Google Search. There is no way to disable the storage of this information if you use Grounding with Google Search.\n- **Session resumption for Gemini Live API:** This feature is disabled by default. It must be enabled by the user every time they call the API by specifying the field in the API request, and project-level privacy is enforced for cached data. Enabling Session Resumption allows the user to reconnect to a previous session within 24 hours by storing cached data, including text, video, and audio prompt data and model outputs, for up to 24 hours. To achieve zero data retention, do not enable this feature. For more information about this feature, including how to enable it, see [Live API](/vertex-ai/generative-ai/docs/live-api#session-resumption).\n\nThis applies to all managed models on Vertex AI, including GA and\npre-GA models.\n\n### Enabling and disabling data caching\n\nYou can use the following curl commands to get\ncaching status, disable caching, or re-enable caching.\nWhen you disable or re-enable caching, the change\napplies to all Google Cloud regions. For more information about using\nIdentity and Access Management to grant permissions required to enable or disable caching, see\n[Vertex AI access control with IAM](/vertex-ai/docs/general/access-control).\nExpand the following sections to learn how to get the current cache setting, to\ndisable caching, and to enable caching. \n\n#### Get current caching setting\n\nRun the following command to determine if caching is enabled or disabled for a\nproject. To run this command, a user must be granted one of the following\nroles: `roles/aiplatform.viewer`, `roles/aiplatform.user`, or\n`roles/aiplatform.admin`. \n\n```\nPROJECT_ID=PROJECT_ID\n# Setup project_id\n$ gcloud config set project PROJECT_ID\n\n# GetCacheConfig\n$ curl -X GET -H \"Authorization: Bearer $(gcloud auth application-default print-access-token)\" -H \"Content-Type: application/json\" https://us-central1-aiplatform.googleapis.com/v1/projects/PROJECT_ID/cacheConfig\n\n# Response if caching is enabled (caching is enabled by default).\n{\n \"name\": \"projects/PROJECT_ID/cacheConfig\"\n}\n\n# Response if caching is disabled.\n{\n \"name\": \"projects/PROJECT_ID/cacheConfig\"\n \"disableCache\": true\n}\n \n``` \n\n#### Disable caching\n\nRun the following curl command to disable caching for a Google Cloud project. To run\nthis command, a user must be granted the Vertex AI administrator role,\n`roles/aiplatform.admin`. \n\n```\nPROJECT_ID=PROJECT_ID\n# Setup project_id\n$ gcloud config set project PROJECT_ID\n\n# Setup project_id.\n$ gcloud config set project ${PROJECT_ID}\n\n# Opt-out of caching.\n$ curl -X PATCH -H \"Authorization: Bearer $(gcloud auth application-default print-access-token)\" -H \"Content-Type: application/json\" https://us-central1-aiplatform.googleapis.com/v1/projects/PROJECT_ID/cacheConfig -d '{\n \"name\": \"projects/PROJECT_ID/cacheConfig\",\n \"disableCache\": true\n}'\n\n# Response.\n{\n \"name\": \"projects/PROJECT_ID/locations/us-central1/projects/PROJECT_ID/cacheConfig/operations/${OPERATION_ID}\",\n \"done\": true,\n \"response\": {\n \"@type\": \"type.googleapis.com/google.protobuf.Empty\"\n }\n}\n \n``` \n\n#### Enable caching\n\nIf you disabled caching for a Google Cloud project and want re-enable it, run the\nfollowing curl command. To run this command, a user must be granted the\nVertex AI administrator role, `roles/aiplatform.admin`. \n\n```\nPROJECT_ID=PROJECT_ID\nLOCATION_ID=\"us-central1\"\n# Setup project_id\n$ gcloud config set project PROJECT_ID\n\n# Setup project_id.\n$ gcloud config set project ${PROJECT_ID}\n\n# Opt in to caching.\n$ curl -X PATCH -H \"Authorization: Bearer $(gcloud auth application-default print-access-token)\" -H \"Content-Type: application/json\" https://us-central1-aiplatform.googleapis.com/v1/projects/PROJECT_ID/cacheConfig -d '{\n \"name\": \"projects/PROJECT_ID/cacheConfig\",\n \"disableCache\": false\n}'\n\n# Response.\n{\n \"name\": \"projects/PROJECT_ID/locations/us-central1/projects/PROJECT_ID/cacheConfig/operations/${OPERATION_NUMBER}\",\n \"done\": true,\n \"response\": {\n \"@type\": \"type.googleapis.com/google.protobuf.Empty\"\n }\n}\n \n```\n\nWhat's next\n-----------\n\n- Learn about [responsible AI best practices and Vertex AI's safety filters](/vertex-ai/generative-ai/docs/learn/responsible-ai).\n- Learn about [Gemini in Google Cloud data governance](/gemini/docs/discover/data-governance)."]]