This page explains how you can ground responses by using your data from Vertex AI Search (Preview).
Grounding Gemini to your data
If you want to do retrieval-augmented generation (RAG), connect your model to your website data or your sets of documents, then use Grounding with Vertex AI Search.
Grounding to your data supports a maximum of 10 Vertex AI Search data sources and can be combined with Grounding with Google Search.
Supported models
This section lists the models that support grounding with your data. To explore how each model generates grounded responses, follow these instructions:
Try a model listed in this table in the Google Cloud console.
Click the Grounding: Your data toggle to the on position.
Click Customize and a Customize Grounding pane displays.
Select Vertex AI Search.
In the Grounding with Vertex AI Search section, enter the path of the Vertex AI datastore. If you don't have a Vertex AI data store, create a data store. For more information, see Create a data store.
Click Save.
Model | Description | Try a model |
---|---|---|
Gemini 2.5 Pro |
Text, code, images, audio, video, video with audio, PDF Doesn't support dynamic retrieval. For more information, see Considerations. |
Try the Gemini 2.5 Pro model |
Gemini 2.0 Flash |
Text, code, images, audio, video, video with audio, PDF Doesn't support dynamic retrieval. For more information, see Considerations. |
Try the Gemini 2.0 Flash model |
Gemini 2.0 Flash-Lite |
Text, code, images, audio, video, video with audio, PDF Doesn't support dynamic retrieval. For more information, see Considerations. |
Try the Gemini 2.0 Flash-Lite model |
Prerequisites
Before you can ground model output to your data, do the following:
- Enable AI Applications and activate the API.
- Create a AI Applications data source and application.
See the Introduction to Vertex AI Search for more.
Enable AI Applications
In the Google Cloud console, go to the Agent Builder page.
Read and agree to the terms of service, then click Continue and activate the API.
AI Applications is available in the global
location, or the eu
and us
multi-region. To
learn more, see AI Applications locations
Create a data store in AI Applications
To create a data store in AI Applications, you can choose to ground with website data or documents.
Website
Open the Create Data Store page from the Google Cloud console.
In Website Content box, click Select.
Specify the websites for your data store pane displays.If Advanced website indexing isn't checked, then select the Advanced website indexing checkbox to turn it on.
Configure your data store pane displays.In the Specify URL patterns to index section, do the following:
- Add URLs for Sites to include.
- Optional: Add URLs for Sites to exclude.
Click Continue.
In the Configure your data store pane,
- Select a value from the Location of your data store list.
- Enter a name in the Your data store name field. The ID is generated. Use this ID when you generate your grounded responses with your data store. For more information, see Generate grounded responses with your data store.
- Click Create.
Documents
Open the Create Data Store page from the Google Cloud console.
In Cloud Storage box, click Select.
Import data from Cloud Storage pane displays.In the Unstructured documents (PDF, HTML, TXT and more) section, select Unstructured documents (PDF, HTML, TXT and more).
Select a Synchronization frequency option.
Select a Select a folder or a file you want to import option, and enter the path in the field.
Click Continue.
Configure your data store pane displays.In the Configure your data store pane,
- Select a value from the Location of your data store list.
- Enter a name in the Your data store name field. The ID is generated.
- To select parsing and chunking options for your documents, expand the Document Processing Options section. For more information about different parsers, see Parse documents.
- Click Create.
Click Create.
Generate grounded responses with your data store
Use the following instructions to ground a model with your data. A maximum of 10 data stores is supported.
If you don't know your data store ID, follow these steps:
In the Google Cloud console, go to the AI Applications page and in the navigation menu, click Data stores.
Click the name of your data store.
On the Data page for your data store, get the data store ID.
Console
To ground your model output to AI Applications by using Vertex AI Studio in the Google Cloud console, follow these steps:
- In the Google Cloud console, go to the Vertex AI Studio Freeform page.
- To turn on grounding, click the Grounding: your data toggle.
- Click Customize.
- Select Vertex AI Search as your source.
- Using this path format, replace your data store's Project ID and
the ID of the data store:
projects/project_id/locations/global/collections/default_collection/dataStores/data_store_id.
- Click Save.
- Enter your prompt in the text box, and click Submit.
Your prompt responses are grounded to AI Applications.
Python
Before trying this sample, follow the Python setup instructions in the Vertex AI quickstart using client libraries. For more information, see the Vertex AI Python API reference documentation.
To authenticate to Vertex AI, set up Application Default Credentials. For more information, see Set up authentication for a local development environment.
REST
To test a text prompt by using the Vertex AI API, send a POST request to the publisher model endpoint.
Before using any of the request data, make the following replacements:
- LOCATION: The region to process the request.
- PROJECT_ID: Your project ID.
- MODEL_ID: The model ID of the multimodal model.
- TEXT: The text instructions to include in the prompt.
HTTP method and URL:
POST https://LOCATION-aiplatform.googleapis.com/v1beta1/projects/PROJECT_ID/locations/LOCATION/publishers/google/models/MODEL_ID:generateContent
Request JSON body:
{ "contents": [{ "role": "user", "parts": [{ "text": "TEXT" }] }], "tools": [{ "retrieval": { "vertexAiSearch": { "datastore": projects/PROJECT_ID/locations/global/collections/default_collection/dataStores/DATA_STORE_ID } } }], "model": "projects/PROJECT_ID/locations/LOCATION/publishers/google/models/MODEL_ID" }
To send your request, expand one of these options:
You should receive a JSON response similar to the following:
{ "candidates": [ { "content": { "role": "model", "parts": [ { "text": "You can make an appointment on the website https://dmv.gov/" } ] }, "finishReason": "STOP", "safetyRatings": [ "..." ], "groundingMetadata": { "retrievalQueries": [ "How to make appointment to renew driving license?" ], "groundingChunks": [ { "retrievedContext": { "uri": "https://vertexaisearch.cloud.google.com/grounding-api-redirect/AXiHM.....QTN92V5ePQ==", "title": "dmv" } } ], "groundingSupport": [ { "segment": { "startIndex": 25, "endIndex": 147 }, "segment_text": "ipsum lorem ...", "supportChunkIndices": [1, 2], "confidenceScore": [0.9541752, 0.97726375] }, { "segment": { "startIndex": 294, "endIndex": 439 }, "segment_text": "ipsum lorem ...", "supportChunkIndices": [1], "confidenceScore": [0.9541752, 0.9325467] } ] } } ], "usageMetadata": { "..." } }
Understand your response
The response from both APIs include the LLM-generated text, which is called a candidate. If your model prompt successfully grounds to your Elasticsearch data source, then the responses include grounding metadata, which identifies the parts of the response that were derived from your Elasticsearch data. However, there are several reasons this metadata might not be provided, and the prompt response won't be grounded. These reasons include low-source relevance or incomplete information within the model's response.
The following is a breakdown of the output data:
- Role: Indicates the sender of the grounded answer. Because the response
always contains grounded text, the role is always
model
. - Text: The grounded answer generated by the LLM.
- Grounding metadata: Information about the grounding source, which contains
the following elements:
- Grounding chunks: A list of results from your Elasticsearch index that support the answer.
- Grounding supports: Information about a specific claim within the answer that can be used to show citations:
- Segment: The part of the model's answer that is substantiated by a grounding chunk.
- Grounding chunk index: The index of the grounding chunks in the grounding chunks list that corresponds to this claim.
- Confidence scores: A number from 0 to 1 that indicates how grounded the claim is in the provided set of grounding chunks.
What's next
- To learn how to send chat prompt requests, see Multiturn chat.
- To learn about responsible AI best practices and Vertex AI's safety filters,
see Safety best practices.