This page introduces you to your choices of a supported vector database on Vertex AI RAG Engine. You can also see how to connect a vector database (vector store) to your RAG corpus.
A common problem with LLMs is that they don't understand private knowledge, that is, your organization's data. With Vertex AI RAG Engine, you can enrich the LLM context with additional private information, because the model can reduce hallucination and answer questions more accurately.
Vector databases play a crucial role in enabling retrieval for RAG applications. Vector databases offer a specialized way to store and query vector embeddings, which are mathematical representations of text or other data that capture semantic meaning and relationships. Vector embeddings allow RAG systems to quickly and accurately find the most relevant information within a vast knowledge base, even when dealing with complex or nuanced queries. When combined with an embedding model, vector databases can help overcome the limitations of LLMs, and provide more accurate, relevant, and comprehensive responses.
Supported vector databases
When creating a RAG corpus, Vertex AI RAG Engine offers RagManagedDb
as the default
choice of a vector database, which requires no additional provisioning or
managing. For Vertex AI RAG Engine to automatically create and manage the vector
database for you, see Create a RAG
corpus.
In addition to the default RagManagedDb
, Vertex AI RAG Engine lets you provision
and bring your vector database for use within your RAG corpus. In this case,
you are responsible for the lifecycle and scalability of your vector database.
Compare vector database options
This table lists your choices of vector databases that are supported within Vertex AI RAG Engine and provides links to pages that explain how to use the vector databases within your RAG corpus.
To identify which vector database meets your needs, use this table to compare the vector-database options:
Vector database | Description | Benefits | Disadvantages | Supported distance metrics in Vertex AI RAG Engine | Search type | Launch stage in Vertex AI RAG Engine | Production ready | Best for |
---|---|---|---|---|---|---|---|---|
RagManagedDb (default) |
RagManagedDb is a regionally-distributed scalable database service that offers consistency and high availability.RagManagedDb can be used for a vector search.
|
|
|
cosine |
KNN | Preview |
|
|
Pinecone | Pinecone is a fully-managed cloud-native vector database designed for a high-performance similarity search. |
|
|
cosine euclidean dot-product |
ANN | GA |
|
|
Weaviate | Weaviate is an open-source vector database with a focus on flexibility and modularity. Weaviate supports various data types and offers built-in graph capabilities. |
|
|
cosine dot-product L2 squared hamming manhattan |
ANN + Hybrid search support | Preview |
|
|
Vector Search | Vector Search is the vector database service within Vertex AI. Vector Search is optimized for machine-learning tasks and integrates with other Google Cloud services. |
|
|
cosine dot-product |
ANN | GA |
|
|
Vertex AI Feature Store | Vertex AI Feature Store is a managed service for organizing, storing, and serving machine-learning features. Vertex AI Feature Store is optimized for machine-learning tasks and integrates with other Google Cloud services. |
|
|
cosine dot-product L2 squared |
ANN | Preview |
|