Vertex AI GenAI API

Service: aiplatform.googleapis.com

To call this service, we recommend that you use the Google-provided client libraries. If your application needs to use your own libraries to call this service, use the following information when you make the API requests.

Discovery document

A Discovery Document is a machine-readable specification for describing and consuming REST APIs. It is used to build client libraries, IDE plugins, and other tools that interact with Google APIs. One service may provide multiple discovery documents. This service provides the following discovery documents:

Service endpoint

A service endpoint is a base URL that specifies the network address of an API service. One service might have multiple service endpoints. This service has the following service endpoint and all URIs below are relative to this service endpoint:

REST Resource: v1.projects.locations

Methods
evaluateInstances POST /v1/{location}:evaluateInstances
Evaluates instances based on a given metric.

REST Resource: v1.projects.locations.endpoints

Methods
generateContent POST /v1/{model}:generateContent
Generate content with multimodal inputs.
predict POST /v1/{endpoint}:predict
Perform an online prediction.
serverStreamingPredict POST /v1/{endpoint}:serverStreamingPredict
Perform a server-side streaming online prediction request for Vertex LLM streaming.
streamGenerateContent POST /v1/{model}:streamGenerateContent
Generate content with multimodal inputs with streaming support.

REST Resource: v1.projects.locations.endpoints.chat

Methods
completions POST /v1/{endpoint}/chat/completions
Exposes an OpenAI-compatible endpoint for chat completions.

REST Resource: v1.projects.locations.models

Methods
getIamPolicy POST /v1/{resource}:getIamPolicy
Gets the access control policy for a resource.
setIamPolicy POST /v1/{resource}:setIamPolicy
Sets the access control policy on the specified resource.
testIamPermissions POST /v1/{resource}:testIamPermissions
Returns permissions that a caller has on the specified resource.

REST Resource: v1.projects.locations.operations

Methods
cancel POST /v1/{name}:cancel
Starts asynchronous cancellation on a long-running operation.
delete DELETE /v1/{name}
Deletes a long-running operation.
get GET /v1/{name}
Gets the latest state of a long-running operation.
list GET /v1/{name}/operations
Lists operations that match the specified filter in the request.
wait POST /v1/{name}:wait
Waits until the specified long-running operation is done or reaches at most a specified timeout, returning the latest state.

REST Resource: v1.projects.locations.publishers.models

Methods
generateContent POST /v1/{model}:generateContent
Generate content with multimodal inputs.
predict POST /v1/{endpoint}:predict
Perform an online prediction.
serverStreamingPredict POST /v1/{endpoint}:serverStreamingPredict
Perform a server-side streaming online prediction request for Vertex LLM streaming.
streamGenerateContent POST /v1/{model}:streamGenerateContent
Generate content with multimodal inputs with streaming support.

REST Resource: v1.projects.locations.tuningJobs

Methods
cancel POST /v1/{name}:cancel
Cancels a TuningJob.
create POST /v1/{parent}/tuningJobs
Creates a TuningJob.
get GET /v1/{name}
Gets a TuningJob.
list GET /v1/{parent}/tuningJobs
Lists TuningJobs in a Location.
rebaseTunedModel POST /v1/{parent}/tuningJobs:rebaseTunedModel
Rebase a TunedModel.

REST Resource: v1beta1.media

Methods
upload POST /v1beta1/{parent}/ragFiles:upload
POST /upload/v1beta1/{parent}/ragFiles:upload
Upload a file into a RagCorpus.

REST Resource: v1beta1.projects

Methods
getCacheConfig GET /v1beta1/{name}
Gets a GenAI cache config.
updateCacheConfig PATCH /v1beta1/{cacheConfig.name}
Updates a cache config.

REST Resource: v1beta1.projects.locations

Methods
augmentPrompt POST /v1beta1/{parent}:augmentPrompt
Given an input prompt, it returns augmented prompt from vertex rag store to guide LLM towards generating grounded responses.
corroborateContent POST /v1beta1/{parent}:corroborateContent
Given an input text, it returns a score that evaluates the factuality of the text.
evaluateInstances POST /v1beta1/{location}:evaluateInstances
Evaluates instances based on a given metric.
retrieveContexts POST /v1beta1/{parent}:retrieveContexts
Retrieves relevant contexts for a query.

REST Resource: v1beta1.projects.locations.cachedContents

Methods
create POST /v1beta1/{parent}/cachedContents
Creates cached content, this call will initialize the cached content in the data storage, and users need to pay for the cache data storage.
delete DELETE /v1beta1/{name}
Deletes cached content
get GET /v1beta1/{name}
Gets cached content configurations
list GET /v1beta1/{parent}/cachedContents
Lists cached contents in a project
patch PATCH /v1beta1/{cachedContent.name}
Updates cached content configurations

REST Resource: v1beta1.projects.locations.endpoints

Methods
countTokens POST /v1beta1/{endpoint}:countTokens
Perform a token counting.
generateContent POST /v1beta1/{model}:generateContent
Generate content with multimodal inputs.
getIamPolicy POST /v1beta1/{resource}:getIamPolicy
Gets the access control policy for a resource.
predict POST /v1beta1/{endpoint}:predict
Perform an online prediction.
serverStreamingPredict POST /v1beta1/{endpoint}:serverStreamingPredict
Perform a server-side streaming online prediction request for Vertex LLM streaming.
setIamPolicy POST /v1beta1/{resource}:setIamPolicy
Sets the access control policy on the specified resource.
streamGenerateContent POST /v1beta1/{model}:streamGenerateContent
Generate content with multimodal inputs with streaming support.
testIamPermissions POST /v1beta1/{resource}:testIamPermissions
Returns permissions that a caller has on the specified resource.

REST Resource: v1beta1.projects.locations.endpoints.chat

Methods
completions POST /v1beta1/{endpoint}/chat/completions
Exposes an OpenAI-compatible endpoint for chat completions.

REST Resource: v1beta1.projects.locations.extensions

Methods
delete DELETE /v1beta1/{name}
Deletes an Extension.
execute POST /v1beta1/{name}:execute
Executes the request against a given extension.
get GET /v1beta1/{name}
Gets an Extension.
import POST /v1beta1/{parent}/extensions:import
Imports an Extension.
list GET /v1beta1/{parent}/extensions
Lists Extensions in a location.
patch PATCH /v1beta1/{extension.name}
Updates an Extension.
query POST /v1beta1/{name}:query
Queries an extension with a default controller.

REST Resource: v1beta1.projects.locations.models

Methods
getIamPolicy POST /v1beta1/{resource}:getIamPolicy
Gets the access control policy for a resource.
setIamPolicy POST /v1beta1/{resource}:setIamPolicy
Sets the access control policy on the specified resource.
testIamPermissions POST /v1beta1/{resource}:testIamPermissions
Returns permissions that a caller has on the specified resource.

REST Resource: v1beta1.projects.locations.operations

Methods
cancel POST /v1beta1/{name}:cancel
Starts asynchronous cancellation on a long-running operation.
delete DELETE /v1beta1/{name}
Deletes a long-running operation.
get GET /v1beta1/{name}
Gets the latest state of a long-running operation.
list GET /v1beta1/{name}/operations
Lists operations that match the specified filter in the request.
wait POST /v1beta1/{name}:wait
Waits until the specified long-running operation is done or reaches at most a specified timeout, returning the latest state.

REST Resource: v1beta1.projects.locations.publishers

Methods
getIamPolicy POST /v1beta1/{resource}:getIamPolicy
Gets the access control policy for a resource.

REST Resource: v1beta1.projects.locations.publishers.models

Methods
countTokens POST /v1beta1/{endpoint}:countTokens
Perform a token counting.
generateContent POST /v1beta1/{model}:generateContent
Generate content with multimodal inputs.
getIamPolicy POST /v1beta1/{resource}:getIamPolicy
Gets the access control policy for a resource.
predict POST /v1beta1/{endpoint}:predict
Perform an online prediction.
serverStreamingPredict POST /v1beta1/{endpoint}:serverStreamingPredict
Perform a server-side streaming online prediction request for Vertex LLM streaming.
streamGenerateContent POST /v1beta1/{model}:streamGenerateContent
Generate content with multimodal inputs with streaming support.

REST Resource: v1beta1.projects.locations.ragCorpora

Methods
create POST /v1beta1/{parent}/ragCorpora
Creates a RagCorpus.
delete DELETE /v1beta1/{name}
Deletes a RagCorpus.
get GET /v1beta1/{name}
Gets a RagCorpus.
list GET /v1beta1/{parent}/ragCorpora
Lists RagCorpora in a Location.
patch PATCH /v1beta1/{ragCorpus.name}
Updates a RagCorpus.

REST Resource: v1beta1.projects.locations.ragCorpora.ragFiles

Methods
delete DELETE /v1beta1/{name}
Deletes a RagFile.
get GET /v1beta1/{name}
Gets a RagFile.
import POST /v1beta1/{parent}/ragFiles:import
Import files from Google Cloud Storage or Google Drive into a RagCorpus.
list GET /v1beta1/{parent}/ragFiles
Lists RagFiles in a RagCorpus.

REST Resource: v1beta1.projects.locations.reasoningEngines

Methods
create POST /v1beta1/{parent}/reasoningEngines
Creates a reasoning engine.
delete DELETE /v1beta1/{name}
Deletes a reasoning engine.
get GET /v1beta1/{name}
Gets a reasoning engine.
list GET /v1beta1/{parent}/reasoningEngines
Lists reasoning engines in a location.
patch PATCH /v1beta1/{reasoningEngine.name}
Updates a reasoning engine.
query POST /v1beta1/{name}:query
Queries using a reasoning engine.

REST Resource: v1beta1.projects.locations.tuningJobs

Methods
cancel POST /v1beta1/{name}:cancel
Cancels a TuningJob.
create POST /v1beta1/{parent}/tuningJobs
Creates a TuningJob.
get GET /v1beta1/{name}
Gets a TuningJob.
list GET /v1beta1/{parent}/tuningJobs
Lists TuningJobs in a Location.
rebaseTunedModel POST /v1beta1/{parent}/tuningJobs:rebaseTunedModel
Rebase a TunedModel.