Starting April 29, 2025, Gemini 1.5 Pro and Gemini 1.5 Flash models are not available in projects that have no prior usage of these models, including new projects. For details, see Model versions and lifecycle.
Stay organized with collections
Save and categorize content based on your preferences.
Full name: projects.locations.publishers.models.predictLongRunning
Endpoint
post
https://aiplatform.googleapis.com/v1/{endpoint}:predictLongRunning
Path parameters
endpoint
string
Required. The name of the Endpoint requested to serve the prediction. Format: projects/{project}/locations/{location}/endpoints/{endpoint} or projects/{project}/locations/{location}/publishers/{publisher}/models/{model}
Request body
The request body contains data with the following structure:
Required. The instances that are the input to the prediction call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the prediction call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' Model'sPredictSchemata'sinstanceSchemaUri.
Optional. The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' Model's PredictSchemata'sparametersSchemaUri.
Response body
If successful, the response body contains an instance of Operation.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-06-27 UTC."],[],[],null,["# Method: models.predictLongRunning\n\n**Full name**: projects.locations.publishers.models.predictLongRunning \n\n### Endpoint\n\npost `https:``/``/aiplatform.googleapis.com``/v1``/{endpoint}:predictLongRunning` \n\n### Path parameters\n\n`endpoint` `string` \nRequired. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}`\n\n### Request body\n\nThe request body contains data with the following structure:\nFields `instances[]` `value (`[Value](https://protobuf.dev/reference/protobuf/google.protobuf/#value)` format)` \nRequired. The instances that are the input to the prediction call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the prediction call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' `Model's` `PredictSchemata's` `instanceSchemaUri`.\n`parameters` `value (`[Value](https://protobuf.dev/reference/protobuf/google.protobuf/#value)` format)` \nOptional. The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' `Model's ` `PredictSchemata's` `parametersSchemaUri`. \n\n### Response body\n\nIf successful, the response body contains an instance of [Operation](/vertex-ai/generative-ai/docs/reference/rest/Shared.Types/ListOperationsResponse#Operation)."]]