Starting April 29, 2025, Gemini 1.5 Pro and Gemini 1.5 Flash models are not available in projects that have no prior usage of these models, including new projects. For details, see Model versions and lifecycle.
Stay organized with collections
Save and categorize content based on your preferences.
OpenAI models on Vertex AI offer fully managed and serverless
models as APIs. To use an OpenAI model on Vertex AI, send
a request directly to the Vertex AI API endpoint. Because
OpenAI models use a managed API, there's no need to provision or
manage infrastructure.
You can stream your responses to reduce the end-user latency perception. A
streamed response uses server-sent events (SSE) to incrementally stream the
response.
Available OpenAI models
The following models are available from OpenAI to use in
Vertex AI. To access an OpenAI model, go to its
Model Garden model card.
gpt-oss 120B
OpenAI gpt-oss 120B is a 120B open-weight language model
released under the Apache
2.0 license. It is well-suited for reasoning and function calling use cases. The
model is optimized for deployment on consumer hardware.
The 120B model achieves near-parity with OpenAI o4-mini on core reasoning
benchmarks, while running on a single 80GB GPU.
OpenAI gpt-oss 20B is a 20B open-weight language model
released under the Apache
2.0 license. It is well-suited for reasoning and function calling use cases. The
model is optimized for deployment on consumer hardware.
The 20B model delivers similar results to OpenAI o3-mini on common benchmarks
and can run on edge devices with 16GB of memory, making it ideal for on-device
use cases, local inference, or rapid iteration without costly infrastructure.
To use OpenAI models with Vertex AI, you must perform the
following steps. The Vertex AI API
(aiplatform.googleapis.com) must be enabled to use
Vertex AI. If you already have an existing project with the
Vertex AI API enabled, you can use that project instead of creating a
new project.
Sign in to your Google Cloud account. If you're new to
Google Cloud,
create an account to evaluate how our products perform in
real-world scenarios. New customers also get $300 in free credits to
run, test, and deploy workloads.
In the Google Cloud console, on the project selector page,
select or create a Google Cloud project.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-09-02 UTC."],[],[],null,["# OpenAI models\n\n| **Note:** OpenAI models are not a Google product, and its availability in Vertex AI is subject to the terms for \"Separate Offerings\" in the AI/ML Services section of the [Service Specific\n| Terms](/terms/service-terms), and separate terms found in the relevant model card.\n\nOpenAI models on Vertex AI offer fully managed and serverless\nmodels as APIs. To use an OpenAI model on Vertex AI, send\na request directly to the Vertex AI API endpoint. Because\nOpenAI models use a managed API, there's no need to provision or\nmanage infrastructure.\n\nYou can stream your responses to reduce the end-user latency perception. A\nstreamed response uses *server-sent events* (SSE) to incrementally stream the\nresponse.\n\nAvailable OpenAI models\n-----------------------\n\nThe following models are available from OpenAI to use in\nVertex AI. To access an OpenAI model, go to its\nModel Garden model card.\n\n### gpt-oss 120B\n\nOpenAI gpt-oss 120B is a 120B open-weight language model\nreleased under the Apache\n2.0 license. It is well-suited for reasoning and function calling use cases. The\nmodel is optimized for deployment on consumer hardware.\n\nThe 120B model achieves near-parity with OpenAI o4-mini on core reasoning\nbenchmarks, while running on a single 80GB GPU.\n\n[Go to the gpt-oss 120B model card](https://console.cloud.google.com/vertex-ai/publishers/openai/model-garden/gpt-oss-120b-maas)\n\n### gpt-oss 20B\n\nOpenAI gpt-oss 20B is a 20B open-weight language model\nreleased under the Apache\n2.0 license. It is well-suited for reasoning and function calling use cases. The\nmodel is optimized for deployment on consumer hardware.\n\nThe 20B model delivers similar results to OpenAI o3-mini on common benchmarks\nand can run on edge devices with 16GB of memory, making it ideal for on-device\nuse cases, local inference, or rapid iteration without costly infrastructure.\n\n[Go to the gpt-oss 20B model card](https://console.cloud.google.com/vertex-ai/publishers/openai/model-garden/gpt-oss-120b-maas)\n\n### Before you begin\n\nTo use OpenAI models with Vertex AI, you must perform the\nfollowing steps. The Vertex AI API\n(`aiplatform.googleapis.com`) must be enabled to use\nVertex AI. If you already have an existing project with the\nVertex AI API enabled, you can use that project instead of creating a\nnew project.\n\n- Sign in to your Google Cloud account. If you're new to Google Cloud, [create an account](https://console.cloud.google.com/freetrial) to evaluate how our products perform in real-world scenarios. New customers also get $300 in free credits to run, test, and deploy workloads.\n- In the Google Cloud console, on the project selector page,\n select or create a Google Cloud project.\n\n [Go to project selector](https://console.cloud.google.com/projectselector2/home/dashboard)\n-\n [Verify that billing is enabled for your Google Cloud project](/billing/docs/how-to/verify-billing-enabled#confirm_billing_is_enabled_on_a_project).\n\n-\n\n\n Enable the Vertex AI API.\n\n\n [Enable the API](https://console.cloud.google.com/flows/enableapi?apiid=aiplatform.googleapis.com)\n\n- In the Google Cloud console, on the project selector page,\n select or create a Google Cloud project.\n\n [Go to project selector](https://console.cloud.google.com/projectselector2/home/dashboard)\n-\n [Verify that billing is enabled for your Google Cloud project](/billing/docs/how-to/verify-billing-enabled#confirm_billing_is_enabled_on_a_project).\n\n-\n\n\n Enable the Vertex AI API.\n\n\n [Enable the API](https://console.cloud.google.com/flows/enableapi?apiid=aiplatform.googleapis.com)\n\n1. Go to one of the following Model Garden model cards, then click **Enable**."]]