Starting April 29, 2025, Gemini 1.5 Pro and Gemini 1.5 Flash models are not available in projects that have no prior usage of these models, including new projects. For details, see Model versions and lifecycle.
Stay organized with collections
Save and categorize content based on your preferences.
Last updated: June 25, 2025
Gemini 2 general FAQ
Help! The model I'm using isn't available anymore!
If your application recently started showing errors related to an unavailable
Palm2, Gemini 1.0, or Gemini 1.5-001 models, this document
covers how you can transition to a supported model.
Google regularly releases new and improved AI models. To make way for these
advancements, older models are retired (or deprecated). We provide
notice when deprecating a model and a transition window before access to the
model is terminated, but we understand it can still cause interruptions.
Here are two options for updating your model:
The quickest way to resolve the error is usually to switch your
application code to one of the supported models.
Test all critical features to make sure everything works as expected, then launch the change like you normally would.
If you have a bit more time, consider our
step-by-step migration process.
This walks you through upgrading to the latest Gemini SDK and
includes best practices to help minimize any risks during the switch.
You can use this approach to minimize any potential risks of model
migration and make sure that you are using the new model in an optimal
way.
How do the Gemini 2 models compare to the 1.5 generation?
The Gemini 2 models feature the following upgrades over our 1.5 models:
Improved multilingual capabilities: Gemini 2 models show
strong advancements in multilingual understanding, with increased scores
in the Global MMLU (Lite) benchmark.
Significant gains in reasoning and knowledge factuality:
Gemini 2.5 Pro shows substantial improvements in GPQA
(domain expert reasoning) and SimpleQA (world knowledge factuality)
indicating enhanced ability to understand and provide accurate
information.
Enhanced mathematical problem-solving: Both
Gemini 2.0 Flash and Gemini 2.5 Pro
demonstrate notable progress in handling complex mathematical problems, as
evidenced by the MATH and HiddenMath benchmarks.
The following table shows the comparison between our Gemini 2 models:
Model name
Description
Upgrade path for
Gemini 2.5 Pro
Strongest model quality (especially for code and world knowledge), with a 1M token-long context window
Gemini 1.5 Pro users who want better quality, or who are particularly invested in long context and code
Gemini 2.0 Flash
Workhorse model for all daily tasks and features enhanced performance and supports real-time Live API
Gemini 1.5 Flash users who want a model with
significantly better quality that's slightly slower
Gemini 1.5 Pro users who want slightly better
quality and real-time latency
Gemini 2.0 Flash-Lite
Our most cost effective offering to support high throughput
Gemini 1.5 Flash users who want better quality
for the same price and speed
Customers looking for the fastest model in the Gemini 2
family
How do I migrate Gemini on Google AI Studio to Vertex AI Studio?
Migrating to Google Cloud's Vertex AI platform offers a suite of
MLOps tools that streamline the usage, deployment, and monitoring of AI
models for efficiency and reliability. To migrate your work to
Vertex AI, import and upload your existing data to Vertex AI Studio and use the Gemini API with Vertex AI.
How does Gemini 2 image generation compare to Imagen 3?
While the experimental version of Gemini 2.0 Flash
supports image generation, Gemini 2 does not currently support
image generation in our generally available models. The experimental version
of Gemini 2.0 Flash shouldn't be used in production-level
code.
If you need image generation in production code, use
Imagen 3. This powerful model offers
high-quality images, low-latency generation, and flexible editing options.
Does Gemini 2 in Vertex AI support compositional function calling?
For the full list of locations that are supported for Gemini 2
models, see Locations.
What are the default quotas for Gemini 2?
Gemini 2.0 Flash and Gemini 2.0 Flash-Lite
use dynamic shared quota and
have no default quota.
Gemini 2.5 Pro is an experimental model and has a
10 queries per minute (QPM) limit.
Monitoring
Why does my quota usage show as 0% percent on API dashboard when I'm sending requests?
For Gemini models on Vertex, we use a Dynamic Shared
Quota (DSQ) system. This innovative approach automatically manages
capacity across all users in a region, ensuring optimal performance without
the need for manual quota adjustments or requests. As a result, you won't
see traditional quota usage displayed in the Quotas & System Limits
tab. Your project will automatically receive the necessary resources based
on real-time availability.
For generative AI applications in production requiring consistent
throughput, we recommend using Provisioned Throughput (PT). PT ensures a
predictable and consistent user experience, critical for time-sensitive
workloads. Additionally, it provides deterministic monthly or weekly cost
structures, enabling accurate budget planning.
What models are supported for Provisioned Throughput?
The list of models supported for Provisioned Throughput, including
throughput, purchase increment, and burndown rate is listed on our
Supported
models page.
To purchase Provisioned Throughput for partner models (such as
Anthropic's Claude models), you must contact Google; you can't order through
the Google Cloud console. For more information, see
Partner models.
How can I monitor my Provisioned Throughput usage?
There are three ways to measure your Provisioned Throughput usage:
When using the built-in monitoring metrics or HTTP response headers, you
can create a chart in the
Metrics Explorer to monitor usage.
What permissions are required to purchase and use Provisioned Throughput?
To buy and manage Provisioned Throughput, follow the instructions in the Permissions section of Purchase Provisioned Throughput. The same permissions for pay-as-you-go apply for Provisioned Throughput usage.
If you still run into issues placing an order, you likely need to add one of
the following roles:
Vertex AI Administrator
Vertex AI Platform Provisioned Throughput Admin
What is a GSU?
A generative AI scale unit (GSU) is an abstract measure of capacity for throughput provisioning
that is fixed and standard across all Google models that support Provisioned
Throughput. A GSU's price and capacity is fixed, but throughput may vary
between models because different models may require different amounts of
capacity to deliver the same throughput.
How can I estimate my GSU needs for Provisioned Throughput?
You can estimate your Provisioned Throughput needs by:
Gather your requirements
Calculate your throughput:
$$
\begin{aligned}
\text{Throughput per sec} = & \\
& \qquad (\text{Inputs per query converted to input chars} \\
& \qquad + \text{Outputs per query converted to input chars}) \\
& \qquad \times \text{QPS}
\end{aligned}
$$
Calculate your GSUs: Use the estimation tool provided in the purchasing console
to calculate the corresponding number of GSUs needed to cover that usage
for the given model and details.
How often am I billed for Provisioned Throughput?
You're invoiced for any charges you incur for Provisioned Throughput usage
over the course of the month at the end of that month.
How long does it take to activate my Provisioned Throughput order?
For small orders or small incremental increases, the order
will be auto-approved and ready within minutes if capacity is available.
Larger increases or orders may take longer and may require us to
communicate with you directly in order to prepare capacity for your order.
We aim to have a decision on each request (either approved or denied)
within 1 week of order submission.
Can I test Provisioned Throughput before placing an order?
While a direct test environment is not available, a 1-week order with a
limited number of GSUs provides a cost-effective way to experience its
benefits and assess its suitability for your requirements.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-08-28 UTC."],[],[],null,["# Frequently asked questions\n\nLast updated: June 25, 2025\n\nGemini 2 general FAQ\n--------------------\n\n### Help! The model I'm using isn't available anymore!\n\n\nIf your application recently started showing errors related to an unavailable\nPalm2, Gemini 1.0, or Gemini 1.5-001 models, this document\ncovers how you can transition to a supported model.\n\n\nGoogle regularly releases new and improved AI models. To make way for these\nadvancements, older models are retired (or *deprecated*). We provide\nnotice when deprecating a model and a transition window before access to the\nmodel is terminated, but we understand it can still cause interruptions.\n\nHere are two options for updating your model:\n\n1. The quickest way to resolve the error is usually to switch your application code to one of the [supported models](/vertex-ai/generative-ai/docs/learn/model-versions). Test all critical features to make sure everything works as expected, then launch the change like you normally would.\n2. If you have a bit more time, consider our [step-by-step migration process](/vertex-ai/generative-ai/docs/migrate-to-v2). This walks you through upgrading to the latest Gemini SDK and includes best practices to help minimize any risks during the switch. You can use this approach to minimize any potential risks of model migration and make sure that you are using the new model in an optimal way.\n\n### How do the Gemini 2 models compare to the 1.5 generation?\n\nThe Gemini 2 models feature the following upgrades over our 1.5 models:\n\n- **Improved multilingual capabilities:** Gemini 2 models show strong advancements in multilingual understanding, with increased scores in the Global MMLU (Lite) benchmark.\n- **Significant gains in reasoning and knowledge factuality:** Gemini 2.5 Pro shows substantial improvements in GPQA (domain expert reasoning) and SimpleQA (world knowledge factuality) indicating enhanced ability to understand and provide accurate information.\n- **Enhanced mathematical problem-solving:** Both Gemini 2.0 Flash and Gemini 2.5 Pro demonstrate notable progress in handling complex mathematical problems, as evidenced by the MATH and HiddenMath benchmarks.\n\nThe following table shows the comparison between our Gemini 2 models:\n\n\nTo see all benchmark capabilities for Gemini 2, visit the\n[Google DeepMind\ndocumentation](https://deepmind.google/technologies/gemini/).\n\n### How do I migrate Gemini on Google AI Studio to Vertex AI Studio?\n\n\nMigrating to Google Cloud's Vertex AI platform offers a suite of\nMLOps tools that streamline the usage, deployment, and monitoring of AI\nmodels for efficiency and reliability. To migrate your work to\nVertex AI, import and upload your existing data to Vertex AI Studio and use the Gemini API with Vertex AI.\n\n\nFor more information, see [Migrate from\nGemini on Google AI to Vertex AI](/vertex-ai/generative-ai/docs/migrate/migrate-google-ai).\n\n### How does Gemini 2 image generation compare to Imagen 3?\n\n\nWhile the experimental version of Gemini 2.0 Flash\nsupports image generation, Gemini 2 does not currently support\nimage generation in our generally available models. The experimental version\nof Gemini 2.0 Flash shouldn't be used in production-level\ncode.\n\n\nIf you need image generation in production code, use\n[Imagen 3](/vertex-ai/generative-ai/docs/image/overview). This powerful model offers\nhigh-quality images, low-latency generation, and flexible editing options.\n\n### Does Gemini 2 in Vertex AI support compositional function calling?\n\n\n[Compositional function calling](https://ai.google.dev/gemini-api/docs/function-calling#compositional-function-calling)\nis only available in [Google AI Studio](https://aistudio.google.com).\n\n### What locations are supported for Gemini 2?\n\n\nFor the full list of locations that are supported for Gemini 2\nmodels, see [Locations](/vertex-ai/generative-ai/docs/learn/locations).\n\n### What are the default quotas for Gemini 2?\n\n\n**Gemini 2.0 Flash** and **Gemini 2.0 Flash-Lite**\nuse [dynamic shared quota](/vertex-ai/generative-ai/docs/dsq) and\nhave no default quota.\n\n\n**Gemini 2.5 Pro** is an experimental model and has a\n10 queries per minute (QPM) limit.\n\nMonitoring\n----------\n\n### Why does my quota usage show as 0% percent on API dashboard when I'm sending requests?\n\n\nFor Gemini models on Vertex, we use a [Dynamic Shared\nQuota (DSQ) system](/vertex-ai/generative-ai/docs/dsq). This innovative approach automatically manages\ncapacity across all users in a region, ensuring optimal performance without\nthe need for manual quota adjustments or requests. As a result, you won't\nsee traditional quota usage displayed in the **Quotas \\& System Limits**\ntab. Your project will automatically receive the necessary resources based\non real-time availability.\n\n\nUse the [**Vertex AI Model Garden (Monitoring)**\ndashboard](https://console.cloud.google.com/monitoring/dashboards/integration/vertex_ai.vertex-ai-model-garden) to monitor usage.\n\nProvisioned Throughput\n----------------------\n\n### When should I use Provisioned Throughput?\n\n\nFor generative AI applications in production requiring consistent\nthroughput, we recommend using Provisioned Throughput (PT). PT ensures a\npredictable and consistent user experience, critical for time-sensitive\nworkloads. Additionally, it provides deterministic monthly or weekly cost\nstructures, enabling accurate budget planning.\n\n\nFor more information, see [Provisioned Throughput overview](/vertex-ai/generative-ai/docs/provisioned-throughput#when-to-use-provisioned-throughput).\n\n### What models are supported for Provisioned Throughput?\n\n\nThe list of models supported for Provisioned Throughput, including\nthroughput, purchase increment, and burndown rate is listed on our\n[Supported\nmodels](/vertex-ai/generative-ai/docs/supported-models#google-models) page.\n\n\nTo purchase Provisioned Throughput for partner models (such as\nAnthropic's Claude models), you must contact Google; you can't order through\nthe Google Cloud console. For more information, see\n[Partner models](/vertex-ai/generative-ai/docs/provisioned-throughput/supported-models#partner-models).\n\n### How can I monitor my Provisioned Throughput usage?\n\n\nThere are three ways to measure your Provisioned Throughput usage:\n\n- Use the [Model Garden monitoring dashboard](https://console.cloud.google.com/monitoring/dashboards/integration/vertex_ai.vertex-ai-model-garden)\n- Use the built-in [](/vertex-ai/generative-ai/docs/use-provisioned-throughput#metrics)monitoring metrics\n- Use the [HTTP response headers](/vertex-ai/generative-ai/docs/use-provisioned-throughput#response_headers)\n\n\nWhen using the built-in monitoring metrics or HTTP response headers, you\ncan [create a chart](/monitoring/charts/metrics-explorer) in the\nMetrics Explorer to monitor usage.\n\n### What permissions are required to purchase and use Provisioned Throughput?\n\n\nTo buy and manage Provisioned Throughput, follow the instructions in the Permissions section of [Purchase Provisioned Throughput](/vertex-ai/generative-ai/docs/purchase-provisioned-throughput#permissions). The same permissions for pay-as-you-go apply for Provisioned Throughput usage.\n\n\nIf you still run into issues placing an order, you likely need to add one of\nthe following roles:\n\n- Vertex AI Administrator\n- Vertex AI Platform Provisioned Throughput Admin\n\n### What is a GSU?\n\n\nA [*generative AI scale unit*](/vertex-ai/generative-ai/docs/measure-provisioned-throughput#gsu-burndown-rate) (GSU) is an abstract measure of capacity for throughput provisioning\nthat is fixed and standard across all Google models that support Provisioned\nThroughput. A GSU's price and capacity is fixed, but throughput may vary\nbetween models because different models may require different amounts of\ncapacity to deliver the same throughput.\n\n### How can I estimate my GSU needs for Provisioned Throughput?\n\n\nYou can estimate your Provisioned Throughput needs by:\n\n- Gather your requirements\n- **Calculate your throughput:** \n $$ \\\\begin{aligned} \\\\text{Throughput per sec} = \\& \\\\\\\\ \\& \\\\qquad (\\\\text{Inputs per query converted to input chars} \\\\\\\\ \\& \\\\qquad + \\\\text{Outputs per query converted to input chars}) \\\\\\\\ \\& \\\\qquad \\\\times \\\\text{QPS} \\\\end{aligned} $$\n- **Calculate your GSUs:** Use the [estimation tool](https://console.cloud.google.com/vertex-ai/provisioned-throughput/price-estimate) provided in the purchasing console to calculate the corresponding number of GSUs needed to cover that usage for the given model and details.\n\n| **Important:** Review the detailed steps in our [example of estimating your Provisioned Throughput needs](https://console.cloud.google.com/vertex-ai/provisioned-throughput/price-estimate).\n\n### How often am I billed for Provisioned Throughput?\n\n\nYou're invoiced for any charges you incur for Provisioned Throughput usage\nover the course of the month at the end of that month.\n\n### How long does it take to activate my Provisioned Throughput order?\n\n- For **small orders** or **small incremental increases**, the order will be auto-approved and ready within minutes if capacity is available.\n- **Larger increases or orders** may take longer and may require us to communicate with you directly in order to prepare capacity for your order. We aim to have a decision on each request (either approved or denied) within 1 week of order submission.\n\n### Can I test Provisioned Throughput before placing an order?\n\n\nWhile a direct test environment is not available, a 1-week order with a\nlimited number of GSUs provides a cost-effective way to experience its\nbenefits and assess its suitability for your requirements.\n\n\nFor more information, see\n[Purchase Provisioned Throughput](/vertex-ai/generative-ai/docs/purchase-provisioned-throughput#place-an-order)."]]