Stay organized with collections
Save and categorize content based on your preferences.
Vertex AI is a machine learning (ML) platform that lets you train
and deploy ML models and AI applications, and customize large language models
(LLMs) for use in your AI-powered applications. Vertex AI combines data
engineering, data science, and ML engineering workflows, enabling your
teams to collaborate using a common toolset and scale your applications using
the benefits of Google Cloud.
Vertex AI provides several options for model training
and deployment:
Model Garden
lets you discover, test, customize, and deploy
Vertex AI and select open-source models and assets.
Generative AI gives you access to Google's large generative AI
models for multiple modalities (text, code, images, speech). You can tune
Google's LLMs to meet your needs, and then deploy them
for use in your AI-powered applications.
After you deploy your models, use Vertex AI's end-to-end MLOps tools to
automate and scale projects throughout the ML lifecycle.
These MLOps tools are run on fully-managed infrastructure that you can customize
based on your performance and budget needs.
You can use the Vertex AI SDK for Python to run the entire machine
learning workflow in Vertex AI Workbench, a Jupyter
notebook-based development environment. You can collaborate with a team
to develop your model in Colab Enterprise,
a version of Colaboratory that is integrated with
Vertex AI. Other available interfaces
include the Google Cloud console, the Google Cloud CLI command line tool, client
libraries, and Terraform (limited support).
Vertex AI and the machine learning (ML) workflow
This section provides an overview of the machine learning workflow and how you
can use Vertex AI to build and deploy your models.
Data preparation: After extracting and cleaning your dataset, perform
exploratory data analysis (EDA) to understand the data schema and
characteristics that are expected by the ML model. Apply data transformations
and feature engineering to the model, and split the data into training,
validation, and test sets.
Explore and visualize data using Vertex AI Workbench
notebooks. Vertex AI Workbench integrates with Cloud Storage and
BigQuery to help you access and process your data faster.
For large datasets, use Dataproc Serverless Spark from a
Vertex AI Workbench notebook to run Spark workloads without having to
manage your own Dataproc clusters.
Model training: Choose a training method to train a model and tune it for
performance.
To train a model without writing code, see the AutoML
overview. AutoML supports tabular, image, and
video data.
To write your own training code and train custom models using your preferred
ML framework, see the Custom training overview.
Optimize hyperparameters for custom-trained models using custom tuning
jobs.
Vertex AI Vizier tunes hyperparameters for you in complex machine
learning (ML) models.
Use Vertex AI Experiments to train your model using
different ML techniques and compare the results.
Register your trained models in the
Vertex AI Model Registry for versioning and hand-off to
production. Vertex AI Model Registry integrates with validation and
deployment features such as model evaluation and endpoints.
Model evaluation and iteration: Evaluate your trained model, make
adjustments to your data based on evaluation metrics, and iterate on your
model.
Use model evaluation metrics, such as precision and recall, to
evaluate and compare the performance of your models. Create evaluations
through Vertex AI Model Registry, or include evaluations in your
Vertex AI Pipelines workflow.
Model serving: Deploy your model to production and get online
inferences or query it directly for batch inferences.
Deploy your custom-trained model using prebuilt or
custom containers to get real-time online
inferences (sometimes called HTTP inference).
Get asynchronous batch inferences, which don't require
deployment to endpoints.
Optimized TensorFlow runtime lets you serve TensorFlow
models at a lower cost and with lower latency than open source based
prebuilt TensorFlow serving containers.
For online serving cases with tabular models, use
Vertex AI Feature Store to serve features from a
central repository and monitor feature health.
Vertex Explainable AI helps you understand how each feature contributes to
model inference (feature attribution) and find mislabeled data from the
training dataset (example-based explanation).
Deploy and get online inferences for models trained with
BigQuery ML.
Model monitoring: Monitor the performance of your deployed model. Use
incoming inference data to retrain your model for improved performance.
Vertex AI Model Monitoring monitors models for
training-serving skew and inference drift and sends you alerts when the
incoming inference data skews too far from the training baseline.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-08-25 UTC."],[],[],null,["# Introduction to Vertex AI\n\nVertex AI is a machine learning (ML) platform that lets you train\nand deploy ML models and AI applications, and customize large language models\n(LLMs) for use in your AI-powered applications. Vertex AI combines data\nengineering, data science, and ML engineering workflows, enabling your\nteams to collaborate using a common toolset and scale your applications using\nthe benefits of Google Cloud.\n\nVertex AI provides several options for model [training](/vertex-ai/docs/start/training-methods)\nand deployment:\n\n- [AutoML](/vertex-ai/docs/beginner/beginners-guide) lets you train tabular, image, or video data\n without writing code or preparing data splits. These models can be\n [deployed for online inference or queried directly for batch inference](/vertex-ai/docs/predictions/overview#get_predictions_from_models).\n\n- [Custom training](/vertex-ai/docs/training/overview) gives you complete control over the training\n process, including using your preferred ML framework, writing your own\n training code, and choosing hyperparameter tuning options. You can import\n your custom-trained model into the [Model Registry](/vertex-ai/docs/model-registry/introduction)\n and [deploy it to an endpoint](/vertex-ai/docs/general/deployment) for online\n inference using [prebuilt](/vertex-ai/docs/predictions/pre-built-containers) or [custom](/vertex-ai/docs/predictions/use-custom-container) containers.\n Or you can\n [query it directly for batch inferences](/vertex-ai/docs/predictions/get-batch-predictions).\n\n- [Model Garden](/vertex-ai/generative-ai/docs/model-garden/explore-models)\n lets you discover, test, customize, and deploy\n Vertex AI and select open-source models and assets.\n\n- [Generative AI](/vertex-ai/generative-ai/docs/overview) gives you access to Google's large generative AI\n models for multiple modalities (text, code, images, speech). You can tune\n Google's LLMs to meet your needs, and then deploy them\n for use in your AI-powered applications.\n\nAfter you deploy your models, use Vertex AI's end-to-end MLOps tools to\nautomate and scale projects throughout the ML lifecycle.\nThese MLOps tools are run on fully-managed infrastructure that you can customize\nbased on your performance and budget needs.\n\nYou can use the [Vertex AI SDK for Python](/vertex-ai/docs/python-sdk/use-vertex-ai-python-sdk) to run the entire machine\nlearning workflow in [Vertex AI Workbench](/vertex-ai/docs/workbench/introduction), a Jupyter\nnotebook-based development environment. You can collaborate with a team\nto develop your model in [Colab Enterprise](/colab/docs/introduction),\na version of [Colaboratory](https://colab.google/) that is integrated with\nVertex AI. Other [available interfaces](/vertex-ai/docs/start/introduction-interfaces)\ninclude the Google Cloud console, the Google Cloud CLI command line tool, client\nlibraries, and Terraform (limited support).\n\nVertex AI and the machine learning (ML) workflow\n------------------------------------------------\n\nThis section provides an overview of the machine learning workflow and how you\ncan use Vertex AI to build and deploy your models.\n\n1. **Data preparation** : After extracting and cleaning your dataset, perform\n [exploratory data analysis (EDA)](/vertex-ai/docs/glossary#exploratory_data_analysis) to understand the data schema and\n characteristics that are expected by the ML model. Apply data transformations\n and feature engineering to the model, and split the data into training,\n validation, and test sets.\n\n - Explore and visualize data using [Vertex AI Workbench](/vertex-ai/docs/workbench/introduction)\n notebooks. Vertex AI Workbench integrates with Cloud Storage and\n BigQuery to help you access and process your data faster.\n\n - For large datasets, use [Dataproc Serverless Spark](/dataproc-serverless/docs/overview) from a\n Vertex AI Workbench notebook to run Spark workloads without having to\n manage your own Dataproc clusters.\n\n2. **Model training**: Choose a training method to train a model and tune it for\n performance.\n\n - To train a model without writing code, see the [AutoML\n overview](/vertex-ai/docs/training-overview#automl). AutoML supports tabular, image, and\n video data.\n\n - To write your own training code and train custom models using your preferred\n ML framework, see the [Custom training overview](/vertex-ai/docs/training/overview).\n\n - Optimize hyperparameters for custom-trained models using [custom tuning\n jobs](/vertex-ai/docs/training/using-hyperparameter-tuning).\n\n - [Vertex AI Vizier](/vertex-ai/docs/vizier/overview) tunes hyperparameters for you in complex machine\n learning (ML) models.\n\n - Use [Vertex AI Experiments](/vertex-ai/docs/experiments/intro-vertex-ai-experiments) to train your model using\n different ML techniques and compare the results.\n\n - Register your trained models in the\n [Vertex AI Model Registry](/vertex-ai/docs/model-registry/introduction) for versioning and hand-off to\n production. Vertex AI Model Registry integrates with validation and\n deployment features such as model evaluation and endpoints.\n\n3. **Model evaluation and iteration**: Evaluate your trained model, make\n adjustments to your data based on evaluation metrics, and iterate on your\n model.\n\n - Use [model evaluation](/vertex-ai/docs/evaluation/introduction) metrics, such as precision and recall, to evaluate and compare the performance of your models. Create evaluations through Vertex AI Model Registry, or include evaluations in your [Vertex AI Pipelines](/vertex-ai/docs/pipelines/introduction) workflow.\n4. **Model serving**: Deploy your model to production and get online\n inferences or query it directly for batch inferences.\n\n - Deploy your custom-trained model using [prebuilt](/vertex-ai/docs/predictions/pre-built-containers) or\n [custom](/vertex-ai/docs/predictions/use-custom-container) containers to get real-time [*online\n inferences*](/vertex-ai/docs/predictions/overview#online_predictions) (sometimes called HTTP inference).\n\n - Get asynchronous [*batch inferences*](/vertex-ai/docs/predictions/overview#batch_predictions), which don't require\n deployment to endpoints.\n\n - [Optimized TensorFlow runtime](/vertex-ai/docs/predictions/optimized-tensorflow-runtime) lets you serve TensorFlow\n models at a lower cost and with lower latency than open source based\n prebuilt TensorFlow serving containers.\n\n - For online serving cases with tabular models, use\n [Vertex AI Feature Store](/vertex-ai/docs/featurestore/overview) to serve features from a\n central repository and monitor feature health.\n\n - [Vertex Explainable AI](/vertex-ai/docs/explainable-ai/overview) helps you understand how each feature contributes to\n model inference (*feature attribution* ) and find mislabeled data from the\n training dataset (*example-based explanation*).\n\n - Deploy and get online inferences for models trained with\n [BigQuery ML](/vertex-ai/docs/beginner/bqml).\n\n5. **Model monitoring**: Monitor the performance of your deployed model. Use\n incoming inference data to retrain your model for improved performance.\n\n - [Vertex AI Model Monitoring](/vertex-ai/docs/model-monitoring/overview) monitors models for training-serving skew and inference drift and sends you alerts when the incoming inference data skews too far from the training baseline.\n\nWhat's next\n-----------\n\n- Learn about [Vertex AI's MLOps features](/vertex-ai/docs/start/introduction-mlops).\n\n- Learn about [interfaces that you can use to interact with\n Vertex AI](/vertex-ai/docs/start/introduction-interfaces)."]]