Send feedback
Class ImageTextModel (1.95.1)
Stay organized with collections
Save and categorize content based on your preferences.
Version latestkeyboard_arrow_down
ImageTextModel ( model_id : str , endpoint_name : typing . Optional [ str ] = None )
Generates text from images.
Examples::
model = ImageTextModel.from_pretrained("imagetext@001")
image = Image.load_from_file("image.png")
captions = model.get_captions(
image=image,
# Optional:
number_of_results=1,
language="en",
)
answers = model.ask_question(
image=image,
question="What color is the car in this image?",
# Optional:
number_of_results=1,
)
Methods
ImageTextModel
ImageTextModel ( model_id : str , endpoint_name : typing . Optional [ str ] = None )
Creates a _ModelGardenModel.
This constructor should not be called directly.
Use {model_class}.from_pretrained(model_name=...)
instead.
ask_question
ask_question (
image : vertexai . vision_models . Image , question : str , * , number_of_results : int = 1
) - > typing . List [ str ]
Answers questions about an image.
from_pretrained
from_pretrained ( model_name : str ) - > vertexai . _model_garden . _model_garden_models . T
Loads a _ModelGardenModel.
Exceptions
Type
Description
ValueError
If model_name is unknown.
ValueError
If model does not support this class.
get_captions
get_captions (
image : vertexai . vision_models . Image ,
* ,
number_of_results : int = 1 ,
language : str = "en" ,
output_gcs_uri : typing . Optional [ str ] = None
) - > typing . List [ str ]
Generates captions for a given image.
Send feedback
Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License , and code samples are licensed under the Apache 2.0 License . For details, see the Google Developers Site Policies . Java is a registered trademark of Oracle and/or its affiliates.
Last updated 2025-08-07 UTC.
Need to tell us more?
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-08-07 UTC."],[],[],null,["# Class ImageTextModel (1.95.1)\n\nVersion latestkeyboard_arrow_down\n\n- [1.95.1 (latest)](/python/docs/reference/vertexai/latest/vertexai.vision_models.ImageTextModel)\n- [1.94.0](/python/docs/reference/vertexai/1.94.0/vertexai.vision_models.ImageTextModel)\n- [1.93.1](/python/docs/reference/vertexai/1.93.1/vertexai.vision_models.ImageTextModel)\n- [1.92.0](/python/docs/reference/vertexai/1.92.0/vertexai.vision_models.ImageTextModel)\n- [1.91.0](/python/docs/reference/vertexai/1.91.0/vertexai.vision_models.ImageTextModel)\n- [1.90.0](/python/docs/reference/vertexai/1.90.0/vertexai.vision_models.ImageTextModel)\n- [1.89.0](/python/docs/reference/vertexai/1.89.0/vertexai.vision_models.ImageTextModel)\n- [1.88.0](/python/docs/reference/vertexai/1.88.0/vertexai.vision_models.ImageTextModel)\n- [1.87.0](/python/docs/reference/vertexai/1.87.0/vertexai.vision_models.ImageTextModel)\n- [1.86.0](/python/docs/reference/vertexai/1.86.0/vertexai.vision_models.ImageTextModel)\n- [1.85.0](/python/docs/reference/vertexai/1.85.0/vertexai.vision_models.ImageTextModel)\n- [1.84.0](/python/docs/reference/vertexai/1.84.0/vertexai.vision_models.ImageTextModel)\n- [1.83.0](/python/docs/reference/vertexai/1.83.0/vertexai.vision_models.ImageTextModel)\n- [1.82.0](/python/docs/reference/vertexai/1.82.0/vertexai.vision_models.ImageTextModel)\n- [1.81.0](/python/docs/reference/vertexai/1.81.0/vertexai.vision_models.ImageTextModel)\n- [1.80.0](/python/docs/reference/vertexai/1.80.0/vertexai.vision_models.ImageTextModel)\n- [1.79.0](/python/docs/reference/vertexai/1.79.0/vertexai.vision_models.ImageTextModel)\n- [1.78.0](/python/docs/reference/vertexai/1.78.0/vertexai.vision_models.ImageTextModel)\n- [1.77.0](/python/docs/reference/vertexai/1.77.0/vertexai.vision_models.ImageTextModel)\n- [1.76.0](/python/docs/reference/vertexai/1.76.0/vertexai.vision_models.ImageTextModel)\n- [1.75.0](/python/docs/reference/vertexai/1.75.0/vertexai.vision_models.ImageTextModel)\n- [1.74.0](/python/docs/reference/vertexai/1.74.0/vertexai.vision_models.ImageTextModel)\n- [1.73.0](/python/docs/reference/vertexai/1.73.0/vertexai.vision_models.ImageTextModel)\n- [1.72.0](/python/docs/reference/vertexai/1.72.0/vertexai.vision_models.ImageTextModel)\n- [1.71.1](/python/docs/reference/vertexai/1.71.1/vertexai.vision_models.ImageTextModel)\n- [1.70.0](/python/docs/reference/vertexai/1.70.0/vertexai.vision_models.ImageTextModel)\n- [1.69.0](/python/docs/reference/vertexai/1.69.0/vertexai.vision_models.ImageTextModel)\n- [1.68.0](/python/docs/reference/vertexai/1.68.0/vertexai.vision_models.ImageTextModel)\n- [1.67.1](/python/docs/reference/vertexai/1.67.1/vertexai.vision_models.ImageTextModel)\n- [1.66.0](/python/docs/reference/vertexai/1.66.0/vertexai.vision_models.ImageTextModel)\n- [1.65.0](/python/docs/reference/vertexai/1.65.0/vertexai.vision_models.ImageTextModel)\n- [1.63.0](/python/docs/reference/vertexai/1.63.0/vertexai.vision_models.ImageTextModel)\n- [1.62.0](/python/docs/reference/vertexai/1.62.0/vertexai.vision_models.ImageTextModel)\n- [1.60.0](/python/docs/reference/vertexai/1.60.0/vertexai.vision_models.ImageTextModel)\n- [1.59.0](/python/docs/reference/vertexai/1.59.0/vertexai.vision_models.ImageTextModel) \n\n ImageTextModel(model_id: str, endpoint_name: typing.Optional[str] = None)\n\nGenerates text from images.\n\nExamples:: \n\n model = ImageTextModel.from_pretrained(\"imagetext@001\")\n image = Image.load_from_file(\"image.png\")\n\n captions = model.get_captions(\n image=image,\n # Optional:\n number_of_results=1,\n language=\"en\",\n )\n\n answers = model.ask_question(\n image=image,\n question=\"What color is the car in this image?\",\n # Optional:\n number_of_results=1,\n )\n\nMethods\n-------\n\n### ImageTextModel\n\n ImageTextModel(model_id: str, endpoint_name: typing.Optional[str] = None)\n\nCreates a _ModelGardenModel.\n\nThis constructor should not be called directly.\nUse `{model_class}.from_pretrained(model_name=...)` instead.\n\n### ask_question\n\n ask_question(\n image: vertexai.vision_models.Image, question: str, *, number_of_results: int = 1\n ) -\u003e typing.List[str]\n\nAnswers questions about an image.\n\n### from_pretrained\n\n from_pretrained(model_name: str) -\u003e vertexai._model_garden._model_garden_models.T\n\nLoads a _ModelGardenModel.\n\n### get_captions\n\n get_captions(\n image: vertexai.vision_models.Image,\n *,\n number_of_results: int = 1,\n language: str = \"en\",\n output_gcs_uri: typing.Optional[str] = None\n ) -\u003e typing.List[str]\n\nGenerates captions for a given image."]]