Class ImageCaptioningModel (1.95.1)
Stay organized with collections
Save and categorize content based on your preferences.
ImageCaptioningModel(model_id: str, endpoint_name: typing.Optional[str] = None)
Generates captions from image.
Examples::
model = ImageCaptioningModel.from_pretrained("imagetext@001")
image = Image.load_from_file("image.png")
captions = model.get_captions(
image=image,
# Optional:
number_of_results=1,
language="en",
)
Methods
ImageCaptioningModel
ImageCaptioningModel(model_id: str, endpoint_name: typing.Optional[str] = None)
Creates a _ModelGardenModel.
This constructor should not be called directly.
Use {model_class}.from_pretrained(model_name=...)
instead.
from_pretrained
from_pretrained(model_name: str) -> vertexai._model_garden._model_garden_models.T
Loads a _ModelGardenModel.
Exceptions |
Type |
Description |
ValueError |
If model_name is unknown. |
ValueError |
If model does not support this class. |
get_captions
get_captions(
image: vertexai.vision_models.Image,
*,
number_of_results: int = 1,
language: str = "en",
output_gcs_uri: typing.Optional[str] = None
) -> typing.List[str]
Generates captions for a given image.
Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.
Last updated 2025-08-07 UTC.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-08-07 UTC."],[],[],null,["# Class ImageCaptioningModel (1.95.1)\n\nVersion latestkeyboard_arrow_down\n\n- [1.95.1 (latest)](/python/docs/reference/vertexai/latest/vertexai.vision_models.ImageCaptioningModel)\n- [1.94.0](/python/docs/reference/vertexai/1.94.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.93.1](/python/docs/reference/vertexai/1.93.1/vertexai.vision_models.ImageCaptioningModel)\n- [1.92.0](/python/docs/reference/vertexai/1.92.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.91.0](/python/docs/reference/vertexai/1.91.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.90.0](/python/docs/reference/vertexai/1.90.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.89.0](/python/docs/reference/vertexai/1.89.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.88.0](/python/docs/reference/vertexai/1.88.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.87.0](/python/docs/reference/vertexai/1.87.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.86.0](/python/docs/reference/vertexai/1.86.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.85.0](/python/docs/reference/vertexai/1.85.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.84.0](/python/docs/reference/vertexai/1.84.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.83.0](/python/docs/reference/vertexai/1.83.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.82.0](/python/docs/reference/vertexai/1.82.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.81.0](/python/docs/reference/vertexai/1.81.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.80.0](/python/docs/reference/vertexai/1.80.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.79.0](/python/docs/reference/vertexai/1.79.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.78.0](/python/docs/reference/vertexai/1.78.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.77.0](/python/docs/reference/vertexai/1.77.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.76.0](/python/docs/reference/vertexai/1.76.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.75.0](/python/docs/reference/vertexai/1.75.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.74.0](/python/docs/reference/vertexai/1.74.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.73.0](/python/docs/reference/vertexai/1.73.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.72.0](/python/docs/reference/vertexai/1.72.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.71.1](/python/docs/reference/vertexai/1.71.1/vertexai.vision_models.ImageCaptioningModel)\n- [1.70.0](/python/docs/reference/vertexai/1.70.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.69.0](/python/docs/reference/vertexai/1.69.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.68.0](/python/docs/reference/vertexai/1.68.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.67.1](/python/docs/reference/vertexai/1.67.1/vertexai.vision_models.ImageCaptioningModel)\n- [1.66.0](/python/docs/reference/vertexai/1.66.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.65.0](/python/docs/reference/vertexai/1.65.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.63.0](/python/docs/reference/vertexai/1.63.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.62.0](/python/docs/reference/vertexai/1.62.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.60.0](/python/docs/reference/vertexai/1.60.0/vertexai.vision_models.ImageCaptioningModel)\n- [1.59.0](/python/docs/reference/vertexai/1.59.0/vertexai.vision_models.ImageCaptioningModel) \n\n ImageCaptioningModel(model_id: str, endpoint_name: typing.Optional[str] = None)\n\nGenerates captions from image.\n\nExamples:: \n\n model = ImageCaptioningModel.from_pretrained(\"imagetext@001\")\n image = Image.load_from_file(\"image.png\")\n captions = model.get_captions(\n image=image,\n # Optional:\n number_of_results=1,\n language=\"en\",\n )\n\nMethods\n-------\n\n### ImageCaptioningModel\n\n ImageCaptioningModel(model_id: str, endpoint_name: typing.Optional[str] = None)\n\nCreates a _ModelGardenModel.\n\nThis constructor should not be called directly.\nUse `{model_class}.from_pretrained(model_name=...)` instead.\n\n### from_pretrained\n\n from_pretrained(model_name: str) -\u003e vertexai._model_garden._model_garden_models.T\n\nLoads a _ModelGardenModel.\n\n### get_captions\n\n get_captions(\n image: vertexai.vision_models.Image,\n *,\n number_of_results: int = 1,\n language: str = \"en\",\n output_gcs_uri: typing.Optional[str] = None\n ) -\u003e typing.List[str]\n\nGenerates captions for a given image."]]