Multilingual E5 Large

Multilingual E5 Large is part of the E5 family of text embedding models. This large variant contains 24 layers. This model is well-suited for tasks such as:

  • Semantic Search: Finding documents or text passages that are semantically relevant to a query.
  • Text Clustering: Grouping similar pieces of text based on their semantic meaning.
  • Informative Response Generation: Following instructions to provide summaries, explanations, or answers to factual queries.

View model card in Model Garden

Model ID multilingual-e5-large-instruct-maas
Launch stage GA
Supported inputs & outputs
  • Inputs:
    Text
  • Outputs:
    Embeddings
Output dimensions Up to 1,024
Number of layers 24
Max sequence length 512 tokens
Supported languages See Supported languages.
Usage types
Versions
  • multilingual-e5-large
    • Launch stage: GA
    • Release date: September 19, 2025
Supported regions

Model availability

  • United States
    • us-central1
  • Europe
    • europe-west4

ML processing

  • United States
    • Multi-region
  • Europe
    • Multi-region
Quota limits

us-central1:

  • TPM: 3,000

europe-west4:

  • TPM: 3,000