[[["易于理解","easyToUnderstand","thumb-up"],["解决了我的问题","solvedMyProblem","thumb-up"],["其他","otherUp","thumb-up"]],[["很难理解","hardToUnderstand","thumb-down"],["信息或示例代码不正确","incorrectInformationOrSampleCode","thumb-down"],["没有我需要的信息/示例","missingTheInformationSamplesINeed","thumb-down"],["翻译问题","translationIssue","thumb-down"],["其他","otherDown","thumb-down"]],["最后更新时间 (UTC):2025-09-04。"],[],[],null,["# Text embeddings overview\n\n| **Important:** This content applies to version 1.14.4 and later.\n\nEmbeddings are numerical representations of text, images, or videos that capture\nrelationships between inputs. Machine learning models, especially\nGenerative AI models, are suited for creating embeddings by identifying\npatterns within large datasets. Applications can use embeddings to process and\nproduce language, recognizing complex meanings and semantic relationships\nspecific to your content.\n\nVertex AI on Google Distributed Cloud (GDC) air-gapped supports\nText Embedding APIs for English and multilingual textual input.\nText Embedding works by converting text into arrays of floating-point\nnumbers called vectors. These vectors are designed to capture the meaning of the\ntext. The length of the embedding array is called the vector's dimensionality.\nFor example, one passage of text might be represented by a vector containing\nhundreds of dimensions. Then, by calculating the numerical distance between the\nvector representations of two pieces of text, an application can determine the\nsimilarity between the objects.\n| **Note:** Only text is supported as an input type for embeddings on GDC.\n\nFor a list of supported models, see\n[Embeddings models](/distributed-cloud/hosted/docs/latest/gdch/application/ao-user/genai/genai-models#embeddings-models).\n\nFor a list of supported multilingual languages, see\n[Supported text embedding languages](/distributed-cloud/hosted/docs/latest/gdch/application/ao-user/genai/supported-text-languages).\n\nText embeddings use cases\n-------------------------\n\nSome common use cases for text embeddings include:\n\n- **Semantic search**: Search text ranked by semantic similarity.\n- **Classification**: Return the class of items with attributes similar to the given text.\n- **Clustering**: Cluster items whose text attributes are similar to the given text.\n- **Outlier detection**: Return items where text attributes are least related to the given text.\n- **Conversational interface**: Cluster groups of sentences that can lead to similar responses, like in a conversation-level embedding space.\n\n### Example use case: Develop a book recommendation chatbot\n\nIf you want to develop a book recommendation chatbot, the first thing to do is\nto use a deep neural network (DNN) to convert each book into an embedding\nvector, where one embedding vector represents one book. You can feed the book\ntitle or text content as input to the DNN. Alternatively, you can use both of\nthese inputs together, along with any other metadata describing the book, such\nas the genre.\n\nThe embeddings in this example could include thousands of book titles with\nsummaries and their genre. It might have representations for books like\n*Wuthering Heights* by Emily Brontë and *Persuasion* by Jane Austen that are\nsimilar to each other (small distance between numerical representation). In\ncontrast, the numerical representation of the book *The Great Gatsby* by F.\nScott Fitzgerald would be further, as the time period, genre, and summary are\nless similar.\n\nThe inputs are the main influence on the orientation of the embedding space. For\nexample, if we only had book title inputs, then two books with similar titles\nbut very different summaries could be close together. However, if we include the\ntitle and summary, then these same books are less similar (further away) in the\nembedding space.\n\nWorking with Generative AI, this book-suggestion chatbot could summarize,\nsuggest, and show you books that you might like (or dislike) based on your\nquery.\n\nWhat's next\n-----------\n\n- Learn how to [get text embeddings](/distributed-cloud/hosted/docs/latest/gdch/application/ao-user/genai/get-text-embeddings)."]]