The natural language speech audio to be processed.
A single request can contain up to 1 minute of speech audio data.
The [transcribed text][google.cloud.dialogflow.cx.v3.QueryResult.transcript] cannot contain more than 256
bytes.
For non-streaming audio detect intent, both config and audio must be
provided.
For streaming audio detect intent, config must be provided in
the first request and audio must be provided in all following requests.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-04-09 UTC."],[[["This page details the `AudioInput` class, which is used to represent natural speech audio for processing in Google Cloud Dialogflow CX V3."],["The latest version of `AudioInput` is 2.23.0, with a history of versions dating back to 1.0.0, all of which can be accessed via links."],["The `AudioInput` class implements interfaces such as `IMessage`, `IEquatable`, `IDeepCloneable`, and `IBufferMessage`, indicating its role within the Google.Protobuf framework."],["`AudioInput` contains two key properties, `Audio` which represents the speech audio itself, and `Config`, which is required to tell the speech recognizer how to process the audio."],["`AudioInput` has a constructor that takes another `AudioInput` object, for use in cloning."]]],[]]