Google Cloud Dialogflow V2 Client - Class StreamingDetectIntentRequest (1.3.2)

Reference documentation and code samples for the Google Cloud Dialogflow V2 Client class StreamingDetectIntentRequest.

The top-level message sent by the client to the Sessions.StreamingDetectIntent method.

Multiple request messages should be sent in order:

  1. The first message must contain session, query_input plus optionally query_params. If the client wants to receive an audio response, it should also contain output_audio_config. The message must not contain input_audio.
  2. If query_input was set to query_input.audio_config, all subsequent messages must contain input_audio to continue with Speech recognition. If you decide to rather detect an intent from text input after you already started Speech recognition, please send a message with query_input.text. However, note that:
    • Dialogflow will bill you for the audio duration so far.
    • Dialogflow discards all Speech recognition results in favor of the input text.
    • Dialogflow will use the language code from the first message. After you sent all input, you must half-close or abort the request stream.

Generated from protobuf message google.cloud.dialogflow.v2.StreamingDetectIntentRequest

Methods

__construct

Constructor.

Parameters
NameDescription
data array

Optional. Data for populating the Message object.

↳ session string

Required. The name of the session the query is sent to. Format of the session name: projects/<Project ID>/agent/sessions/<Session ID>, or projects/<Project ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>. If Environment ID is not specified, we assume default 'draft' environment. If User ID is not specified, we are using "-". It's up to the API caller to choose an appropriate Session ID and User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the Session ID and User ID must not exceed 36 characters. For more information, see the API interactions guide. Note: Always use agent versions for production traffic. See Versions and environments.

↳ query_params Google\Cloud\Dialogflow\V2\QueryParameters

The parameters of this query.

↳ query_input Google\Cloud\Dialogflow\V2\QueryInput

Required. The input specification. It can be set to: 1. an audio config which instructs the speech recognizer how to process the speech audio, 2. a conversational query in the form of text, or 3. an event that specifies which intent to trigger.

↳ single_utterance bool

Please use InputAudioConfig.single_utterance instead. If false (default), recognition does not cease until the client closes the stream. If true, the recognizer will detect a single spoken utterance in input audio. Recognition ceases when it detects the audio's voice has stopped or paused. In this case, once a detected intent is received, the client should close the stream and start a new request with a new stream as needed. This setting is ignored when query_input is a piece of text or an event.

↳ output_audio_config Google\Cloud\Dialogflow\V2\OutputAudioConfig

Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.

↳ output_audio_config_mask Google\Protobuf\FieldMask

Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.

↳ input_audio string

The input audio content to be recognized. Must be sent if query_input was set to a streaming input audio config. The complete audio over all streaming messages must not exceed 1 minute.

getSession

Required. The name of the session the query is sent to.

Format of the session name: projects/<Project ID>/agent/sessions/<Session ID>, or projects/<Project ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>. If Environment ID is not specified, we assume default 'draft' environment. If User ID is not specified, we are using "-". It's up to the API caller to choose an appropriate Session ID and User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the Session ID and User ID must not exceed 36 characters. For more information, see the API interactions guide. Note: Always use agent versions for production traffic. See Versions and environments.

Returns
TypeDescription
string

setSession

Required. The name of the session the query is sent to.

Format of the session name: projects/<Project ID>/agent/sessions/<Session ID>, or projects/<Project ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>. If Environment ID is not specified, we assume default 'draft' environment. If User ID is not specified, we are using "-". It's up to the API caller to choose an appropriate Session ID and User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the Session ID and User ID must not exceed 36 characters. For more information, see the API interactions guide. Note: Always use agent versions for production traffic. See Versions and environments.

Parameter
NameDescription
var string
Returns
TypeDescription
$this

getQueryParams

The parameters of this query.

Returns
TypeDescription
Google\Cloud\Dialogflow\V2\QueryParameters|null

hasQueryParams

clearQueryParams

setQueryParams

The parameters of this query.

Parameter
NameDescription
var Google\Cloud\Dialogflow\V2\QueryParameters
Returns
TypeDescription
$this

getQueryInput

Required. The input specification. It can be set to:

  1. an audio config which instructs the speech recognizer how to process the speech audio,
  2. a conversational query in the form of text, or
  3. an event that specifies which intent to trigger.
Returns
TypeDescription
Google\Cloud\Dialogflow\V2\QueryInput|null

hasQueryInput

clearQueryInput

setQueryInput

Required. The input specification. It can be set to:

  1. an audio config which instructs the speech recognizer how to process the speech audio,
  2. a conversational query in the form of text, or
  3. an event that specifies which intent to trigger.
Parameter
NameDescription
var Google\Cloud\Dialogflow\V2\QueryInput
Returns
TypeDescription
$this

getSingleUtterance

Please use InputAudioConfig.single_utterance instead. If false (default), recognition does not cease until the client closes the stream. If true, the recognizer will detect a single spoken utterance in input audio. Recognition ceases when it detects the audio's voice has stopped or paused. In this case, once a detected intent is received, the client should close the stream and start a new request with a new stream as needed. This setting is ignored when query_input is a piece of text or an event.

Returns
TypeDescription
bool

setSingleUtterance

Please use InputAudioConfig.single_utterance instead. If false (default), recognition does not cease until the client closes the stream. If true, the recognizer will detect a single spoken utterance in input audio. Recognition ceases when it detects the audio's voice has stopped or paused. In this case, once a detected intent is received, the client should close the stream and start a new request with a new stream as needed. This setting is ignored when query_input is a piece of text or an event.

Parameter
NameDescription
var bool
Returns
TypeDescription
$this

getOutputAudioConfig

Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.

Returns
TypeDescription
Google\Cloud\Dialogflow\V2\OutputAudioConfig|null

hasOutputAudioConfig

clearOutputAudioConfig

setOutputAudioConfig

Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.

Parameter
NameDescription
var Google\Cloud\Dialogflow\V2\OutputAudioConfig
Returns
TypeDescription
$this

getOutputAudioConfigMask

Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level.

If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.

Returns
TypeDescription
Google\Protobuf\FieldMask|null

hasOutputAudioConfigMask

clearOutputAudioConfigMask

setOutputAudioConfigMask

Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level.

If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.

Parameter
NameDescription
var Google\Protobuf\FieldMask
Returns
TypeDescription
$this

getInputAudio

The input audio content to be recognized. Must be sent if query_input was set to a streaming input audio config. The complete audio over all streaming messages must not exceed 1 minute.

Returns
TypeDescription
string

setInputAudio

The input audio content to be recognized. Must be sent if query_input was set to a streaming input audio config. The complete audio over all streaming messages must not exceed 1 minute.

Parameter
NameDescription
var string
Returns
TypeDescription
$this