public sealed class StreamingAnalyzeContentRequest : IMessage<StreamingAnalyzeContentRequest>, IEquatable<StreamingAnalyzeContentRequest>, IDeepCloneable<StreamingAnalyzeContentRequest>, IBufferMessage, IMessageReference documentation and code samples for the Google Cloud Dialogflow v2 API class StreamingAnalyzeContentRequest.
The top-level message sent by the client to the [Participants.StreamingAnalyzeContent][google.cloud.dialogflow.v2.Participants.StreamingAnalyzeContent] method.
Multiple request messages should be sent in order:
The first message must contain [participant][google.cloud.dialogflow.v2.StreamingAnalyzeContentRequest.participant], [config][google.cloud.dialogflow.v2.StreamingAnalyzeContentRequest.config] and optionally [query_params][google.cloud.dialogflow.v2.StreamingAnalyzeContentRequest.query_params]. If you want to receive an audio response, it should also contain [reply_audio_config][google.cloud.dialogflow.v2.StreamingAnalyzeContentRequest.reply_audio_config]. The message must not contain [input][google.cloud.dialogflow.v2.StreamingAnalyzeContentRequest.input].
If [config][google.cloud.dialogflow.v2.StreamingAnalyzeContentRequest.config] in the first message was set to [audio_config][google.cloud.dialogflow.v2.StreamingAnalyzeContentRequest.audio_config], all subsequent messages must contain [input_audio][google.cloud.dialogflow.v2.StreamingAnalyzeContentRequest.input_audio] to continue with Speech recognition. However, note that:
- Dialogflow will bill you for the audio so far.
- Dialogflow discards all Speech recognition results in favor of the text input.
- If [StreamingAnalyzeContentRequest.config][google.cloud.dialogflow.v2.StreamingAnalyzeContentRequest.config] in the first message was set to [StreamingAnalyzeContentRequest.text_config][google.cloud.dialogflow.v2.StreamingAnalyzeContentRequest.text_config], then the second message must contain only [input_text][google.cloud.dialogflow.v2.StreamingAnalyzeContentRequest.input_text]. Moreover, you must not send more than two messages.
After you sent all input, you must half-close or abort the request stream.
Implements
IMessageStreamingAnalyzeContentRequest, IEquatableStreamingAnalyzeContentRequest, IDeepCloneableStreamingAnalyzeContentRequest, IBufferMessage, IMessageNamespace
Google.Cloud.Dialogflow.V2Assembly
Google.Cloud.Dialogflow.V2.dll
Constructors
StreamingAnalyzeContentRequest()
public StreamingAnalyzeContentRequest()StreamingAnalyzeContentRequest(StreamingAnalyzeContentRequest)
public StreamingAnalyzeContentRequest(StreamingAnalyzeContentRequest other)| Parameter | |
|---|---|
| Name | Description |
other |
StreamingAnalyzeContentRequest |
Properties
AssistQueryParams
public AssistQueryParameters AssistQueryParams { get; set; }Parameters for a human assist query.
| Property Value | |
|---|---|
| Type | Description |
AssistQueryParameters |
|
AudioConfig
public InputAudioConfig AudioConfig { get; set; }Instructs the speech recognizer how to process the speech audio.
| Property Value | |
|---|---|
| Type | Description |
InputAudioConfig |
|
ConfigCase
public StreamingAnalyzeContentRequest.ConfigOneofCase ConfigCase { get; }| Property Value | |
|---|---|
| Type | Description |
StreamingAnalyzeContentRequestConfigOneofCase |
|
CxParameters
public Struct CxParameters { get; set; }Additional parameters to be put into Dialogflow CX session parameters. To remove a parameter from the session, clients should explicitly set the parameter value to null.
Note: this field should only be used if you are connecting to a Dialogflow CX agent.
| Property Value | |
|---|---|
| Type | Description |
Struct |
|
EnableDebuggingInfo
public bool EnableDebuggingInfo { get; set; }If true, StreamingAnalyzeContentResponse.debugging_info will get
populated.
| Property Value | |
|---|---|
| Type | Description |
bool |
|
EnableExtendedStreaming
public bool EnableExtendedStreaming { get; set; }Optional. Enable full bidirectional streaming. You can keep streaming the audio until timeout, and there's no need to half close the stream to get the response.
Restrictions:
- Timeout: 3 mins.
- Audio Encoding: only supports [AudioEncoding.AUDIO_ENCODING_LINEAR_16][google.cloud.dialogflow.v2.AudioEncoding.AUDIO_ENCODING_LINEAR_16] and [AudioEncoding.AUDIO_ENCODING_MULAW][google.cloud.dialogflow.v2.AudioEncoding.AUDIO_ENCODING_MULAW]
- Lifecycle: conversation should be in
Assist Stage, go to [Conversations.CreateConversation][google.cloud.dialogflow.v2.Conversations.CreateConversation] for more information.
InvalidArgument Error will be returned if the one of restriction checks failed.
You can find more details in https://cloud.google.com/agent-assist/docs/extended-streaming
| Property Value | |
|---|---|
| Type | Description |
bool |
|
EnablePartialAutomatedAgentReply
public bool EnablePartialAutomatedAgentReply { get; set; }Optional. Enable partial responses from Dialogflow CX agent. If this flag
is not enabled, response stream still contains only one final response even
if some Fulfillments in Dialogflow CX agent have been configured to
return partial responses.
| Property Value | |
|---|---|
| Type | Description |
bool |
|
HasInputAudio
public bool HasInputAudio { get; }Gets whether the "input_audio" field is set
| Property Value | |
|---|---|
| Type | Description |
bool |
|
HasInputText
public bool HasInputText { get; }Gets whether the "input_text" field is set
| Property Value | |
|---|---|
| Type | Description |
bool |
|
InputAudio
public ByteString InputAudio { get; set; }The input audio content to be recognized. Must be sent if audio_config
is set in the first message. The complete audio over all streaming
messages must not exceed 1 minute.
| Property Value | |
|---|---|
| Type | Description |
ByteString |
|
InputCase
public StreamingAnalyzeContentRequest.InputOneofCase InputCase { get; }| Property Value | |
|---|---|
| Type | Description |
StreamingAnalyzeContentRequestInputOneofCase |
|
InputDtmf
public TelephonyDtmfEvents InputDtmf { get; set; }The DTMF digits used to invoke intent and fill in parameter value.
This input is ignored if the previous response indicated that DTMF input is not accepted.
| Property Value | |
|---|---|
| Type | Description |
TelephonyDtmfEvents |
|
InputText
public string InputText { get; set; }The UTF-8 encoded natural language text to be processed. Must be sent if
text_config is set in the first message. Text length must not exceed
256 bytes for virtual agent interactions. The input_text field can be
only sent once, and would cancel the speech recognition if any ongoing.
| Property Value | |
|---|---|
| Type | Description |
string |
|
Participant
public string Participant { get; set; }Required. The name of the participant this text comes from.
Format: projects/<Project ID>/locations/<Location
ID>/conversations/<Conversation ID>/participants/<Participant ID>.
| Property Value | |
|---|---|
| Type | Description |
string |
|
ParticipantAsParticipantName
public ParticipantName ParticipantAsParticipantName { get; set; }ParticipantName-typed view over the Participant resource name property.
| Property Value | |
|---|---|
| Type | Description |
ParticipantName |
|
QueryParams
public QueryParameters QueryParams { get; set; }Parameters for a Dialogflow virtual-agent query.
| Property Value | |
|---|---|
| Type | Description |
QueryParameters |
|
ReplyAudioConfig
public OutputAudioConfig ReplyAudioConfig { get; set; }Speech synthesis configuration. The speech synthesis settings for a virtual agent that may be configured for the associated conversation profile are not used when calling StreamingAnalyzeContent. If this configuration is not supplied, speech synthesis is disabled.
| Property Value | |
|---|---|
| Type | Description |
OutputAudioConfig |
|
TextConfig
public InputTextConfig TextConfig { get; set; }The natural language text to be processed.
| Property Value | |
|---|---|
| Type | Description |
InputTextConfig |
|